Newsletters




Big Data

The well-known three Vs of Big Data - Volume, Variety, and Velocity – are increasingly placing pressure on organizations that need to manage this data as well as extract value from this data deluge for Predictive Analytics and Decision-Making. Big Data technologies, services, and tools such as Hadoop, MapReduce, Hive and NoSQL/NewSQL databases and Data Integration techniques, In-Memory approaches, and Cloud technologies have emerged to help meet the challenges posed by the flood of Web, Social Media, Internet of Things (IoT) and machine-to-machine (M2M) data flowing into organizations.



Big Data Articles

At Data Summit 2015 in New York City, Tony Shan, chief architect, Wipro, gave a talk on the key components of a successful big data methodology and shared lessons learned from real world big data implementations. According to Shan, there is an 8-step process for a big data framework with specific techniques and methods.

Posted May 14, 2015

The data lake is one of the hottest topics in the data industry today. It is a massive storage reservoir that allows data to be stored in its rawest forms. Hadoop Day at Data Summit 2015 concluded with a panel on everything data lake featuring James Casaletto, solutions architect for MapR, Joe Caserta, president and founder of Caserta Concepts, and George Corugedo, CTO with RedPoint Global Inc.

Posted May 14, 2015

With the influx of big data solutions and technologies comes a bevy of new problems, according to Data Summit 2015 panelists Miles Kehoe, search evangelist at Avalon Consulting, and Anne Buff, business solutions manager for SAS best practices at the SAS Institute. Kehoe and Buff opened the second day of Data Summit with a keynote discussion focusing on resolving data conundrums.

Posted May 14, 2015

To transform data into value, IT must move from thinking about what it does to data, and instead focus on business outcomes and what can be done with the data to advance the business, according to Edd Dumbill, vice president, strategy, Silicon Valley Data Science, who gave the welcome keynote at Data Summit 2015.

Posted May 14, 2015

If used correctly, machine data can provide a company a significant advantage in terms of understanding user and machine behavior, fighting cyber security risks and fraudulent behavior, service levels and customer behavior. In his talk at Data Summit 2015, Dejan Deklich, vice president, engineering platform and cloud at Splunk, discussed issues around machine data analysis and showcased some prominent use cases.

Posted May 13, 2015

In order to break down barriers in creating and storing data, understanding the modern data architecture is key. That was the focus of Mike Lamble, CEO at Clarity Solution Group, and Ron Huizenga's, product manager at Embarcadero Technologies, presentation at Data Summit 2015.

Posted May 12, 2015

Capgemini is extending its long-standing strategic partnership with SAP, allowing Capgemini to act as a single point of contact for customers globally, and delivering SAP products and support services through one consolidated framework. By signing a global value-added reseller (VAR) agreement with SAP, Capgemini is among a select group of global SAP partners that are part of the global program, which has specific entry requirements that include global reach, reseller capabilities and revenue targets.

Posted May 12, 2015

Splice Machine is partnering with Talend to enable customers to simplify data integration and streamline data workflows on Hadoop. Through this partnership, organizations building operational data lakes with Splice Machine can augment Talend's data integration technology with its data quality capabilities.

Posted May 12, 2015

Pentaho users will now be able to use Apache Spark within Pentaho thanks to a new native integration solution that will enable the orchestration of all Spark jobs. Pentaho Data Integration (PDI), an effort initiated by Pentaho Labs, will enable customers to increase productivity, reduce maintenance costs, and dramatically lower the skill sets required as Spark is incorporated into big data projects.

Posted May 12, 2015

HP has made multiple contributions to the OpenStack Kilo release, including new converged storage management automation and new flash storage technologies to support flexible, enterprise-class clouds. HP's storage contributions to the OpenStack Kilo release focus on two strategic goals.

Posted May 11, 2015

Teradata has made enhancements to the Teradata Database's hybrid row and column capabilities to provide quicker access to data stored on columnar tables and drive faster query performance. Other relational database management systems store data tables in rows or columns, and each method offers benefits, depending on the application and type of data. However, they have been mutually exclusive. Teradata's new hybrid row and column capabilities allow the best of both worlds.

Posted May 08, 2015

Cloudera is now offering support for Capgemini's new reference architecture for the SAP HANA platform and Cloudera Enterprise. "By bringing the power of Cloudera's enterprise data hub offering to the ecosystem in support of SAP HANA, we can enable Capgemini's clients to expand the amount of data they have within their environment in a cost-efficient manner," said Tim Stevens, vice president of corporate and business development at Cloudera.

Posted May 08, 2015

The certification enables Nimble Storage to participate in SAP's program for SAP HANA tailored data center integration using its certified solutions. Through participation in the program, customers can leverage their existing hardware and infrastructure components for their SAP HANA-based environments, providing further choice for organizations even when working in heterogeneous environments.

Posted May 07, 2015

Software AG has made updates to its Terracotta In-Memory Data Management platform. New improvements to Terracotta Open Source Kit 4.3 include distributed storage and off-heap storage.The platform is used for boosting performance, scalability, and building real-time applications. Additionally, Terracotta helps developers leverage in-memory storage for current and emerging data workloads.

Posted May 07, 2015

Tableau's cloud analytic solution, Tableau Online, is being upgraded to version 9.0. The new release enables faster performance, and provides additional live database connection support, single sign-on, and other new features designed to help users do more with their data in the cloud. The new update brings a complete redesign of Tableau Online to deliver a faster, more scalable, resilient, and extensible platform with capabilities such as parallel queries, query fusion, vectorization and smarter query caches that will make Tableau Online as much as 10 times faster.

Posted May 07, 2015

When databases are built from a well-designed data model, the resulting structures provide increased value to the organization. The value derived from the data model exhibits itself in the form of minimized redundancy, maximized data integrity, increased stability, better data sharing, increased consistency, more timely access to data, and better usability.

Posted May 06, 2015

Pivotal has made updates to its big data suite that include upgrades to the Pivotal HD enterprise-grade Apache Hadoop distribution, which is now based on the Open Data Platform core, and performance improvements for Pivotal Greenplum Database.

Posted May 05, 2015

The Spring 2015 release of the SnapLogic Elastic Integration Platform extends the platform's cloud and big data integration capabilities to the Internet of Things (IoT) with support for Message Queuing Telemetry Transport (MQTT), a lightweight machine-to-machine connectivity protocol.

Posted May 05, 2015

Deep Information Sciences has closed $8 million in Series A funding. The round brings the total invested in Deep to $18 million. The funding will assist in the growth of the Deep Engine, which break downs the performance, speed and scale limitations of databases to help businesses achieve new insights and opportunities from big data.

Posted May 05, 2015

Dell is partnering with Datawatch Corporation to continue growing its analytics business by integrating Datawatch's interactive visualization and dashboarding capabilities directly into its Statistica advanced analytics platform.

Posted April 30, 2015

CA Workload Automation Advanced Integration 1.0 for SAP Business Warehouse has received SAP certification. Specifically, the SAP Integration and Certification Center has certified that CA Workload Automation Advanced Integration 1.0 integrates with SAP Business Warehouse to provide a unified view for jobs running in both SAP and non-SAP applications.

Posted April 30, 2015

BackOffice Associates' HiT Software division, a provider of data replication and change data capture solutions for heterogeneous database environments, has announced the release of version 8.5 of its flagship product DBMoto.

Posted April 29, 2015

Splice Machine, a provider of Hadoop RDMS, announced that it is partnering with mrc (michaels, ross & cole ltd), to allow Splice Machine's Hadoop RDBMS to be certified and integrated with mrc's m-Power platform. "Our partnership with mrc gives businesses a solution that can speed real-time application deployment on Hadoop with the staff and tools they currently have, while also offering affordable scale-out on commodity hardware for future growth," said Monte Zweben, co-founder and CEO, Splice Machine.

Posted April 28, 2015

Embarcadero Technologies, a provider of software solutions for application and database development, has unveiled the new XE7 version of ER/Studio, its flagship data architecture suite.

Posted April 28, 2015

ProfitBricks, a provider of cloud infrastructure for IaaS, has announced the release of a Node.js SDK and an SDK for Ruby, written against its recently launched REST API.

Posted April 27, 2015

Cloud technology was a dominant focus at COLLABORATE 15, which took place earlier this month, according to Melissa English, president of the Oracle Applications Users Group (OAUG). "What's on top of everybody's mind is cloud strategy," English noted.

Posted April 27, 2015

Predixion Software, a developer of cloud-based predictive analytics (PA) software, announced that Software AG will lead the company's series D funding round. the company says that this fourth round of funding, which includes participation from existing financial and strategic investors, including GE Software Ventures, will support Predixion's move into the Internet of Things (IoT) analytics market.

Posted April 27, 2015

Pivotal HAWQ is now available on the Hortonworks Data Platform (HDP), enabling the benefits of SQL on Hadoop to be leveraged by enterprises that are investing in HDP. This marks the first time that the features and capabilities of Pivotal HAWQ have been made available outside of Pivotal. The availability aligns with a common Open Data Platform (ODP) Core that allows users to leverage the best-of-breed technology across providers.

Posted April 27, 2015

The future will flourish with machines. We've been told this in pop culture for decades, from the helpful robots of the Jetsons, to the infamous Skynet of the Terminator movies, to the omniscient "computer" of Star Trek. Smart, connected devices will be ubiquitous and it's up to us, the humans, to decide what's next. But the Internet of Things (IoT) is about more than devices and data.

Posted April 23, 2015

SUSE and Veristorm are partnering to provide certified high-performance Hadoop solutions that run directly on Linux on IBM z Systems, IBM Power Systems, and x86-64. Customers with IBM z Systems can team SUSE Linux Enterprise Server for System z with Veristorm zDoop, a commercial distribution of Hadoop supported on mainframes.

Posted April 23, 2015

There are actually many advantages to adopting or subscribing to a cloud-based data services infrastructure. For starters—and this may be the only reason companies need to make the move—there's the simplicity cloud and data as a service can offer. In many ways, cloud and data as a service free enterprises and their data teams from the technical intricacies of deploying systems and solutions.

Posted April 22, 2015

Progress Software has introduced a preview program for a standards-based connectivity solution to deliver fast transactions and analytics for SAP HANA. Called "Progress DataDirect ODBC" for SAP HANA, the connectivity solution will support both high-volume transactional workloads and massive analytics, provide connectivity to virtually any application including all major BI and analytics tools, and meet the demands of low latency, real-time query and analysis with superior throughput and CPU efficiency.

Posted April 22, 2015

SAP SE has announced an Industry 4.0 implementation project with GEA to address condition monitoring and predictive maintenance. GEA, a supplier for the food processing industry and a wide range of process industries, will work with SAP to optimize the performance of its separator and decanter machinery with the SAP Predictive Maintenance and Service solution, cloud edition. Based on SAP HANA Cloud Platform, the solution aims to bring together technology, sensors, and machine data with business processes, applications, and practices.

Posted April 22, 2015

A new release of the HP Haven Big Data Enterprise and OnDemand Platform incorporates advanced analytics and predictive capabilities for enterprises working with large volumes and varieties of information.

Posted April 22, 2015

While the new data stores and other software components are generally open source and incur little or no licensing costs, the architecture of the new stacks grows ever more complex, and this complexity is creating a barrier to adoption for more modestly sized organizations.

Posted April 22, 2015

To help organizations answer questions with data spread across disparate analytics systems and data repositories, Teradata has expanded its QueryGrid technologies. "With this announcement we have our foot on the gas pedal," Imad Birouty, director of product marketing, Teradata. "We have seven updates. We are announcing new connectors that are on their way, announcing that we have delivered on the connectors that we previously announced, and we are refreshing previously-released connector versions of the technologies."

Posted April 20, 2015

Unstructured data types and new database management systems are playing an increasing role in the modern data ecosystem, but structured data in relational database management systems (RDBMS) remains the foundation of the information infrastructure in most companies. In fact, structured data still makes up 75% of data under management for more than two-thirds of organizations, with nearly one-third of organizations not yet actively managing unstructured data at all, according to a new survey commissioned by Dell Software and conducted by Unisphere Research, a division of Information Today, Inc.

Posted April 15, 2015

Percona, a company that makes MySQL and OpenStack faster and reliable for customers, is acquiring Tokutek, allowing Percona to design, service, and support remote management for both MySQL and the ACID-compliant NoSQL database. Tokutek is known for delivering big data processing power across two open source data management platforms, MySQL and MongoDB.

Posted April 14, 2015

AtScale, Inc. has introduced a platform that will enable interactive, multi-dimensional analysis on Hadoop, directly from standard business intelligence tools such as Microsoft Excel, Tableau Software or QlikView. Dubbed the "AtScale Intelligence Platform," the new offering provides a Hadoop-native analysis server that allows users to analyze big data at full scale and top speed, while leveraging the existing BI tools they already own.

Posted April 14, 2015

Informatica, an independent provider of data integration software, has released a new tool aimed at taking a data-centric approach to information security by empowering organizations to identify and visualize sensitive data wherever it resides, inside or outside the corporate perimeter.

Posted April 13, 2015

Think Big, a Teradata company, has introduced the Dashboard Engine for Hadoop, which enables organizations to access and report on big data in Hadoop-based data lakes to make agile business decisions. "There are endless streams of data from web browsers, set top boxes, and contact centers that often land in Hadoop, but sometimes don't make their way into downstream analytics," said Ron Bodkin, president, Think Big.

Posted April 13, 2015

Oracle has unveiled Oracle Data Integrator for Big Data to help make big data integration more accessible and actionable for customers. The goal with the new data integration capabilities is to bring together disparate communities that have emerged within the Oracle client base and allow the mainstream DBAs and ETL developers as well as the big data development organization to be brought together on a single platform for collaboration, said Jeff Pollock, vice president of product management at Oracle.

Posted April 08, 2015

In order to truly appreciate Apache Drill, it is important to understand the history of the projects in this space, as well as the design principles and the goals of its implementation.

Posted April 08, 2015

Translated to an analytical setting, Ockham's principle, also known as Ockham's razor, basically states that analytical models should be as simple as possible, free of any unnecessary complexities and/or assumptions.

Posted April 08, 2015

Tableau Software has announced the general availability of Tableau 9.0, a major release of its visual analytics software. New capabilities in the release help people to achieve more with data, answering questions at interactive speed, according to the vendor.

Posted April 07, 2015

Pages
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102

Sponsors