Newsletters




Hadoop

The Apache Hadoop framework for the processing of data on commodity hardware is at the center of the Big Data picture today. Key solutions and technologies include the Hadoop Distributed File System (HDFS), YARN, MapReduce, Pig, Hive, Security, as well as a growing spectrum of solutions that support Business Intelligence (BI) and Analytics.



Hadoop Articles

At Strata + Hadoop World in New York, Microsoft announced an update to Microsoft Azure HDInsight, its cloud-based distribution of Hadoop. Customers can now process millions of Hadoop events in near real time, with Microsoft's preview of support for Apache Storm clusters in Azure HDInsight. In addition, as part of its integration with the Azure platform, Hortonworks announced that the Hortonworks Data Platform (HDP) has achieved Azure Certification.

Posted October 20, 2014

With Cloudera 5.2 the focus is on building products to deliver on the promise of the enterprise data hub that Cloudera introduced last year, said Clarke Patterson, senior director of product marketing at Cloudera. In particular, new capabilities make the technology more accessible to users who are not data scientists and also increase the level of security, two hurdles which can stand in the way of Hadoop adoption.

Posted October 15, 2014

Generally available today, EMC and Pivotal have announced the Data Lake Hadoop Bundle 2.0 that includes EMC's Data Computing Appliance (DCA), a high-performance big data computing appliance for deployment and scaling of Hadoop and advanced analytics, Isilon scale-out NAS (network attached storage), as well as the Pivotal HD Hadoop distribution and the Pivotal HAWQ parallel SQL query engine. The idea is to provide a turn-key offering that combines compute, analytics and storage for customers building scale-out data lakes for enterprise predictive analytics.

Posted October 14, 2014

Building on its data lake approach, Pivotal today announced the next step in this vision with the implementation of an architecture that builds upon disk-based storage with memory-centric processing frameworks.

Posted October 14, 2014

The new Dell In-Memory Appliance for Cloudera Enterprise is designed to provide customers with a processing engine combined with interactive analytics in a preconfigured and scalable solution, and will begin shipping Oct. 15, 2014.

Posted October 14, 2014

Attunity has introduced Replicate 4.0 which provides high-performance data loading and extraction for Apache Hadoop. The solution has been certified with the Hortonworks and Cloudera Hadoop distributions.

Posted October 14, 2014

Dataguise, a provider of security and data governance solutions for big data, has expanded its DgSecure platform to support Hadoop in the cloud, including full support for Amazon EMR (Elastic MapReduce). Additionally, big data cloud service providers, Altiscale and Qubole, have joined Dataguise's Big Data Protection Partner Program (BDP3) to leverage DgSecure in providing comprehensive discovery, protection and visibility to sensitive data for their cloud-based Hadoop customers.

Posted October 13, 2014

Splunk, which provides software for machine-generated big data analysis, has announced Splunk Enterprise 6.2, Splunk Mint, and Splunk Hunk 6.2. "What we are doing with this release is fundamentally broadening the number of users that can do advanced analytics," stated Shay Mowlem, VP, product marketing at Splunk.

Posted October 13, 2014

IBM is adding new analytics capabilities to the mainframe platform, helping enable better data security and providing clients with the ability to integrate Hadoop big data. By applying analytic tools to business transactions as they are occurring, mainframe systems can enable clients to have true real-time insights. With the analytics on the System z platform, clients can also incorporate social media into their real-time analytic.

Posted October 13, 2014

Teradata is introducing Teradata Loom 2.3, a platform that provides integrated metadata management, data lineage, and ata wrangling for enterprise Hadoop. Teradata has also launched Teradata Cloud for Hadoop, a turnkey, full service cloud environment, and a broad technology and marketing partnership with Cloudera. "Increasingly, customers want a one-stop shop for their data analytics needs," said Chris Twogood, vice president of products and services at Teradata.

Posted October 09, 2014

One feature of the big data revolution is the acknowledgement that a single database management system architecture cannot meet all needs. However, the Lambda Architecture provides a useful pattern for combining multiple big data technologies to achieve multiple enterprise objectives. First proposed by Nathan Marz, it attempts to provide a combination of technologies that together can provide the characteristics of a web-scale system that can satisfy requirements for availability, maintainability, and fault-tolerance.

Posted October 08, 2014

Join DBTA for a webcast on Thursday, October 9, to learn about the key use cases, data replication strategies and methods for exploring data more efficiently through Hadoop.

Posted October 06, 2014

Big analytics and visualization company Datameer is releasing version 5.0 of the company's data analytics application for Hadoop. "Our vision overall is to make data simple and accessible for everyone," said Matt Schumpert, director of product management with Datameer, about version 5.0 improvements.

Posted October 01, 2014

Ron Bodkin founded Think Big Analytics to help organizations gain value from big data. Before that, he was vice president of engineering at Quantcast where he led the data science and engineering teams deploying Hadoop and NoSQL for batch and real-time decision making. In this interview, Bodkin who is CEO of Think Big, now a Teradata company, discusses the challenges organizations face in getting big data projects off the ground and what they need to consider when they embark on projects to leverage data from the Internet of Things and social media.

Posted September 25, 2014

Many in the industry have begun to look to data lakes and Hadoop as the future for data storage. To help shed light on the data lake approach, the pros and cons of this data repository were considered in a recent Unisphere webcast presented by Peter Evans, BI and analytics product evangelist and product technologist consultant, Dell Software; and Elliot King, Unisphere Research analyst.

Posted September 25, 2014

To help customers to meet demanding data challenges, Embarcadero Technologies has announced a major new XE6 release of its portfolio of database tools, including ER/Studio, DBArtisan and Rapid SQL.

Posted September 24, 2014

One of the major issues for companies trying to leverage big data is the length of time it takes for data to be analyzed. While being able to gather and store the data is essential, big data is useless if it cannot be analyzed. As data continues to grow, the processes for moving and analyzing it only become slower and more tedious.

Posted September 23, 2014

MapR Technologies has introduced the 4.0.1 version of the MapR Distribution including Hadoop. The latest MapR release expands real-time capabilities for use cases including operational applications, interactive queries and stream processing with new versions of Apache Drill, Apache Spark, and Apache HBase, while leveraging Hadoop 2.4, including YARN.

Posted September 23, 2014

Broadening the range of tools at its disposal to help customers who are grappling with emerging technologies for leveraging unstructured data, Teradata this week announced its acquisition of Think Big Analytics, a consulting and solutions company that is focused exclusively on Hadoop and open source big data solutions.

Posted September 04, 2014

To help companies evaluate big data systems, a new benchmark has been launched by the Transaction Processing Performance Council (TPC). TPCx-HS is a major achievement for two reasons, noted Raghunath Nambiar, chairman of the TPCx-HS committee, and a distinguished engineer at Cisco. It is the first vendor-neutral benchmark focused on big data systems - which have become a critical part of the enterprise IT ecosystem. TPCx-HS is also the first Express-class benchmark issued by the TPC.

Posted September 03, 2014

Why do tens of thousands of Oracle customers, partners, and consultants descend on Moscone Center in San Francisco for the annual Oracle OpenWorld conference? Despite the proliferation of conferences and online events competing for attention, OpenWorld remains the single, central place for all things Oracle. Here, DBTA presents the annual Who to See @ Oracle OpenWorld special section.

Posted August 27, 2014

With the opportunities and obstacles presented by developments such as big data, cloud, and mobility, the challenges of managing and extracting value from data have never been greater. At the same time, the array of technology options for storing, protecting, integrating, enhancing and analyzing data has exploded. To help add perspective and acknowledge the products and services that have been deemed to provide unique value to customers, DBTA created this new competition. Here are the winners of the 2014 DBTA Readers' Choice Awards.

Posted August 27, 2014

The deluge of big data available for analysis presents great opportunity, but many organizations are having difficulty managing it. This has led to companies not being able to process valuable data, or having to delete sets of data to clear room for the massive amounts of new data.

Posted August 26, 2014

2014 is turning out to be a banner year for Hadoop. The big data giant is continuing to move forward and expand and evolve as big data technology and big data analytics become more mainstream. Here are six key points that demonstrate the advances being made.

Posted August 19, 2014

At the summer 2014 SHARE conference in Pittsburgh, Brian Peterson, who has been president of SHARE for the past two years, spoke about the challenges facing companies today as they deal with emerging IT requirements and the successful approaches for dealing with the mainframe skills shortage. With SHARE's 60th anniversary coming up, Peterson also reflected on the role SHARE has played in the world economy by supporting technology that is critical to industries such as banking, transportation, and manufacturing.

Posted August 18, 2014

Splice Machine, a provider of a Hadoop RDBMS, has added a $3 million extension to its Series B financing provided by venture capital firm Correlation Ventures, along with database industry leaders Roger J. Sippl and Roger Bamford. The new funding adds to the initial $15 million Series B investment from Interwest Partners and Mohr Davidow Ventures (MDV), bringing the total round to $18 million.

Posted August 12, 2014

Actian has started the Big Data 2.0 Clear Path Program to help companies jumpstart analytics projects and move existing ones into production more rapidly. Developed by data scientists in Actian's Analytics Center of Excellence (COE), the Big Data 2.0 Clear Path Program applies big data concepts for faster time to value, higher accuracy and deeper discovery, leveraging the Actian Analytics Platform.

Posted August 05, 2014

X15 Enterprise, which provides an end-to-end solution for ingesting, indexing, searching and analyzing machine data, is now generally available from X15 Software. According to the vendor, X15 Enterprise allows machine data to remain in HDFS and does not require it to be duplicated into proprietary storage before it can be indexed and analyzed.

Posted August 05, 2014

As demand for IT services and data volumes grow, so do the challenges with managing databases. For Oracle's vast and growing network of independent software vendors, value-added resellers, systems integrators, and consultants, these are fast-changing times. With Oracle OpenWorld rapidly approaching, DBTA examines Oracle's current product line-up and evolving ecosystem.

Posted August 05, 2014

The converging forces of open source, with its rapid crowd-sourced innovation, cloud, with its unlimited capacity and on-demand deployment options, and NoSQL database technologies, with their ability to handle unstructured data, are helping companies address the new challenges and opportunities presented by big data. Here are the winners of the 2014 DBTA Readers' Choice Awards for Best Big Data Solution.

Posted August 04, 2014

The staggering variety of data—with much of it unstructured, including business documents, presentations, emails, log files and social media data—means that this data does not fit neatly into the rows and columns of relational database management systems. But Hadoop, which is an open source technology for storing and processing data that runs on industry-standard hardware, embraces this mixed bag of data types and enables companies to store and analyze data sets with no limits in size. Here are the winners of the 2014 DBTA Readers' Choice Awards for Best Hadoop Solution.

Posted August 04, 2014

Novetta Identity Analytics has achieved Hortonworks Certification, and Novetta has joined the Hortonworks Technology Partner Program. Novetta Identity Analytics provides a central, multi-dimensional view of the entities across the data silos and uncovers the relationships within and among those entities to support customer intimacy, churn prediction, risk profiling, fraud analysis and detection, and other use cases.

Posted August 04, 2014

Tata Consultancy Services (TCS), an IT services, consulting and business solutions organization, has formed a new partnership with MapR Technologies. The two companies are developing turn-key solutions to address big data challenges.

Posted July 29, 2014

Databricks and SAP have collaborated on a Databricks-certified Apache Spark distribution offering for the SAP HANA platform. This production-ready distribution offering is the first result of a new partnership between Databricks and SAP.

Posted July 23, 2014

In "Big Data at Work," Tom Davenport explains to readers why big data is important to them and their organizations, what technology they actually need to manage it, and where to start capitalizing on its potential. Here, the author shares an excerpt from his recent book.

Posted July 23, 2014

Helping companies get more from big, unstructured data, Actian has unveiled the Actian Analytics Platform - Hadoop SQL Edition. In addition, as a result of a partnership with Logi Analytics, the Actian Analytics Platform combines with Logi Info to offer customers a comprehensive analytics platform to connect, visualize, analyze and act on big data.

Posted July 22, 2014

InfiniDB has released version 4.6 of its fourth generation columnar data platform, which includes enhanced support for large-scale join operations and support for additional data load commands to speed the extraction, transformation, and load (ETL) process. The company also introduced a new 60-day enterprise trial that includes InfiniDB Enterprise Manager, a management console to simplify the administration of InfiniDB, and provide real-time visibility into the performance and availability of users' InfiniDB servers.

Posted July 15, 2014

Oracle today introduced Oracle Big Data SQL, which allows customers to run one SQL query across Hadoop, NoSQL, and Oracle Database, minimizing data movement while overcoming data silos. According to Oracle the new solution helps customers gain a competitive advantage by making it easier to uncover insights faster, and allows them to leverage existing SQL skills while protecting data security and enforcing governance. Oracle Big Data SQL runs on Oracle Big Data Appliance and can work in conjunction with Oracle Exadata Database Machine.

Posted July 15, 2014

Pages
1
2
3

Sponsors