Newsletters




Hadoop

The Apache Hadoop framework for the processing of data on commodity hardware is at the center of the Big Data picture today. Key solutions and technologies include the Hadoop Distributed File System (HDFS), YARN, MapReduce, Pig, Hive, Security, as well as a growing spectrum of solutions that support Business Intelligence (BI) and Analytics.



Hadoop Articles

Currently, the IT industry is the midst of a major transition as it moves from the last generation - the internet generation - to the new generation of cloud and big data, said Andy Mendelsohn, Oracle's EVP of Database Server Technologies, who recently talked with DBTA about database products that Oracle is bringing to market to support customers' cloud initiatives. "Oracle has been around a long time. This is not the first big transition we have gone through," said Mendelsohn.

Posted February 05, 2016

Oracle has introduced a new Big Data Preparation Cloud Service. Despite the increasing talk about the need for companies to become "data-driven," and the perception that people who work with business data spend most of their time on analytics, Oracle contends that in reality many organizations devote much more time and effort on importing, profiling, cleansing, repairing, standardizing, and enriching their data.

Posted February 04, 2016

Hewlett Packard Enterprise (HPE) has announced the availability of HPE Investigative Analytics, a new software solution to enable financial institutions and other organizations in highly regulated industries to use big data technologies to detect patterns, relationships, behaviors, and anomalies across structured and unstructured data stores. The software is aimed at helping companies reduce risk by proactively preventing fraudulent actions.

Posted February 02, 2016

The combination of location data with high speed data in VoltDB version 6.0 will allow VoltDB to better support businesses that need to maximize the value of real-time data streams including location data to make personalized offers to customers and improve decision making.

Posted January 28, 2016

The Winter '15 release allows administrators to deploy Paxata in heterogeneous environments including the Hortonworks Data Platform on YARN and with multiple versions of Apache Spark. The latest release also improves the way business analysts find, access, and apply data by delivering additional ease of use capabilities supported by machine learning innovations, and provides enterprise-grade security and a multi-tenant governance model.

Posted January 27, 2016

Data Summit, a comprehensive educational experience designed to guide attendees through the key issues in data management and analysis, is coming to the New York Hilton Midtown. Data Summit also features two co-located events: Hadoop Day and Virtualization Day. In addition, the IOUG will participate in Data Summit again this year presenting a track focused on big data in the cloud and the evolution of the data warehouse.

Posted January 26, 2016

Splice Machine, which provides a database powered by Hadoop and Spark, has added $9 million in new funding to accelerate product, sales and marketing efforts, bringing the company's total funding since its inception to approximately $31 million.

Posted January 26, 2016

ClearStory Data is making advancements and core improvements in the upcoming release of its native Apache Spark platform. With Apache Spark 1.6, ClearStory further speeds exploration on big, diverse data when business users need unrestricted data discovery and free-form exploration to answer new questions.

Posted January 26, 2016

For decades, the enterprise data warehouse (EDW) has been the aspirational analytic system for just about every organization. It has taken many forms throughout the enterprise, but all share the same core concepts of integration/consolidation of data from disparate sources, governing that data to provide reliability and trust, and enabling reporting and analytics. The last few years, however, have been very disruptive to the data management landscape. The "big data" era has introduced new technologies and techniques that provide alternatives to the traditional EDW approach, and in many cases, exceeding its capabilities. Many claim we are now in a post-EDW era and the concept itself is legacy.

Posted January 19, 2016

The year 2015 started out with people recognizing that the Hadoop ecosystem is here to stay, and ended as the year in which organizations achieved real success within the Hadoop ecosystem. Today, more projects are popping up within the Hadoop ecosystem that can run both with and without Hadoop. The great thing about this trend is that it lowers the barrier to entry for people to get started with these technologies. More importantly, all of these new technologies work best at large scale within the rest of the Hadoop ecosystem, while Hadoop MapReduce has begun its ride off into the sunset.

Posted January 19, 2016

Rocana has unveiled the latest version of its solution for managing and analyzing event-oriented machine data that introduces new advanced analytics and anomaly detection abilities. In addition to the new features, the platform update also introduces support for Hortonworks along with Cloudera, further deepening the platform's reach.

Posted January 12, 2016

Join IT practitioners and business stakeholders alike for the third annual Data Summit conference at the New York Hilton, May 9-11, 2016. Discounted pricing is available for a limited time.

Posted January 08, 2016

Though MySQL has evolved into a robust database engine that now handles large-scale operations for mission-critical workloads, the complexity of performance tuning can still prove challenging for DBAs who are increasingly being tasked with managing multiple database platforms in production for larger and complex workloads.

Posted January 06, 2016

Traditionally, the cardinal rule has been to model data first and load it later. But with new technologies and repositories such as Hadoop, NoSQL, and data lakes, and big data itself, the rule is being flipped to load first and model later. And, with SQL remaining an effective and widely embraced query language, companies have to balance working with traditional methods against the need for some of the newer methods as well.

Posted January 06, 2016

Servers.com, provider of cloud servers and cloud storage, is collaborating with Percona to deliver automated deployment of the Percona Server for MySQL database.

Posted December 22, 2015

In 2015, big data, mobility, IoT, expanding requirements for security and real-time analytics and the introduction of the Cognitive Era continued to place greater pressure on IT organizations. Linux and open source technologies are at the center of many of the most innovative software and hardware solutions that are addressing emerging enterprise requirements. Here's a look back at some of the most significant announcements in Linux and open source technology of 2015.

Posted December 16, 2015

Looking ahead to 2016 David Jonker, senior director of big data at SAP, offered six key business and technology trends that will take the market by storm, including a continued interest in in-memory processing, enablement of real-time data and analytics, and an increased focus on advanced analytics and machine learning.

Posted December 16, 2015

It is often said that the only constant is change. For data executives and professionals, the coming year will only bring a lot more of it. Developments as diverse as cloud, big data, real time, NoSQL, analytics, and the Internet of Things (IoT) will continue to reshape enterprise data operations and opportunities as we know them. Here are 16 trends that will shape the enterprise data landscape in 2016.

Posted December 16, 2015

What's ahead for 2016 in terms of cloud, IoT, big data, analytics, and open source technologies? IT executives gaze into their crystal balls, and weigh in on the upcoming challenges and opportunities ahead in the next year - and beyond.

Posted December 16, 2015

Dataguise, provider of data-centric security solutions for detecting and protecting confidential information, is now securing relational databases such as Oracle and SQL Server, as well as Teradata data warehouses, deployed on the Microsoft Azure cloud infrastructure. Now Azure users can utilize big data and cloud technologies for improved business intelligence and performance.

Posted December 16, 2015

SnapLogic is receiving $37.5 million of series E financing, fueling the company's expansion into the world of big data. The world's biggest technology company and private equity firms join other top tier investors, including Andreessen Horowitz, Ignition Partners and Triangle Peak Partners to provide Snaplogic with the finacing which will be used for accelerating SnapLogic's expansion with further investments in technology and extending field operations to give the company a broader reach in sales and marketing to continue growing internationally.

Posted December 15, 2015

Hortonworks revealed upcoming advancements of Hortonworks Data Platform (HDP) with the in-memory analytic capabilities of Spark. Hortonworks has also launched Hortonworks Community Connection (HCC), a new online collaboration destination for developers, DevOps, customers and partners to get answers to questions, collaborate on technical articles and share code examples from GitHub.

Posted December 10, 2015

Harnessing operational big data does not come with a "one size fits all" solution. Organizations are at different points in their data management cycles, and whether that they are building new applications or optimizing existing ones, each needs a unique solution. This was among the key points made during a special DBTA roundtable webinar on harnessing operational big data. The webinar featured Matt Allen, senior product marketing manager with MarkLogic, Kevin Petrie, senior director with Attunity, and Jason Paul Kazarian, senior architect with Hewlett Packard Enterprise (HPE).

Posted December 07, 2015

It's commonly asserted—and generally accepted—that the era of the "one-size-fits-all" database is over. We expect that enterprises will use a combination of database technologies to meet the distinct needs created by various application architectures.

Posted December 02, 2015

As 2015 draws to a close it's time to look back on the year's big changes in data management and reflect on some of the most insightful observations by leading data experts. Here, Big Data Quarterly presents a round-up of thought-provoking articles that explore some of the key advances of 2015, how organizations are harnessing what big data has to offer, and the challenges they face as they seek to compete on analytics.

Posted December 02, 2015

Datameer, a provider of a big data analytics platform, has announced a range of product enhancements, including the availability of Datameer Enterprise in the Microsoft Azure Marketplace via Microsoft HDInsight, a new Spark connector, and native support for Amazon Redshift and updated support for leading Hadoop distributions.

Posted December 01, 2015

More and more facets of data management are moving towards the cloud. To address this trend, IBM's BigInsights, a Hadoop as a service offering, gives organizations the opportunity to deploy Hadoop in the cloud.

Posted December 01, 2015

Vormetric is joining the Hortonworks Technology partner program, promising users a comprehensive enterprise solution that is made for big data environments and optimizes performance. The advantage for the customers is that they will be able to use the Vormetric solution across their entire infrastructure, said Ashvin Kamaraju, vice president of product development and partner management at Vormetric. "What our customers like is for us to be the provider of all the data security solutions across their entire enterprise."

Posted December 01, 2015

Hadoop distribution provider Cloudera has introduced Cloudera Enterprise 5.5, including Cloudera Navigator Optimizer, a new product targeted at helping organizations improve big data workload performance and efficiency. Cloudera Navigator Optimizer, now in beta, is expected to be generally available in 2016. The new release of Cloudera Enterprise has three main areas of focus, according to Anupam Singh, head of data management at Cloudera.

Posted November 19, 2015

Almost every commercial endeavor and, indeed, almost every human undertaking, has software at its core. Yet, with software at the core of so much of our society, it's surprising to realize it's getting harder and harder to actually make a living selling software. In his recent book, "The Software Paradox," Stephen O'Grady - co-founder of analyst firm RedMonk - provides a cohesive and persuasive analysis of what those of us in the software business have been experiencing for several years - it's getting increasingly difficult to generate revenues selling "shrink-wrapped" software.

Posted November 18, 2015

Clearlake Capital Group, L.P. has completed the acquisition of Syncsort. "Our organic product investments and related software acquisitions over the last couple of years have been singularly focused on one of the most important secular growth opportunities in the technology industry today: demand for software that liberates data and existing budgets for use in advanced machine learning and business intelligence systems," said Lonne Jaffe, CEO of Syncsort.

Posted November 18, 2015

With the Hadoop ecosystem expanding rapidly, Hortonworks' Shaun Connolly recently discussed how Hadoop and related technologies are growing and being used, as well as the factors for a successful deployment.

Posted November 13, 2015

Franz, Inc., an early innovator in artificial intelligence (AI) and a supplier of semantic graph database technology, is releasing Gruff v6.0, the company's graph visualization software for exploring and discovering connections within data. Gruff provides novice users and graph experts with the ability to visually build queries and visualize connections between data without writing code, which speeds discovery and enhances the ability to uncover hidden connections within data.

Posted November 10, 2015

The concept of the data lake has become a hot topic. The data lake retains data in its original format to allow the data to be more flexible for everyone involved. While this sounds fine in theory, it is a more complicated in practice due to the need for governance and security.

Posted November 09, 2015

To better manage the data explosion now and with scalable options for the future, existing data architecture is evolving beyond traditional databases, data stores, data warehouses, and the like into a more unfiltered repository known as the data lake.

Posted November 05, 2015

Following its separation from Hewlett-Packard Company, Hewlett Packard Enterprise made its debut today as an enterprise technology provider with $53 billion in annual revenue and a comprehensive product portfolio. HP announced plans to separate into two new publicly traded companies a year ago with one including HP's enterprise technology infrastructure, software and services businesses, doing business as Hewlett-Packard Enterprise, and the other comprising HP's personal systems and printing businesses, operating as HP Inc. and retain the current logo.

Posted November 02, 2015

There's unrelenting pressure on businesses to compete on analytics and to be able to anticipate customer needs and trends ahead of the curve. Enterprises are looking to expand BI and analytics capabilities as far and wide as technologies and budgets will allow them to go. As a result, the continuing advance of analytic capabilities across the enterprise has reached a "tipping point."

Posted October 28, 2015

Oracle has announced new services - Oracle Big Data Preparation Cloud Service, Oracle GoldenGate Cloud Service, Oracle Big Data Discovery Cloud Service and Oracle NoSQL Database Cloud Service - that complement Oracle's Big Data Cloud Service and Big Data SQL Cloud Service, to form a comprehensive portfolio of solutions for big data in the cloud.

Posted October 28, 2015

AtScale, a company specializing in connecting business users to their data in Hadoop, is releasing a new version of its platform, introducing a new tool that will enhance performance of big data on Hadoop.

Posted October 21, 2015

MapR is including Apache Drill 1.2 in its Apache Hadoop distribution and is also now offering a new Data Exploration Quick Start Solution, leveraging Drill to help customers get started more rapidly with big data projects. Apache Drill is an open source, low-latency query engine for Hadoop that delivers secure, interactive SQL analytics at petabyte scale. With the two announcements, MapR says customers and partners will able to more quickly leverage Drill to get fast business insights from all their data in Hadoop and other sources. MapR also released a comprehensive SQL-based test framework to the open source community.

Posted October 21, 2015

Teradata is introducing a series of platforms that will capture streaming IoT data, land that data in Hadoop, and use advanced Aster machine learning natively in Hadoop to analyze data.

Posted October 20, 2015

Pages
1
2
3
4
5
6
7
8
9
10

Sponsors