Newsletters




Hadoop

The Apache Hadoop framework for the processing of data on commodity hardware is at the center of the Big Data picture today. Key solutions and technologies include the Hadoop Distributed File System (HDFS), YARN, MapReduce, Pig, Hive, Security, as well as a growing spectrum of solutions that support Business Intelligence (BI) and Analytics.



Hadoop Articles

IBM says it is making it easier and faster for organizations to access and analyze data in-place on the IBM z Systems mainframe with a new z/OS Platform for Apache Spark. The platform enables Spark to run natively on the z/OS mainframe operating system.

Posted April 04, 2016

Databricks, the company behind Apache Spark, is releasing a new set of APIs that will enable enterprises to automate their Spark infrastructure to accelerate the deployment of production data-driven applications.

Posted April 01, 2016

ManageEngine is introducing a new application performance monitoring solution, enabling IT operations teams in enterprises to gain operational intelligence into big data platforms. Applications Manager enables performance monitoring of Hadoop clusters to minimize downtime and performance degradation. Additionally, the platform's monitoring support for Oracle Coherence provides insights into the health and performance of Coherence clusters and facilitates troubleshooting of issues.

Posted April 01, 2016

It's become almost a standard career path in Silicon Valley: A talented engineer creates a valuable open source software commodity inside of a larger organization, then leaves that company to create a new startup to commercialize the open source product. Indeed, this is virtually the plot line for the hilarious HBO comedy series, Silicon Valley. Jay Krepes, a well-known engineer at LinkedIn and creator of the NoSQL database system, Voldemort, has such a story.

Posted March 31, 2016

With well over a hundred open source projects now part of the Hadoop ecosystem, it can be hard to know which technologies are best for which requirements. To help users get started with Hadoop and understand their technology choices, James Casaletto will present "Harnessing the Hadoop Ecosystem" at Data Summit 2016 in NYC. Casaletto is a solutions architect for MapR, where he develops and deploys big data solutions with Apache Hadoop.

Posted March 31, 2016

What can you learn from the structure of an email and what really constitutes as a "good" post? What kinds of data can you grab to create the best marketing campaign? Matt Laudato will address those questions during his presentation, titled "Supercharging Your Marketing with Big Data," at Data Summit 2016 in NYC.

Posted March 31, 2016

Denodo, a provider of data virtualization software, is releasing Denodo Platform 6.0, further accelerating its "fast data" strategy. "It's a major release for us," said Ravi Shankar, Denodo CMO. There are three important areas that nobody else is focusing on in the industry, he noted. "This, we hope, will change how data virtualization, and in a broader sense, data integration will shape up this year."

Posted March 31, 2016

NoSQL databases were born out of the need to scale transactional persistence stores more efficiently. In a world where the relational database management system (RDBMS) was king, this was easier said than done.

Posted March 29, 2016

MapR is now available as part of Bigstep's big data platform-as-a-service, supporting a wide range of Hadoop applications.

Posted March 29, 2016

Reltio is releasing an enhanced version of Reltio Cloud 2016.1, adding new analytics integration, collaboration, and recommendation capabilities to help companies be right faster.

Posted March 29, 2016

Teradata has introduced a new "design pattern" approach for data lake deployment. The company says its concept of a data lake pattern leverages IP from its client engagements, as well as services and technology to help organizations more quickly and securely get to successful data lake deployment.

Posted March 28, 2016

The data lake has been the subject of more than its fair share of critics since its inception. Pundits claim it's a source of chaos and risk. Analysts often slam the concept, calling it a "data swamp" or "data dump." As a result of this scrutiny, the definition and understanding of the definition of the data lake are rather murky.

Posted March 24, 2016

The rise of big data technologies in enterprise IT is now seen as an inevitability, but adoption has occurred at a slower pace than expected, according to Joe Caserta, president and CEO of Caserta Concepts, a firm focused on big data strategy consulting and technology implementation. Caserta recently discussed the trends in big data projects, the technologies that offer key advantages now, and why he thinks big data is reaching a turning point.

Posted March 23, 2016

Informatica has launched an end-to-end solution to help customers gain greater insight from big data.

Posted March 23, 2016

SAP SE's newest in memory query engine, SAP HANA Vora, is now generally available, equipping enterprises with contextual analytics across all data stored in Hadoop, enterprise systems, and other distributed data sources.

Posted March 23, 2016

To spread the word on the best information management solutions in the marketplace, Database Trends and Applications has launched the 2016 DBTA Readers' Choice Awards, a program in which the winners will be selected by the experts whose opinions count above all others - you. Time is running out so make your nomination now.

Posted March 23, 2016

As more businesses leverage applications that are hosted in the cloud, the lines between corporate networks and the internet become blurred. Accordingly, enterprises need to develop an effective strategy for ensuring security. The problem is, many of today's most common approaches simply don't work in this new cloud-based environment.

Posted March 23, 2016

Data discovery should drive BI in order to support data-driven competencies for evolving data-centric organizations. However, organizations often struggle with bottlenecks and excessive friction while retrieving meaningful insights.

Posted March 22, 2016

The OAUG volunteers planning COLLABORATE 16: Technology and Applications Forum for the Oracle Community (April 10-14 at Mandalay Bay in Las Vegas) are themselves Oracle users and technologists, understanding innately the myriad options and challenges faced by the wider user community in a period of rapid change and transformation. With participation and contributions from all corners of the user community, COLLABORATE offers the information and perspective to make sense of it all.

Posted March 21, 2016

Oracle has released a free and open API and developer kit for its Data Analytics Accelerator (DAX) in SPARC processors through its Software in Silicon Developer Program. "Through our Software in Silicon Developer Program, developers can now apply our DAX technology to a broad spectrum of previously unsolvable challenges in the analytics space because we have integrated data analytics acceleration into processors, enabling unprecedented data scan rates of up to 170 billion rows per second," said John Fowler, executive vice president of Systems, Oracle.

Posted March 16, 2016

Available now, Talend says its Integration Cloud Spring '16 release adds enhancements to help IT organizations execute big data and data integration projects running on AWS Redshift or AWS Elastic MapReduce (EMR) with greater ease - using fewer resources, and at a reduced cost.

Posted March 16, 2016

Analytics and the platforms that support big data are constantly evolving, being shaped by the need to deliver data faster to users and gain effective insights throughout the organization. Hadoop, Spark, Kafka, and the cloud are some of the technologies that can handle the demand the future will bring, according to Kevin Petrie, senior director at Attunity.

Posted March 14, 2016

Attivio is receiving $31 million in investment financing that will help expand the company as it accelerates its offerings into the big data market.

Posted March 09, 2016

IDERA, a provider of database lifecycle management solutions, is extending its product portfolio by adding Embarcadero Technologies' ER/Studio and DB PowerStudio tools, allowing organizations to rely on a single vendor for all their database lifecycle needs.

Posted March 09, 2016

In a new book titled "Next Generation Databases," Guy Harrison, an executive director of R&D at Dell, shares what every data professional needs to know about the future of databases in a world of NoSQL and big data.

Posted March 08, 2016

As more and more data comes into the enterprise, companies are looking to build real-time big data architectures to keep up with an increased amount of information.

Posted March 07, 2016

Syncsort is introducing new capabilities to its data integration software, DMX-h, that allow organizations to work with mainframe data in Hadoop or Spark in its native format, which is necessary for preserving data lineage and maintaining compliance.

Posted March 07, 2016

Infobright, the columnar database analytics platform, has unveiled its new Infobright Approximate Query (IAQ) solution for large-scale data environments, allowing users to gain insights faster and efficiently. "This technology is being delivered on the basis of rethinking the business problem and using technology in a very meaningful way to solve problems that would otherwise be unsolvable using a traditional approach," said Don DeLoach, CEO.

Posted February 26, 2016

SAP SE is introducing new predictive capabilities within its platforms with the release of SAP HANA Cloud Platform predictive services 1.0 and SAP Predictive Analytics 2.5.

Posted February 24, 2016

The promise of the data lake is an enduring repository of raw data that can be accessed now and in the future for different purposes. To help companies on their journey to the data lake, Information Builders has unveiled the iWay Hadoop Data Manager, a new solution that provides an interface to generate portable, reusable code for data integration tasks in Hadoop.

Posted February 23, 2016

It is hard to think of a technology that is more identified with the rise of big data than Hadoop. Since its creation, the framework for distributed processing of massive datasets on commodity hardware has had a transformative effect on the way data is collected, managed, and analyzed - and also grown well beyond its initial scope through a related ecosystem of open source projects. With 2016 recognized as the 10-year anniversary for Hadoop, Big Data Quarterly chose this time to ask technologists, consultants, and researchers to reflect on what has been achieved in the last decade, and what's ahead on the horizon.

Posted February 18, 2016

Currently, the IT industry is the midst of a major transition as it moves from the last generation - the internet generation - to the new generation of cloud and big data, said Andy Mendelsohn, Oracle's EVP of Database Server Technologies, who recently talked with DBTA about database products that Oracle is bringing to market to support customers' cloud initiatives. "Oracle has been around a long time. This is not the first big transition we have gone through," said Mendelsohn.

Posted February 17, 2016

Over the last half decade, we've watched SQL purists butt heads with NoSQL upstarts, Hadoop triumphalists clash with Hadump pessimists, database geeks war with application developers, and so on. In the midst of all this warring, we've tried to fit—and, in many cases, to cram—the new into the old, the old into the new, with the result that at one time or another, we've asked the impossible of all of the components in our ever-expanding technology portfolios.

Posted February 16, 2016

Enterprises can't seem to pack enough big data and analytics solutions into their data centers, executive suites, and everywhere else across their organizations. Just about every venture-capital-cash-fueled startup from Silicon Valley to Boston has an analytics component to it. As these firms rapidly gain traction, they are being scarfed up by larger vendors looking to solidify their leadership of the analytics space.

Posted February 16, 2016

Hewlett Packard Enterprise (HPE) has selected RedPoint Data Management platform as the underlying platform for a new HPE Risk Data Aggregation and Reporting (RDAR) integrated solution to support financial institutions' compliance with BDBS 239.

Posted February 16, 2016

Addressing the shift toward business-user-oriented visual interactive data preparation, Trillium Software has launched a new solution that integrates self-service data preparation with data quality capabilities to improve big data analytics.

Posted February 16, 2016

The Data Summit 2016 conference schedule of sessions and speakers is now available online. Presented by Database Trends and Applications and Big Data Quarterly, the 2-day event (Tuesday, May 10, and Wednesday, May 11) will bring together IT managers, data architects, application developers, data analysts, project managers, and business managers for a deep dive into the key technologies and strategies for becoming a data-informed business.

Posted February 12, 2016

Say what you will about Oracle, it certainly can't be accused of failing to move with the times. Typically, Oracle comes late to a technology party but arrives dressed to kill.

Posted February 10, 2016

Oracle has introduced a new Big Data Preparation Cloud Service. Despite the increasing talk about the need for companies to become "data-driven," and the perception that people who work with business data spend most of their time on analytics, Oracle contends that in reality many organizations devote much more time and effort on importing, profiling, cleansing, repairing, standardizing, and enriching their data.

Posted February 10, 2016

Looker, provider of a BI platform, has added support for Presto and Spark SQL as well as updates to its support for Impala and Hive.

Posted February 09, 2016

Trifacta, provider of a self-service data preparation platform, is receiving $35 million in growth-stage financing that will be used to continue expanding the company globally and allow for additional projects. "The multi-billion-dollar big data and IoT revolution requires a modern, innovative approach to preparing data and empowering end users," said Ping Li, partner at Accel and director of the company's Big Data Fund. "Trifacta delivers that platform. We've backed Trifacta's incredible team from the beginning and are excited to help the company scale to take advantage of this significant market opportunity."

Posted February 09, 2016

Franz Inc. is releasing an updated version of its AllegroGraph platform along with receiving certification on the latest release of Cloudera Enterprise through the Cloudera Certified Technology Program (CCTP).

Posted February 08, 2016

EMC has begun shipping the latest version of EMC Elastic Cloud Storage (ECS), a multi-purpose, shared global storage that scales into Exabytes to store both small and large files. ECS was first launched in spring 2014 to provide high-density storage with exabyte scalability in a private cloud.

Posted February 08, 2016

Pages
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21

Sponsors