Big Data Quarterly Articles



Splice Machine has released version 1.5 of its Hadoop RDBMS, which adds multiple enterprise-ready features. The new release adds functionality and performance improvements to enable companies to increase the benefits of using Splice Machine to support real-time applications, run operational data lakes, and accelerate their ETL pipelines.

Posted October 20, 2015

In the future, the clear winners in the IoT Revolution will be those companies that not only embrace the Internet of Things, but use it to transform their businesses. Those winning companies will be the ones that integrate IoT into their operations, products, and customer interactions to create new business models and sources of value. In fact, McKinsey estimates there could be as much as $11 trillion per year by 2020 in new economic value from adopting IoT.

Posted October 13, 2015

We live in an "as-it-happens" world, and consumers expect on-demand everything—TV, taxi service, you name it. That same culture is now (and rightly so) expected in businesses. Don't just think about "good enough"—think and plan for real time. The technology is here now to leverage real time in your business for shorter feedback loops, improved time-to-market, and happier customers. The sooner, the better. It's never too early for real time. "Good enough" simply isn't, well, good enough anymore.

Posted October 13, 2015

MapR Technologies has added native JSON support to the MapR-DB NoSQL database. The in-Hadoop document database will allow developers to quickly deliver scalable applications that also leverage continuous analytics on real-time data. A developer preview of MapR-DB with sample code is available for download and general availability of these new capabilities in MapR-DB will be available in Q4 2015.

Posted October 07, 2015

One of the noticeable changes this year at Strata + Hadoop World 2015 was the rise of Apache Spark, an engine for large scale data processing. In recent months, many companies have extended support to Spark, which can be complementary to Hadoop, but can also be deployed without it.

Posted October 05, 2015

MarkLogic, which bills itself as the only enterprise NoSQL database provider, completed a $102 million financing round earlier this year that it will use to accelerate the pace of growth in the $36 billion operational database market. Recently, Big Data Quarterly spoke with Joe Pasqua, executive vice president of products at MarkLogic, about the changing database management market, and what MarkLogic is doing to meet emerging enterprise customer requirements.

Posted September 24, 2015

StreamSets Inc., a company that aims to speed access to enterprise big data, has closed a $12.5 million round of Series A funding. The single biggest barrier to a successful enterprise analytics platform is the effective and efficient ingest of data, the company says.

Posted September 24, 2015

Anyone who thought Hadoop was a fly-by-night technology was wrong. Hadoop has rapidly evolved—improving and gaining mainstream adoption as a technology and framework for enabling data applications previously out of reach for all but the savviest of companies. The open source Apache Hadoop developer community (and distribution vendors) continuously contributes advances to meet the demands of companies seeking more powerful—and useful—data applications, while also focusing on requirements for improved data management, security, metadata, and governance. Hadoop is not only stable but worthy of consideration for core IT strategies.

Posted September 14, 2015

By now, we all know the value of big data analytics. In the 1990s, data mining and business intelligence (BI) efforts were used mostly for after-the-fact reporting and problem solving. The need for these capabilities will always remain, and high levels of confidence and investment in these disciplines will continue. However, the current focus on big data analytics emphasizes predicting and influencing the future, based on real-time analysis of 100% of enterprise data.

Posted September 14, 2015

It's no secret that the world of information technology is changing fast. Data is being created in ways not possible a few years ago. It is now feasible to collect and analyze data from a wide range of sources, including mobile devices, machines, social media, documents, and emails. To help organizations navigate the rapidly changing big data landscape, Big Data Quarterly presents the "Big Data 50," a list of companies driving innovation.

Posted September 09, 2015

IBM reports that nearly 15 petabytes of data are created every day; eight times greater than the amount of information stored in all of the libraries in the U.S. To keep up, enterprises are deploying web-scale architectures that enable virtualization, compute and storage functionality on a vast scale.

Posted September 08, 2015

Databricks is enhancing its cloud-based platform to strengthen its security, manageability and ease of application development. According to the vendor, the new features securely manage data access for large teams while streamlining Spark application development and deployment for enterprises dealing with complex and fast-paced environments.

Posted September 02, 2015

Hadoop distribution provider Cloudera has released Cloudera Director 1.5, an integrated solution for deploying and managing enterprise-grade Hadoop in cloud environments that enables deployment flexibility through an open API. Google Cloud Platform (GCP) is the first to integrate with the API.

Posted August 12, 2015

One of the side effects of the growth in data is the vast variety that data now comes in. While it presents advantages, the newer data types and sources also make data integration more complex. To highlight new technologies and techniques for overcoming data silos and enabling the delivery of information across the enterprise - where and when it is needed - Database Trends and Applications recently hosted a special roundtable webinar sponsored by Attunity, MarkLogic, and SnapLogic. The webinar covered the key big data integration issues facing enterprises and presented real-world solutions and best practices.

Posted August 10, 2015

Faced suddenly with the challenge of quickly extracting valuable consumer data from a massive file for customer acquisition programs, DX Marketing selected Oracle Database Cloud Service and Oracle Marketing Cloud Service to replace its on-premise computing platform.

Posted August 07, 2015

Addressing the need for more big data skills training, Apache Hadoop distribution provider MapR Technologies has announced availability of a complete Apache HBase design and development curriculum on its free Hadoop On-Demand Training program.

Posted August 06, 2015

Unless you have been trapped under an elephant, you have heard a lot about big data. Today, most IT teams are either contemplating a Hadoop strategy or are well under way in the hopes of making data actionable.

Posted August 06, 2015

A typical organization loses 5% of its revenues to fraud each year, and in the U.S. alone the total cost of (non-health) insurance fraud is estimated at more than $40 billion per year, according to Bart Baesens, one of the authors of a new book on data science techniques for fraud detection. While opportunities for fraud are continuously evolving, fraud detection always has the biggest impact if it is done quickly. Fortunately, says Baesens, big data analytics can help to make early fraud detection possible.

Posted August 05, 2015

Predixion, a developer of cloud-based advanced analytics software, is partnering with PLAT.One to deliver predictive insights for IoT applications.

Posted July 30, 2015

With the continued increase in data over the recent years, data infrastructures are changing as well. The one-size-fits-all approach for data environments is going by the wayside due to the expanding range of data types. With the growing understanding that the world of big data is heterogeneous, businesses are now employing versatile hybrid data architectures. But while they provide the right tools for the myriad requirements of big data, hybrid architectures must be properly integrated and managed to deliver their full value.

Posted July 28, 2015

Continuing its global expansion, CoolaData, which provides a solution that tracks user behavior across all channels and aggregates the data to help monetize the insights, has opened an office in New York. The company's investors include 83North (formerly Greylock IL) and Carmel Ventures.

Posted July 22, 2015

The Hortonworks Data Platform (HDP) version 2.3 is now available, with enhancements for user experience for both operators and developers, new security and data governance capabilities, a new cluster monitoring service for support subscription customers.

Posted July 22, 2015

Franz, a supplier of semantic graph database technology, has added patented N-dimensional analysis capabilities to its flagship product AllegroGraph.

Posted July 20, 2015

Kyvos Insights, a big data analytics company, emerged from stealth mode to introduce a solution called Kyvos that it says provides insights from all corporate data, regardless of size and granularity.

Posted July 13, 2015

Responding to a growing need to deliver more data to the right people at the right time, Cloudera and Teradata have announced the Teradata Appliance for Hadoop with Cloudera—the enterprise-ready Hadoop distribution.

Posted July 09, 2015

Graph databases are especially good for managing data about the connections among resources, and they are here to stay as a vital mechanism for managing data. In some enterprises, graph databases are even replacing relational databases as the primary database of record.

Posted July 08, 2015

These days, everybody talks about big data and analytics but does anybody really know how to tackle it? Everybody thinks the others are doing it and hence also claim to do it as well, but it's likely that only the fortunate have had (positive) experience(s) with it so far.

Posted June 26, 2015

Describing it as potentially the most important new open source project in a decade, IBM announced a major commitment to Apache Spark.

Posted June 15, 2015

Data is only going to continue to grow and so must data processing capabilities.

Posted June 15, 2015

Syncsort is working with Dell to help businesses improve operational efficiency and lower costs by shifting expensive workloads and associated data from enterprise data warehouses (EDW) to Hadoop.

Posted June 10, 2015

Map R is releasing version 5.0 of the MapR Distribution including Hadoop that will process big and fast data on a single data platform that enables a new class of real-time applications.

Posted June 09, 2015

Unstructured data presents many challenges — it's hard to manage, datasets can be extremely large, and it does not have a pre-defined schema. Still, enterprises and service providers who manage to tame and mine unstructured data will have the ability to drive true business transformation based on the new insights it provides.

Posted June 05, 2015

MapR Technologies, Inc., a provider of a distribution for Apache Hadoop, is including Apache Drill 1.0 in the MapR Distribution.

Posted May 19, 2015

The shortage of skilled talent and data scientists in Western Europe and the U.S. has triggered the question of whether to outsource analytical activities. This need is further amplified by competitive pressure to reduce time to market and lower costs.

Posted May 19, 2015

As the excitement and opportunity provided by big data tools develop, many organizations find their big data initiatives originating outside existing data management policies. As a result, many concepts of formal data governance are either intentionally or unintentionally omitted as these enterprises race to ingest huge new data streams at a feverish pace in the hope of increased insight and new analytic value.

Posted May 19, 2015

Similar to the dot-com revolution, the Internet of Things is the culmination of radical advances in four core technology pillars.

Posted May 19, 2015

Google white papers have inspired many great open source projects. What has been missing until now, however, has been a way of bringing these technologies together such that any data-centric organization can benefit from the capabilities of each technology across its entire data center, and in new ways not documented by any single white paper. This is called the "Zeta Architecture."

Posted May 19, 2015

Business pressures, including cost reduction, scalability, and "just-in-time" application software implementation, are just some of the requirements prompting businesses to "cloudify" at least some aspect of their IT infrastructure.

Posted May 19, 2015

Data-driven companies continue to explore data management technologies that better unify operational, analytical, and other disparate or siloed data in a way that offers tangible business value and data management relief.

Posted May 19, 2015

The demand for effective data management is intensifying. At the same time, the database market has expanded into a wide array of solutions—from traditional relational database management systems to alternative databases such as NoSQL, NewSQL, cloud, and in-memory offerings.

Posted May 19, 2015

Data preparation is gaining considerable visibility as a distinct aspect of data management and analytics work.

Posted May 19, 2015

Just when you thought NoSQL meant the end of SQL, think again, and realize why you need to hold on to your relational database administrator like it was 1999. NoSQL has proven to be a resilient next-generation database technology for increasingly common internet-era specialized workloads. Now approaching a decade after its arrival on the scene, NoSQL is moving beyond architectural marvels to practical tools in the software development toolkit and, in that process, unveiling tried-and-true capabilities formerly known to be the scalpels of the enterprise relational database. Let's go back to the future and take a look at how the DBA is becoming as relevant as ever while NoSQL evolves for the enterprise.

Posted May 19, 2015

RedPoint Global was founded in 2006 by Dale Renner, Lewis Clemmens, and George Corugedo, who previously had worked together at Accenture. Based in Wellesley, Mass., RedPoint collaborates with clients around the world in 11 different verticals. "We have always been very focused on the data, and recognize that a lot of business problems live and die by the quality of the data," says Corugedo.

Posted May 19, 2015

Deep Information Sciences has closed $8 million in Series A funding. The round brings the total invested in Deep to $18 million. The funding will assist in the growth of the Deep Engine, which break downs the performance, speed and scale limitations of databases to help businesses achieve new insights and opportunities from big data.

Posted May 05, 2015

Perhaps no other technology is more intertwined with the promise of big data than Apache Hadoop, the open source framework that emerged 10 years ago. Hadoop was first leveraged at big web companies for its ability to process large quantities of varied-format data using affordable commodity servers. Today, the Hadoop ecosystem is expanding swiftly, and its value in the enterprise is being recognized.

Posted April 15, 2015

Pivotal has proposed "Project Geode" for incubation by the Apache Software Foundation (ASF). A distributed in-memory database, Geode will be the open source core of Pivotal GemFire, and is now available for review at network.pivotal.io. Pivotal plans to contribute to, support, and help build the Project Geode community while simultaneously producing its commercial distribution of Pivotal GemFire.

Posted April 13, 2015

Hortonworks, a contributor to and provider of enterprise Apache Hadoop, has signed a definitive agreement to acquire SequenceIQ. "This acquisition complements our strategy of providing enterprise customers the broadest choice of consumption options for Hortonworks Data Platform, from on-premise deployments to cloud architectures,"said Rob Bearden, chief executive officer of Hortonworks.

Posted April 13, 2015

To fully take advantage of big data tools and architectures, businesses need to adapt a different mindset, according to Edd Dumbill, who contends that looking at the data value chain is the first step to understanding the value of data.

Posted April 08, 2015

Data is being collected everywhere - and from everything. The idea is that it can provide the power of insights never before possible into everything from the patient care to the health of machinery to customer sentiment about products and services. But to reveal these valuable insights, this data also has to be captured and analyzed in ways never before possible.

Posted April 08, 2015

The emergence of the Internet of Things is the driving force behind the latest evolutionary change in the way we think about database platforms and the most important change since the development of the original "big data" technologies a decade ago.

Posted April 08, 2015

Pages
1
2
3
4
5
6
7
8
9

Newsletters

Subscribe to Big Data Quarterly E-Edition