Big Data Quarterly Articles



Enabled by a partnership with Pentaho, a Hitachi Group Company, and integration with Pentaho's Big Data Integration and Analytics platform, Melissa Data's data quality tools and services can now be scaled across the Hadoop cluster to cleanse and verify data center records.

Posted April 27, 2016

Cloudera, provider of a data management and analytics platform built on Apache Hadoop and open source technologies, has announced the general availability of Cloudera Enterprise 5.7. According to the vendor, the new release offers an average 3x improvement for data processing with added support of Hive-on-Spark, and an average 2x improvement for business intelligence analytics with updates to Apache Impala (incubating).

Posted April 26, 2016

The core reason for implementing in-memory technology is to improve performance. To help accelerate adoption of in-memory technologies and provide a universal standard for columnar in-memory processing and interchange, the lead developers of 13 major open source big data projects have joined forces to create Apache Arrow, a new top level project within the Apache Software Foundation (ASF).

Posted April 24, 2016

Percona, a provider of MySQL and MongoDB solutions and services, is releasing an updated version of Percona Server for MongoDB .

Posted April 21, 2016

Teradata, the big data analytics and marketing applications company, is making key investments in the Internet of Things (IoT) and the Analytics of Things (AoT), along with updating its signature platforms.

Posted April 18, 2016

First created as part of a research project at UC Berkeley AMPLab, Spark is an open source project in the big data space, built for sophisticated analytics, speed, and ease of use. It unifies critical data analytics capabilities such as SQL, advanced analytics, and streaming in a single framework. Databricks is a company that was founded by the team that created and continues to lead both the development and training around Apache Spark.

Posted April 14, 2016

Thanks to the digital business transformation, the world around us is changing—and quickly—to a very consumer- and data-centric economy, where companies must transform to remain competitive and survive. The upshot is that for many companies today, it is a full-on Darwinian experience of survival of the fittest.

Posted April 08, 2016

SnapLogic is releasing its hybrid execution framework Snaplex on the Microsoft Azure Marketplace as Azureplex, giving users the ability to gain business insights faster with self-service data integration from a plethora of sources.

Posted April 07, 2016

The emergence of big data, characterized in terms of its four V's—volume, variety, velocity, and veracity—has created both opportunities and challenges for credit scoring.

Posted April 04, 2016

As data visualization increasingly becomes top-of-mind for data-driven organizations, it's time to introduce the concept of data visualization competency. There is a need today to provide a framework to fingerprint data visualizations as unique digital assets in the business for maximum impact and consistent execution against strategic business practices and goals.

Posted April 01, 2016

Hershey's LLC recently deployed the Infosys Information Platform on AWS to analyze retail store data.

Posted March 31, 2016

Over the past 2 years, there have been big announcements from all of the major car manufacturers about their connected car initiatives, lots of M&A activity in the technology industry as they race to supply the revolution, and major global alliances of telecom providers being formed to provide the underlying connectivity and infrastructure. But, most of all, we are actually starting to see some of the promised transformational benefits of the Internet of Things becoming a reality.

Posted March 31, 2016

The pervasive corporate mindset to transition all levels of infrastructure to some cloud, somewhere, is accelerating the growth of the cloud industry with a rapidity so far unseen in the history of computing. This phenomenon has resulted in weighty pressure on CIOs to develop and deploy an effective and comprehensive cloud strategy or risk their organization falling behind this undeniable trend. The internet changed the information technology game, but now the cloud constitutes an entirely different league.

Posted March 31, 2016

There's a need to enable better decision making today with faster access to data. But many organizations are still weighed down by integration and management processes that are not keeping up with the increasing volume, variety, and velocity of data. A greater emphasis on cloud and self-service tools may provide an approach to remedy the situation.

Posted March 30, 2016

Trifacta, a provider of data wrangling technology, is introducing the Photon Compute Framework, providing users with an interactive platform for large in-memory datasets.

Posted March 29, 2016

NoSQL databases were born out of the need to scale transactional persistence stores more efficiently. In a world where the relational database management system (RDBMS) was king, this was easier said than done.

Posted March 29, 2016

The digital economy promises to redefine nearly every aspect of a company's operations—from raw material procurement through post-sale services. Yet, some of the most dramatic changes will be seen in how companies evolve their product portfolios and leverage digital capabilities.

Posted March 25, 2016

One of the most common challenges organizations face when developing an enterprise data governance program is the presence of data silos or pockets of data initiatives scattered throughout the company with little to no association or collaboration of efforts. While many describe their data systems as "siloed," the disjunction is less about the underlying technology and far more about the division created between lines of business.

Posted March 25, 2016

The emerging Internet of Things (IoT) represents a huge opportunity for today's enterprises but also poses risks. Many organizations are challenged to open up their systems to ever-growing networks of devices, sensors, and systems that are relentlessly spewing data that may or may not be of value. But questions about security and systems performance swirl around as well.

Posted March 25, 2016

The data lake has been the subject of more than its fair share of critics since its inception. Pundits claim it's a source of chaos and risk. Analysts often slam the concept, calling it a "data swamp" or "data dump." As a result of this scrutiny, the definition and understanding of the definition of the data lake are rather murky.

Posted March 24, 2016

The rise of big data technologies in enterprise IT is now seen as an inevitability, but adoption has occurred at a slower pace than expected, according to Joe Caserta, president and CEO of Caserta Concepts, a firm focused on big data strategy consulting and technology implementation. Caserta recently discussed the trends in big data projects, the technologies that offer key advantages now, and why he thinks big data is reaching a turning point.

Posted March 23, 2016

Analytics and the platforms that support big data are constantly evolving, being shaped by the need to deliver data faster to users and gain effective insights throughout the organization. Hadoop, Spark, Kafka, and the cloud are some of the technologies that can handle the demand the future will bring, according to Kevin Petrie, senior director at Attunity.

Posted March 14, 2016

In a new book titled "Next Generation Databases," Guy Harrison, an executive director of R&D at Dell, shares what every data professional needs to know about the future of databases in a world of NoSQL and big data.

Posted March 08, 2016

As more and more data comes into the enterprise, companies are looking to build real-time big data architectures to keep up with an increased amount of information.

Posted March 07, 2016

Throughout the year, AAA National Action Center receives information from its federated local clubs. There is a need to align that information to help AAA further understand relevancy and deliver strong engagement opportunities for its members. After a lengthy process of exploring and evaluating solutions in industry and benchmarking different products, AAA National decided to partner with RedPoint Global.

Posted February 26, 2016

SnapLogic is unveiling its Winter 2016 release of its flagship platform, further connecting users to flexible Spark capabilities and reliable big data integration solutions.

Posted February 23, 2016

The promise of the data lake is an enduring repository of raw data that can be accessed now and in the future for different purposes. To help companies on their journey to the data lake, Information Builders has unveiled the iWay Hadoop Data Manager, a new solution that provides an interface to generate portable, reusable code for data integration tasks in Hadoop.

Posted February 23, 2016

It is hard to think of a technology that is more identified with the rise of big data than Hadoop. Since its creation, the framework for distributed processing of massive datasets on commodity hardware has had a transformative effect on the way data is collected, managed, and analyzed - and also grown well beyond its initial scope through a related ecosystem of open source projects. With 2016 recognized as the 10-year anniversary for Hadoop, Big Data Quarterly chose this time to ask technologists, consultants, and researchers to reflect on what has been achieved in the last decade, and what's ahead on the horizon.

Posted February 18, 2016

Every company is undoubtedly concerned about keeping outside attackers away from its sensitive data, but understanding who has access to that data from within the organization can be an equally challenging task. The goal of every attacker is to gain privileged access. An excessively privileged user account can be used as a weapon of destruction in the enterprise, and if a powerful user account is compromised by a malicious attacker, all bets are off.

Posted February 17, 2016

Over the last half decade, we've watched SQL purists butt heads with NoSQL upstarts, Hadoop triumphalists clash with Hadump pessimists, database geeks war with application developers, and so on. In the midst of all this warring, we've tried to fit—and, in many cases, to cram—the new into the old, the old into the new, with the result that at one time or another, we've asked the impossible of all of the components in our ever-expanding technology portfolios.

Posted February 16, 2016

Enterprises can't seem to pack enough big data and analytics solutions into their data centers, executive suites, and everywhere else across their organizations. Just about every venture-capital-cash-fueled startup from Silicon Valley to Boston has an analytics component to it. As these firms rapidly gain traction, they are being scarfed up by larger vendors looking to solidify their leadership of the analytics space.

Posted February 16, 2016

The Internet of Things holds great promise for everything from better healthcare to decreased traffic accidents and more efficient manufacturing processes. Michael Morton is currently CTO at Dell Boomi, which he joined in 2013 after a career with IBM where he became an IBM Master Inventor and worked directly with Fortune 100 companies. Recently, Morton talked with BDQ about some of the opportunities and challenges and the role that Boomi plays in the emerging IoT market.

Posted February 11, 2016

Looker, provider of a BI platform, has added support for Presto and Spark SQL as well as updates to its support for Impala and Hive.

Posted February 09, 2016

Deep Information Sciences is unveiling a new solution that combines a MySQL-compliant database with cloud- and resource-awareness.

Posted February 04, 2016

With Hadoop marking its 10th anniversary this year, Sean Suchter, CEO of Pepperdata, recently reflected on his experience with the platform and speculated on what the next 10 years may bring.

Posted February 03, 2016

Pentaho, a Hitachi Group company, is integrating Pentaho Labs with Python, increasing productivity and data governance, while allowing users to focus on predictive analytics and machine learning.

Posted February 02, 2016

To address a growing need for data professionals with graph database expertise, Neo4j is offering a certification exam online through its GraphAcademy.

Posted February 02, 2016

Redis Labs, is integrating with Spark SQL, and releasing a Spark-Redis connector package that promises to accelerate processing time.

Posted February 02, 2016

Snowflake Computing, a cloud data warehousing company, has formed a technology and go-to-market partnership with Looker, which provides a data exploration and business intelligence platform.

Posted February 01, 2016

Data constraint - which hinders the ability to easily provision parity data from production data sources for use in development environments, QA, UAT (user acceptance testing), and integration testing - can lead to complications in application development such as insufficient resources, bad quality data which leads to bugs, and slow environment builds.

Posted February 01, 2016

To address increasing demands posed by modern applications, Aerospike has released version 3.7 of its NoSQL database with new features that enable more personalized applications, more efficient operations, and greater stability across public and private clouds.

Posted January 28, 2016

For decades, the enterprise data warehouse (EDW) has been the aspirational analytic system for just about every organization. It has taken many forms throughout the enterprise, but all share the same core concepts of integration/consolidation of data from disparate sources, governing that data to provide reliability and trust, and enabling reporting and analytics. The last few years, however, have been very disruptive to the data management landscape. The "big data" era has introduced new technologies and techniques that provide alternatives to the traditional EDW approach, and in many cases, exceeding its capabilities. Many claim we are now in a post-EDW era and the concept itself is legacy.

Posted January 19, 2016

The year 2015 started out with people recognizing that the Hadoop ecosystem is here to stay, and ended as the year in which organizations achieved real success within the Hadoop ecosystem. Today, more projects are popping up within the Hadoop ecosystem that can run both with and without Hadoop. The great thing about this trend is that it lowers the barrier to entry for people to get started with these technologies. More importantly, all of these new technologies work best at large scale within the rest of the Hadoop ecosystem, while Hadoop MapReduce has begun its ride off into the sunset.

Posted January 19, 2016

Rocana has unveiled the latest version of its solution for managing and analyzing event-oriented machine data that introduces new advanced analytics and anomaly detection abilities. In addition to the new features, the platform update also introduces support for Hortonworks along with Cloudera, further deepening the platform's reach.

Posted January 12, 2016

Greenwave Systems, an Internet of Things (IoT) software and managed services leader, has raised $45 million in its Series C financing round in two phases. The company's stated mission is to drive mass adoption of IoT by enabling leading brands to use technology in ways that foster deeper customer relationships.

Posted January 05, 2016

The continual evolution in technology has allowed for more data sources than previously thought possible. The growth of SaaS tools provides many benefits, but there is a downside as well. Bringing these cloud data sources into a coherent system for reporting is perpetually a challenge for IT and business intelligence teams. A recent DBTA roundtable webcast covered the issues of combining different SaaS applications into a cloud based enterprise data and leveraging the simple data pipe. Presenters included with Sarah Maston, solution architect with IBM Cloud Data Services, and Erin Franz, alliances data analyst with Looker.

Posted January 04, 2016

We can expect to see every year within the next 5 years be "The Year of IoT." IoT promises to unlock value and rapidly transform how organizations manage, operationalize, and monetize their assets. With IoT, physical assets become liquid, easily indexed and tracked, enabling identification of idle capacity or over utilization.

Posted December 22, 2015

Today's organizations don't need to be hesitant or "scared" of unstructured data. In actuality, the real fear should be about missing the opportunity to tap these resources. The magic really starts to happen when all datasets - structured and unstructured, inside and outside the enterprise - are analyzed in parallel.

Posted December 22, 2015

Hortonworks revealed upcoming advancements of Hortonworks Data Platform (HDP) with the in-memory analytic capabilities of Spark. Hortonworks has also launched Hortonworks Community Connection (HCC), a new online collaboration destination for developers, DevOps, customers and partners to get answers to questions, collaborate on technical articles and share code examples from GitHub.

Posted December 10, 2015

The challenges that big data companies face when dealing with live data range from scalability and availability to real-time demands and structure. The world of data streaming is a complex one as it involves millions of events gathered per day.

Posted December 09, 2015

Pages
1
2
3
4
5
6
7
8
9

Newsletters

Subscribe to Big Data Quarterly E-Edition