Big Data Quarterly Articles



One of the most common challenges organizations face when developing an enterprise data governance program is the presence of data silos or pockets of data initiatives scattered throughout the company with little to no association or collaboration of efforts. While many describe their data systems as "siloed," the disjunction is less about the underlying technology and far more about the division created between lines of business.

Posted March 25, 2016

The emerging Internet of Things (IoT) represents a huge opportunity for today's enterprises but also poses risks. Many organizations are challenged to open up their systems to ever-growing networks of devices, sensors, and systems that are relentlessly spewing data that may or may not be of value. But questions about security and systems performance swirl around as well.

Posted March 25, 2016

The data lake has been the subject of more than its fair share of critics since its inception. Pundits claim it's a source of chaos and risk. Analysts often slam the concept, calling it a "data swamp" or "data dump." As a result of this scrutiny, the definition and understanding of the definition of the data lake are rather murky.

Posted March 24, 2016

The rise of big data technologies in enterprise IT is now seen as an inevitability, but adoption has occurred at a slower pace than expected, according to Joe Caserta, president and CEO of Caserta Concepts, a firm focused on big data strategy consulting and technology implementation. Caserta recently discussed the trends in big data projects, the technologies that offer key advantages now, and why he thinks big data is reaching a turning point.

Posted March 23, 2016

Analytics and the platforms that support big data are constantly evolving, being shaped by the need to deliver data faster to users and gain effective insights throughout the organization. Hadoop, Spark, Kafka, and the cloud are some of the technologies that can handle the demand the future will bring, according to Kevin Petrie, senior director at Attunity.

Posted March 14, 2016

In a new book titled "Next Generation Databases," Guy Harrison, an executive director of R&D at Dell, shares what every data professional needs to know about the future of databases in a world of NoSQL and big data.

Posted March 08, 2016

As more and more data comes into the enterprise, companies are looking to build real-time big data architectures to keep up with an increased amount of information.

Posted March 07, 2016

Throughout the year, AAA National Action Center receives information from its federated local clubs. There is a need to align that information to help AAA further understand relevancy and deliver strong engagement opportunities for its members. After a lengthy process of exploring and evaluating solutions in industry and benchmarking different products, AAA National decided to partner with RedPoint Global.

Posted February 26, 2016

SnapLogic is unveiling its Winter 2016 release of its flagship platform, further connecting users to flexible Spark capabilities and reliable big data integration solutions.

Posted February 23, 2016

The promise of the data lake is an enduring repository of raw data that can be accessed now and in the future for different purposes. To help companies on their journey to the data lake, Information Builders has unveiled the iWay Hadoop Data Manager, a new solution that provides an interface to generate portable, reusable code for data integration tasks in Hadoop.

Posted February 23, 2016

It is hard to think of a technology that is more identified with the rise of big data than Hadoop. Since its creation, the framework for distributed processing of massive datasets on commodity hardware has had a transformative effect on the way data is collected, managed, and analyzed - and also grown well beyond its initial scope through a related ecosystem of open source projects. With 2016 recognized as the 10-year anniversary for Hadoop, Big Data Quarterly chose this time to ask technologists, consultants, and researchers to reflect on what has been achieved in the last decade, and what's ahead on the horizon.

Posted February 18, 2016

Every company is undoubtedly concerned about keeping outside attackers away from its sensitive data, but understanding who has access to that data from within the organization can be an equally challenging task. The goal of every attacker is to gain privileged access. An excessively privileged user account can be used as a weapon of destruction in the enterprise, and if a powerful user account is compromised by a malicious attacker, all bets are off.

Posted February 17, 2016

Over the last half decade, we've watched SQL purists butt heads with NoSQL upstarts, Hadoop triumphalists clash with Hadump pessimists, database geeks war with application developers, and so on. In the midst of all this warring, we've tried to fit—and, in many cases, to cram—the new into the old, the old into the new, with the result that at one time or another, we've asked the impossible of all of the components in our ever-expanding technology portfolios.

Posted February 16, 2016

Enterprises can't seem to pack enough big data and analytics solutions into their data centers, executive suites, and everywhere else across their organizations. Just about every venture-capital-cash-fueled startup from Silicon Valley to Boston has an analytics component to it. As these firms rapidly gain traction, they are being scarfed up by larger vendors looking to solidify their leadership of the analytics space.

Posted February 16, 2016

The Internet of Things holds great promise for everything from better healthcare to decreased traffic accidents and more efficient manufacturing processes. Michael Morton is currently CTO at Dell Boomi, which he joined in 2013 after a career with IBM where he became an IBM Master Inventor and worked directly with Fortune 100 companies. Recently, Morton talked with BDQ about some of the opportunities and challenges and the role that Boomi plays in the emerging IoT market.

Posted February 11, 2016

Looker, provider of a BI platform, has added support for Presto and Spark SQL as well as updates to its support for Impala and Hive.

Posted February 09, 2016

Deep Information Sciences is unveiling a new solution that combines a MySQL-compliant database with cloud- and resource-awareness.

Posted February 04, 2016

With Hadoop marking its 10th anniversary this year, Sean Suchter, CEO of Pepperdata, recently reflected on his experience with the platform and speculated on what the next 10 years may bring.

Posted February 03, 2016

Pentaho, a Hitachi Group company, is integrating Pentaho Labs with Python, increasing productivity and data governance, while allowing users to focus on predictive analytics and machine learning.

Posted February 02, 2016

To address a growing need for data professionals with graph database expertise, Neo4j is offering a certification exam online through its GraphAcademy.

Posted February 02, 2016

Redis Labs, is integrating with Spark SQL, and releasing a Spark-Redis connector package that promises to accelerate processing time.

Posted February 02, 2016

Snowflake Computing, a cloud data warehousing company, has formed a technology and go-to-market partnership with Looker, which provides a data exploration and business intelligence platform.

Posted February 01, 2016

Data constraint - which hinders the ability to easily provision parity data from production data sources for use in development environments, QA, UAT (user acceptance testing), and integration testing - can lead to complications in application development such as insufficient resources, bad quality data which leads to bugs, and slow environment builds.

Posted February 01, 2016

To address increasing demands posed by modern applications, Aerospike has released version 3.7 of its NoSQL database with new features that enable more personalized applications, more efficient operations, and greater stability across public and private clouds.

Posted January 28, 2016

For decades, the enterprise data warehouse (EDW) has been the aspirational analytic system for just about every organization. It has taken many forms throughout the enterprise, but all share the same core concepts of integration/consolidation of data from disparate sources, governing that data to provide reliability and trust, and enabling reporting and analytics. The last few years, however, have been very disruptive to the data management landscape. The "big data" era has introduced new technologies and techniques that provide alternatives to the traditional EDW approach, and in many cases, exceeding its capabilities. Many claim we are now in a post-EDW era and the concept itself is legacy.

Posted January 19, 2016

The year 2015 started out with people recognizing that the Hadoop ecosystem is here to stay, and ended as the year in which organizations achieved real success within the Hadoop ecosystem. Today, more projects are popping up within the Hadoop ecosystem that can run both with and without Hadoop. The great thing about this trend is that it lowers the barrier to entry for people to get started with these technologies. More importantly, all of these new technologies work best at large scale within the rest of the Hadoop ecosystem, while Hadoop MapReduce has begun its ride off into the sunset.

Posted January 19, 2016

Rocana has unveiled the latest version of its solution for managing and analyzing event-oriented machine data that introduces new advanced analytics and anomaly detection abilities. In addition to the new features, the platform update also introduces support for Hortonworks along with Cloudera, further deepening the platform's reach.

Posted January 12, 2016

Greenwave Systems, an Internet of Things (IoT) software and managed services leader, has raised $45 million in its Series C financing round in two phases. The company's stated mission is to drive mass adoption of IoT by enabling leading brands to use technology in ways that foster deeper customer relationships.

Posted January 05, 2016

The continual evolution in technology has allowed for more data sources than previously thought possible. The growth of SaaS tools provides many benefits, but there is a downside as well. Bringing these cloud data sources into a coherent system for reporting is perpetually a challenge for IT and business intelligence teams. A recent DBTA roundtable webcast covered the issues of combining different SaaS applications into a cloud based enterprise data and leveraging the simple data pipe. Presenters included with Sarah Maston, solution architect with IBM Cloud Data Services, and Erin Franz, alliances data analyst with Looker.

Posted January 04, 2016

We can expect to see every year within the next 5 years be "The Year of IoT." IoT promises to unlock value and rapidly transform how organizations manage, operationalize, and monetize their assets. With IoT, physical assets become liquid, easily indexed and tracked, enabling identification of idle capacity or over utilization.

Posted December 22, 2015

Today's organizations don't need to be hesitant or "scared" of unstructured data. In actuality, the real fear should be about missing the opportunity to tap these resources. The magic really starts to happen when all datasets - structured and unstructured, inside and outside the enterprise - are analyzed in parallel.

Posted December 22, 2015

Hortonworks revealed upcoming advancements of Hortonworks Data Platform (HDP) with the in-memory analytic capabilities of Spark. Hortonworks has also launched Hortonworks Community Connection (HCC), a new online collaboration destination for developers, DevOps, customers and partners to get answers to questions, collaborate on technical articles and share code examples from GitHub.

Posted December 10, 2015

The challenges that big data companies face when dealing with live data range from scalability and availability to real-time demands and structure. The world of data streaming is a complex one as it involves millions of events gathered per day.

Posted December 09, 2015

Platfora has added $30 million in growth-stage financing from new investors HSBC and Harmony Partners, with participation from earlier-stage backers, including Allegis Capital, Andreessen Horowitz, Battery Ventures, Citi Ventures, Cisco, Sutter Hill Ventures and Tenaya Capital. This new round brings Platfora's total amount raised to $95 million.

Posted December 08, 2015

IBM is acquiring Clearleap to boost its IBM Cloud platform and accelerate the delivery of video service as a strategic source of data for business. The acquisition is part of IBM's moves to help clients improve access to "dark data" which, because it is unstructured and dark to computer systems, cannot be effectively managed or exploited.

Posted December 08, 2015

Harnessing operational big data does not come with a "one size fits all" solution. Organizations are at different points in their data management cycles, and whether that they are building new applications or optimizing existing ones, each needs a unique solution. This was among the key points made during a special DBTA roundtable webinar on harnessing operational big data. The webinar featured Matt Allen, senior product marketing manager with MarkLogic, Kevin Petrie, senior director with Attunity, and Jason Paul Kazarian, senior architect with Hewlett Packard Enterprise (HPE).

Posted December 07, 2015

As 2015 draws to a close it's time to look back on the year's big changes in data management and reflect on some of the most insightful observations by leading data experts. Here, Big Data Quarterly presents a round-up of thought-provoking articles that explore some of the key advances of 2015, how organizations are harnessing what big data has to offer, and the challenges they face as they seek to compete on analytics.

Posted December 02, 2015

Hadoop distribution provider Cloudera has introduced Cloudera Enterprise 5.5, including Cloudera Navigator Optimizer, a new product targeted at helping organizations improve big data workload performance and efficiency. Cloudera Navigator Optimizer, now in beta, is expected to be generally available in 2016. The new release of Cloudera Enterprise has three main areas of focus, according to Anupam Singh, head of data management at Cloudera.

Posted November 19, 2015

Alpine Data has announced the general availability of the Alpine Custom Operator Framework, a methodology for developing custom algorithms that can be plugged into Alpine's parallel machine learning engine. Complementing Alpine Touchpoints, the Custom Operator Framework enables data science and business analyst teams to create, manage and distribute frequently-requested analytic assets to business users directly into their existing activities and workflows.

Posted November 17, 2015

Not too long ago, large volumes of information were available only to the select few organizations able to afford the expensive IT infrastructure to collect, store, manage, and analyze it—the big-budget companies with seemingly bottomless pockets, or the professional research services that built empires on collecting and interpreting data. But now, through the realized effects of Moore's law, along with the consumerization of BI tools, this data is increasingly available to everyone—and without having to dig deep into budgets or employ robust IT departments.

Posted November 13, 2015

As the business value of big data increases, vendors are offering cloud-hosted big data technology, known as big data as a service (BDaaS). Why choose BDaaS instead of on-premises deployments? Do you lack internal IT expertise in big data? Is the upfront cost for a cluster an issue? Or do you not have the luxury of time to build a cluster? These are all important questions when considering BDaaS.

Posted November 13, 2015

With the Hadoop ecosystem expanding rapidly, Hortonworks' Shaun Connolly recently discussed how Hadoop and related technologies are growing and being used, as well as the factors for a successful deployment.

Posted November 13, 2015

A typical organization loses about 5% of its revenues to fraud each year. The total cost of non-health insurance fraud in the U.S. is estimated to be more than $40 billion per year. These numbers stress the importance and need of finding sophisticated tools to both detect and prevent fraud. Big data and analytics offer a new valuable toolkit in the fight against fraud.

Posted November 13, 2015

"Caveat emptor" is Latin for "Let the buyer beware." In the realm of the modern information technology cloud, this sage advice rings especially true.

Posted November 13, 2015

With increased demand for mass customization and personalization, the emergence of Web 2.0, and one-to-one marketing, and the need for better risk management and timely fraud detection, the pressure is on for organizations to improve their ability to extract, understand, and exploit analytical patterns of customer behavior and strategic intelligence.

Posted November 12, 2015

To better manage the data explosion now and with scalable options for the future, existing data architecture is evolving beyond traditional databases, data stores, data warehouses, and the like into a more unfiltered repository known as the data lake.

Posted November 05, 2015

A new Intel IoT Platform reference architecture and new hardware and software products have been introduced as part of the company's effort to build out comprehensive offerings for the Internet of Things (IoT) marketplace.

Posted November 05, 2015

Trove is the database as a service component of OpenStack that lets administrators and DevOps manage multiple instances of different database management systems (DBMS), both relational and NoSQL, using a common infrastructure. It makes database capacity available that can be consumed on-demand. This makes it much easier and faster to select and operate a database - while retaining all its capabilities - in a secure private or public cloud. The result is that routine tasks like provisioning, and managing regular administrative tasks like clustering, replication, backup and restore are handled in a simple, unified way.

Posted November 04, 2015

MapR is including Apache Drill 1.2 in its Apache Hadoop distribution and is also now offering a new Data Exploration Quick Start Solution, leveraging Drill to help customers get started more rapidly with big data projects. Apache Drill is an open source, low-latency query engine for Hadoop that delivers secure, interactive SQL analytics at petabyte scale. With the two announcements, MapR says customers and partners will able to more quickly leverage Drill to get fast business insights from all their data in Hadoop and other sources. MapR also released a comprehensive SQL-based test framework to the open source community.

Posted October 21, 2015

Citus Data, which provides solutions to scale out PostgreSQL for real-time big data, has secured $9.5 million in Series A funding led by Khosla Ventures, with participation by previous Citus Data investors including Data Collective and Vaizra Investments. The company says the funding round will be used to grow the Citus Data development team and expand its sales and marketing efforts.

Posted October 20, 2015

Pages
1
2
3
4
5
6
7
8
9

Newsletters

Subscribe to Big Data Quarterly E-Edition