Big Data Quarterly Articles



There has been a sea of change in how enterprises are thinking about Apache Hadoop and big data. Today, a majority of enterprises are thinking about the cloud first, not on-premises, and are increasingly relying on ecosystem standards to drive their Apache Hadoop distribution selection.

Posted April 07, 2017

It is difficult to find someone not talking about or considering using containers to deploy and manage their enterprise applications. A container just looks like another process running on a system; a dedicated CPU and pre-allocated memory aren't required in order to run a container. The simplicity of building, deploying, and managing containers is among the reasons that containers are growing rapidly in popularity.

Posted April 07, 2017

Alation and Trifacta say they are extending their partnership to jointly deliver an integrated solution for self-service data discovery and preparation that enables users to access the data catalog and data wrangling features within a single interface.

Posted March 15, 2017

SAP has announced advancements in the SAP Vora solution to help customers accelerate project implementations and improve their enterprise business analytics.

Posted March 15, 2017

Dataguise, a provider of sensitive data governance, has announced that DgSecure now provides sensitive data monitoring and masking in Apache Hive.

Posted March 15, 2017

The rise of big data and the growing popularity of cloud is a combination that presents valuable new opportunities to leverage data with greater efficiency. But organizations also need to be aware of some key differences between on-premise and cloud deployments, says Charles Zedlewski, senior vice president, products, at Cloudera.

Posted March 15, 2017

Ash Munshi, Pepperdata CEO, recently discussed the need for DevOps for big data, and the role of the Dr. Elephant project, which was open sourced in 2016 by LinkedIn and is available under the Apache v2 License.

Posted March 07, 2017

Tableau Software is releasing an updated version of its namesake platform, bringing advanced mapping capabilities to the analytics solution. Tableau 10.2 will make complex geospatial analysis easier, simplify data prep with new ways to combine and clean data, and give enterprises more tools to deliver self-service analytics at scale, according to the company.

Posted March 03, 2017

Kong Yang, head geek at SolarWinds, believes the rise of the mobile workforce and the pressure to implement new technologies means that modern IT professionals must be able to quickly evolve beyond the confines of on-premises and shift into the realm of hybrid IT. Here, Yang reflects on some of the ways that IT professionals can begin that journey.

Posted February 24, 2017

MapR Technologies, Inc., which provides a converged data platform, has introduced persistent storage for containers with complete state access to files, database tables, and message streams from any location. The MapR Converged Data Platform for Docker includes the MapR Persistent Client Container (PACC) that enables stateful applications and microservices to access data for greater application agility and faster time-to-value.

Posted February 07, 2017

As news of data breaches continues to grab headlines, data security is becoming a greater enterprise concern. However, at the same time, it is becoming clear that many organizations are actually doing things that make their data more vulnerable. Recently, Joe Pasqua, executive vice president of products at MarkLogic, provider of enterprise NoSQL database technology, discussed the ways organizations and their employees are unintentionally putting their data at greater risk—and the ways to address it.

Posted February 03, 2017

Hadoop Fundamentals and Key Technologies in the Evolving Hadoop Ecosystem

Posted February 03, 2017

Cost of Data Breaches on the Rise

Posted January 27, 2017

vArmour has been awarded a patent by the USPTO for security technology for container microservices. Marc Woolward, CTO of vArmour described what's changing in the world of clouds, containers, and microservices.

Posted January 24, 2017

IoT has massive implications for businesses of all kinds, and for individuals at all organizational levels, as well. Bart Schouw, IoT solutions director at Software AG, recently reflected on the changes taking place and explained why 2020 will be a critical year for IoT.

Posted January 20, 2017

Patrick Hubbard, head geek and technical product marketing director at SolarWinds, recently discussed key themes that will emerge on the IT front in 2017. Smarter use of container technologies, a greater emphasis on security, and the continued shifting of IT roles, he says are some of the key trends that will take hold in the year ahead.

Posted January 13, 2017

Arcadia Data, a provider of visual analytics software, has added new native integration features for Arcadia Enterprise and Cloudera Enterprise to deliver a real-time, Hadoop-native analytics platform.

Posted January 11, 2017

Hortonworks has forged an open source collaboration with Neustar, a provider of real-time information services, on security and identity management tools for IoT devices.

Posted January 11, 2017

Xplenty has announced new $4 million in funding from Bain Capital Ventures, True Ventures, and Rembrandt Venture Partners, and with participation in the funding round from existing Xplenty investors Magma Venture Partners and Waarde Capital.

Posted January 04, 2017

The Data Warehousing Sanity Check

Posted January 04, 2017

Pitney Bowes has joined Hortonworks Partnerworks in the Modern Data Solutions (MDS) partner program. According to the vendors, location-based data, in particular, is becoming more important in how businesses understand their customers because it is one of the most consistent ways to link people, places, and things.

Posted January 04, 2017

Using Data Lake Management Strategies for Big Data Analytics

Posted January 03, 2017

ZeroPoint technology focuses on analyzing documents, email, web content and server traffic for hazardous content such as malicious code

Posted December 13, 2016

When software providers consider transitioning to (or at the very least adding) a SaaS offering, they think about the impact to their business of moving from a perpetual license model to a recurring revenue stream. And while it's easy to remember and consider such migration costs as application-level rearchitecture, other upfront and ongoing costs - such as infrastructure and service-related costs - are often severely underestimated.

Posted December 12, 2016

The Modern Heterogeneous Enterprise Data Architecture Takes Shape

Posted December 08, 2016

It has become all too clear that no organization is immune from the risk of a data breach, and that anyone accessing data can pose a threat - including trusted employees and partners. Here, IT executives speculate on the impact newer technologies such as IoT, blockchain, and cloud, as well as the need for data protection, including disaster recovery plans, encryption, and comprehensive oversight.

Posted December 07, 2016

Many providers of cloud services market the idea that all critical computing functions should be run using their public cloud services because this paradigm is the future and the future is now. While we do share that long-term vision, the reality is less impressive, and the solution is not yet complete. Amazon itself does not run 100% of its critical business systems in the AWS Public Cloud, a fact that was revealed in The Wall Street Journal article, "Cloud-Computing Kingpins Slow to Adapt to Own Movement." This is also true for Google, Microsoft, and other top cloud providers.

Posted November 15, 2016

The definition of "data visualization" often varies depending on whom you ask. For some, it's a process of visually transforming data for exploration or analysis. For others, it's a tool to share analytical insights or invite discovery.

Posted November 15, 2016

Data as a service (DaaS) is a business-centric service that transforms raw data into meaningful and reusable data assets, and delivers these data assets on-demand via a standard connectivity protocol in a pre-determined, configurable format and frequency for internal and external consumption.

Posted November 04, 2016

New data sources such as sensors, social media, and telematics along with new forms of analytics such as text and graph analysis have necessitated a new data lake design pattern to augment traditional design patterns such as the data warehouse. Unlike the data warehouse - an approach based on structuring and packaging data for the sake of quality, consistency, reuse, ease of use, and performance - the data lake goes in the other direction by storing raw data that lowers data acquisition costs and provides a new form of analytical agility.

Posted November 03, 2016

A new semantic-based graph data model has emerged within the enterprise. This data model has all of the advantages of the relational data model, but goes even further in providing for more intelligence built into the database itself, enabling greater elasticity to absorb the inevitable changes to data requirements, at cloud scales.

Posted November 02, 2016

Data has become a disruptive force for global businesses and a catalyst for digital transformation. But data can only be leveraged for BI initiatives to the extent it can be accessed and trusted. And, while today's self-service BI and analytics tools satisfy a user's craving for more "consumerized" technology, they often leave an analyst stuck in neutral because the users, first and foremost, cannot find the data they need to perform any analysis.

Posted November 02, 2016

Kinetica, provider of an in-memory database accelerated by GPUs (graphics processing units) has introduced two new software and services offerings designed to help customers ingest and use streaming datasets through use of GPUs.

Posted October 31, 2016

The focus of data governance should not be on creating bureaucracy and rules, but instead on business enablement within context of use. To do this, I suggest looking at data governance not as enforcement of a discipline, but instead as a process of guiding a data expedition. Let's look at what a data expedition entails and how data governance will be the guide of this ongoing journey.

Posted October 13, 2016

Data Modeling for the Modern World

Posted October 10, 2016

The rise of big data with new sources of data for analytics represents new opportunity to put data to work in organizations for a wide range of uses. A developing use case for leveraging data analytics on large datasets is fraud discovery.

Posted October 05, 2016

Choosing when to leverage cloud infrastructure is a topic that should not be taken lightly. There are a few issues that should be considered when debating cloud as part of a business strategy.

Posted October 04, 2016

NoSQL and Hadoop—two foundations of the emerging agile data architecture—have been on the scene for several years now, and, industry observers say, adoption continues to accelerate—especially within mainstream enterprises that weren't necessarily at the cutting edge of technology in the past.

Posted October 04, 2016

At Strata + Hadoop World, Hortonworks showcased its technology solutions for streaming analytics, security, governance, and Apache Spark at scale.

Posted September 30, 2016

Data lakes are quickly transitioning from interesting idea to priority project. A recent study, "Data Lake Adoption and Maturity," from Unisphere Research showed that nearly half of respondents have an approved budget or have requested budget to launch a data lake project. What's driving this rapid rush to the lake?

Posted September 27, 2016

At Strata + Hadoop World, MapR Technologies announced support for microservices that leverage continuous analytics, automated actions, and rapid response to better impact business as it happens. The new capabilities in the MapR Platform range from microservices application monitoring and management to integrated support for agile microservices application development.

Posted September 27, 2016

Big Data 50 - Companies Driving Innovation

Posted September 14, 2016

Conventional wisdom insists that IT will migrate to the cloud entirely at some point. But practical experience shows that enterprises that have invested in legacy architecture that still has many years of life left in it are not likely to rip and replace, at potentially astronomical costs. Instead, implementing a Bimodal IT approach supported by SDDC on integrated systems will allow companies to address scalability needs with agility, while also ensuring the mission-critical functions of their legacy systems are not compromised.

Posted September 12, 2016

Data lakes are quickly transitioning from interesting idea to priority project. A recent study, "Data Lake Adoption and Maturity," from Unisphere Research showed that nearly half of respondents have an approved budget or have requested budget to launch a data lake project.

Posted September 12, 2016

The elastic and distributed technologies that run modern applications require a new approach to operations — one that learns about your infrastructure and assists IT operators with maintenance and problem-solving. The interdependencies between new applications are creating chaos in existing systems and surfacing the operational challenges of modern systems. Solutions such as micro services architectures alleviate the scalability pains of centralized proprietary services but at a tremendous cost in complexity.

Posted August 25, 2016

Getting to Know Hadoop and its Advantages

Posted August 25, 2016

Perhaps the biggest and most overlooked is how to create accurate test data. You're implementing a new system in order to deal with a massive amount of data, and perhaps your relational database can't handle the volume, so it's vitally important to properly test this new system and ensure that it doesn't fall over as soon as the data floods in.

Posted August 23, 2016

Informatica is releasing five new Informatica Cloud offerings in Amazon Web Services Marketplace (AWS Marketplace) to help organizations jumpstart data management projects in the cloud.

Posted August 12, 2016

Paxata is unleashing new native push/pull seamless connectivity options to and from Amazon Web Services (AWS) that include the Amazon Redshift data warehouse and Amazon Simple Storage Service (Amazon S3).

Posted August 11, 2016

Nimbus Data is releasing a new all-flash platform for cloud, big data, virtualization, and massive digital content that will offer unprecedented scale and efficiency.

Posted August 09, 2016

Pages
1
2
3
4
5
6
7
8
9
10
11
12
13
14

Newsletters

Subscribe to Big Data Quarterly E-Edition