Newsletters




Big Data

The well-known three Vs of Big Data - Volume, Variety, and Velocity – are increasingly placing pressure on organizations that need to manage this data as well as extract value from this data deluge for Predictive Analytics and Decision-Making. Big Data technologies, services, and tools such as Hadoop, MapReduce, Hive and NoSQL/NewSQL databases and Data Integration techniques, In-Memory approaches, and Cloud technologies have emerged to help meet the challenges posed by the flood of Web, Social Media, Internet of Things (IoT) and machine-to-machine (M2M) data flowing into organizations.



Big Data Articles

Revolution Analytics, a commercial provider of open source R software, has released Revolution R Open and Revolution R Plus.

Posted October 15, 2014

With Cloudera 5.2 the focus is on building products to deliver on the promise of the enterprise data hub that Cloudera introduced last year, said Clarke Patterson, senior director of product marketing at Cloudera. In particular, new capabilities make the technology more accessible to users who are not data scientists and also increase the level of security, two hurdles which can stand in the way of Hadoop adoption.

Posted October 15, 2014

Share the tips and best practices you use in your work all the time—the code snippets and favorite queries you keep within easy reach. "Best practices" include acknowledged and sensible ways of carrying out an activity; "tips" illustrate how some activity could be done in a better way or reveal some undocumented or not-so-well-known feature.

Posted October 15, 2014

Informatica PowerCenter v. 9.6.1 and Data Quality v. 9.6.1 have achieved Oracle Exadata Optimized and Oracle SuperCluster Optimized status through the Oracle PartnerNetwork (OPN). Customers can utilize Informatica PowerCenter and Data Quality to ingest, cleanse and transform various types of data into Oracle Exadata and Oracle SuperCluster to maximize the value of their engineered systems investment.

Posted October 15, 2014

The newest release of Oracle Exalytics In-Memory Machine, an engineered system for business analytics, includes Intel Xeon processors customized for Oracle business analytics workloads, supporting 50% speed, 50% more processing cores and 50% more memory compared to the previous generation. The Oracle Database In-Memory has also been certified with Oracle Exalytics In-Memory Machine, expanding the scope of in-memory analytics to include the full capabilities of the Oracle Database.

Posted October 15, 2014

SAP SE has announced the SAP Cloud for Planning solution, an enterprise performance management (EPM) solution designed around user experience and built for the cloud. The SAP Cloud for Planning solution will be built natively on SAP HANA Cloud Platform, the in-memory platform-as-a-service (PaaS) from SAP.

Posted October 14, 2014

Generally available today, EMC and Pivotal have announced the Data Lake Hadoop Bundle 2.0 that includes EMC's Data Computing Appliance (DCA), a high-performance big data computing appliance for deployment and scaling of Hadoop and advanced analytics, Isilon scale-out NAS (network attached storage), as well as the Pivotal HD Hadoop distribution and the Pivotal HAWQ parallel SQL query engine. The idea is to provide a turn-key offering that combines compute, analytics and storage for customers building scale-out data lakes for enterprise predictive analytics.

Posted October 14, 2014

ParStream has introduced an analytics platform purpose-built for the speed and scale of the Internet of Things (IoT). The ParStream Analytics Platform is designed to scale to handle the massive volumes and high velocity of IoT data and is expected to help companies generate actionable insights by enabling analysis with greater flexibility and closer to the source.

Posted October 14, 2014

Building on its data lake approach, Pivotal today announced the next step in this vision with the implementation of an architecture that builds upon disk-based storage with memory-centric processing frameworks.

Posted October 14, 2014

The new Dell In-Memory Appliance for Cloudera Enterprise is designed to provide customers with a processing engine combined with interactive analytics in a preconfigured and scalable solution, and will begin shipping Oct. 15, 2014.

Posted October 14, 2014

In its first server announcement since completing the IBM System x server acquisition, Lenovo has announced plans to collaborate with VMware. This alliance extends the 16-year development relationship between System x and VMware and broadens the partnership to include the full range of Lenovo's expanded server business.

Posted October 14, 2014

MongoDB has introduced enhancements to MongoDB Management Service (MMS), a cloud service to simplify operations for MongoDB deployments and reduce operational overhead.

Posted October 14, 2014

To help IT organizations extend the agility provided by continuous integration into continuous delivery, VMware also unveiled a major update of VMware vRealize Operations (formerly VMware vCenterOperations Management Suite. The new VMware vRealize Code Stream enables DevOps teams to deliver frequent, reliable software releases.

Posted October 14, 2014

Attunity has introduced Replicate 4.0 which provides high-performance data loading and extraction for Apache Hadoop. The solution has been certified with the Hortonworks and Cloudera Hadoop distributions.

Posted October 14, 2014

Today, many companies still have most of their transactional data in relational database management systems which support various business-critical applications, from order entry to financials. But in order to maintain processing performance, most companies limit the amount of data stored there, making it less useful for in-depth analysis. One alternative, according to a recent DBTA webcast presented by Bill Brunt, product manager, SharePlex, at Dell, and Unisphere Research analyst Elliot King, is moving the data to Hadoop to allow it to be inexpensively stored and analyzed for new business insight.

Posted October 13, 2014

Dataguise, a provider of security and data governance solutions for big data, has expanded its DgSecure platform to support Hadoop in the cloud, including full support for Amazon EMR (Elastic MapReduce). Additionally, big data cloud service providers, Altiscale and Qubole, have joined Dataguise's Big Data Protection Partner Program (BDP3) to leverage DgSecure in providing comprehensive discovery, protection and visibility to sensitive data for their cloud-based Hadoop customers.

Posted October 13, 2014

Two former Facebook engineers, Bobby Johnson and Lior Abraham, and former Intel engineer, Ann Johnson, have formed Interana to address what they say is an analytics void in event data. Espousing the philosophy that event data holds the key business metrics that companies care about most, Interana's solution is a database that has been specifically designed for event time data. Many methods in the past involved using general-purpose systems which were not designed to answer the types of questions posed by event data and it also took days to process, according to the company.

Posted October 13, 2014

Splunk, which provides software for machine-generated big data analysis, has announced Splunk Enterprise 6.2, Splunk Mint, and Splunk Hunk 6.2. "What we are doing with this release is fundamentally broadening the number of users that can do advanced analytics," stated Shay Mowlem, VP, product marketing at Splunk.

Posted October 13, 2014

IBM is adding new analytics capabilities to the mainframe platform, helping enable better data security and providing clients with the ability to integrate Hadoop big data. By applying analytic tools to business transactions as they are occurring, mainframe systems can enable clients to have true real-time insights. With the analytics on the System z platform, clients can also incorporate social media into their real-time analytic.

Posted October 13, 2014

GT Software has added enhancements to its flagship Ivory Service Suite line, incorporating greater support for big data elements and messaging formats.

Posted October 13, 2014

IBM, which has made a billion-dollar investment to broaden the use of cognitive computing, is announcing the launch of Watson World HQ today at 51 Astor Place. IBM said it chose NYC's Silicon Alley for Watson World to tap into the ecosystem of talent and capital centered around New York University, Columbia University, CUNY and Cooper Union, as well as venture capital firms and an expanding tech startup and developer community. Starting now, Watson's cognitive services and tools will be available to all users of Bluemix, IBM's open, cloud-based platform for mobile and web app development.

Posted October 13, 2014

Teradata is introducing Teradata Loom 2.3, a platform that provides integrated metadata management, data lineage, and ata wrangling for enterprise Hadoop. Teradata has also launched Teradata Cloud for Hadoop, a turnkey, full service cloud environment, and a broad technology and marketing partnership with Cloudera. "Increasingly, customers want a one-stop shop for their data analytics needs," said Chris Twogood, vice president of products and services at Teradata.

Posted October 09, 2014

SAP and BI provider Birst have formed a partnership to provide analytics in the cloud on the SAP HANA Cloud Platform. This collaboration intends to bring together the next-generation cloud platform from SAP with Birst's two-tier data architecture to provide instant access to an organization's data and help eliminate BI wait time.

Posted October 09, 2014

One feature of the big data revolution is the acknowledgement that a single database management system architecture cannot meet all needs. However, the Lambda Architecture provides a useful pattern for combining multiple big data technologies to achieve multiple enterprise objectives. First proposed by Nathan Marz, it attempts to provide a combination of technologies that together can provide the characteristics of a web-scale system that can satisfy requirements for availability, maintainability, and fault-tolerance.

Posted October 08, 2014

Today, as never before, public sector agencies have had to become more proactive and operate more like private sector businesses and are required to maintain an immense amount of data in order to make "just in time" decisions as well as to forecast for the long term. This is where business intelligence (BI) tools come in handy—not only to provide systems that facilitate the collection of data, but more importantly, providing a means to sift through the vast amounts of information for which the public sector is custodian.

Posted October 08, 2014

Organizations have been collecting data for years, but never before has there been such urgency to have it immediately available. The business need is pressing—decision makers need up-to-the-minute situational awareness in a volatile global economy.

Posted October 08, 2014

The data profession itself is splintering into an array of new specialties and tasks—away from database administration and programming and toward higher-level data science and business consulting tasks. A new survey of 281 data managers and professionals who are subscribers to Database Trends and Applications reveals that data managers and professionals are embracing their expanded roles and are motivated not just by money, but by the technology challenges and learning opportunities the database field now offers.

Posted October 08, 2014

In his keynote address, Thomas Kurian, executive vice president, Product Development, Oracle, showcased a brand new product, Oracle Big Data Discovery - a visual face to Hadoop - "that allows you to go to a browser, profile the data, explore the data, analyze the data, and do prediction and correlation," he said.

Posted October 08, 2014

Oracle CTO Larry Ellison took the stage at the Moscone Center to kick off the Oracle OpenWorld conference on Sunday evening. As he does each year, Ellison outlined major announcements for the week and explained the company's technology vision to set the tone for the conference. Confirming analysts' predictions that 2014 would be the year of the cloud for Oracle, Ellison said 2014 is an inflection point for the company. Ellison's key announcement was Oracle's upgraded platform-as-a-service capability: a fourth generation upward-compatible database.

Posted October 08, 2014

Hadoop RDBMS provider Splice Machine is partnering with LucidWorks to enable Splice Machine customers to access and analyze their unstructured data via LucidWorks Search.

Posted October 08, 2014

ConnectX, a collaboration of experts in aeronautics, mathematics, physics, engineering, security and business, says terrestrial computing is too limited, and aims to replace the current cloud computing, storage and network infrastructure being used for big data with a space-based supercomputing platform.

Posted October 07, 2014

Join DBTA for a webcast on Thursday, October 9, to learn about the key use cases, data replication strategies and methods for exploring data more efficiently through Hadoop.

Posted October 06, 2014

Approaching the fourth year of its planned five-year turnaround, HP today announced its intention to separate into two new publicly traded companies. One will include HP's enterprise technology infrastructure, software and services businesses, which will do business as Hewlett-Packard Enterprise, and the other will include HP's personal systems and printing businesses, which will operate as HP Inc. and retain the current logo.

Posted October 06, 2014

NoSQL database provider Couchbase has released a major new release of its NoSQL database Couchbase Server. The 3.0 release includes major architectural enhancements, new SDKs, and introduces a beta release of the ForestDB Storage Engine. "It is the biggest release ever for us with more than 200 features - small, medium, and large," said Couchbase CEO Bob Wiederhold.

Posted October 06, 2014

Aerospike Inc., provider of an in-memory NoSQL database, has announced a new startup special and trade-in program. The new program gives free access to the enterprise Edition of Aerospike with no limits on nodes, TPS and volume of data managed. To qualify, startups must have revenue of under $2 million and funding of under $20 million.

Posted October 03, 2014

Denodo Technologies has introduced Denodo Express, a free version of its data virtualization solution that provides the same technical features of as the enterprise version of the Denodo platform. The only difference is there are certain data restrictions when using the Express version.

Posted October 02, 2014

Mainframes represent some of the most important log data available, since they host the most mission-critical applications. However, according to the companies, the terabytes of data in more than 200 different log types produced by a typical mainframe system were previously inaccessible to Splunk software without significant work. Delivering on the technology alliance recently announced by Splunk and Syncsort to pull mainframe data into Splunk Enterprise and Splunk Cloud, Syncsort has introduced Ironstream.

Posted October 02, 2014

Cisco brought the fifth annual Data Virtualization Day to the Waldorf Astoria in New York City to share details about advancements coming in Cisco Information Server 7.0, the advantages of data virtualization, and the importance of the network. A key component of 7.0 release of Cisco Information Server which will be shipped next month, is the Business Directory, which will support greater access to data among more users on a self-service basis.

Posted October 01, 2014

Big analytics and visualization company Datameer is releasing version 5.0 of the company's data analytics application for Hadoop. "Our vision overall is to make data simple and accessible for everyone," said Matt Schumpert, director of product management with Datameer, about version 5.0 improvements.

Posted October 01, 2014

With data volumes spiraling out of control, companies are searching for less expensive and more efficient storage technologies. The factors that are involved with these decisions were discussed in a recent DBTA webcast, "Building a Memory-First Database Architecture for Speed and Scalability," presented by Joe McKendrick, Unisphere Research analyst, and Shane Johnson, product marketing manager with Couchbase, provider of Couchbase Server, a document-oriented NoSQL database.

Posted September 30, 2014

GridGain Systems, a provider of open source in-memory computing solutions, has launched the GridGain In-Memory Data Fabric. According to the company, the GridGain In-Memory Data Fabric is designed to serve the emerging needs of the real-time enterprise, which does not view datasets as historical events, but instead as actionable intelligence for immediate decision-making, and as a way to deal with changing market forces. Although analytics and transactional data functions have traditionally operated separately, companies now need to view them as part of the same "data fabric," and should implement strategies that are able to connect enterprise-wide data from all sources.

Posted September 30, 2014

"Digital disruption" driven by technology is having an impact on the decisions IT organizations are making, Mark Hurd told the audience at the Moscone Center in his first OpenWorld keynote as Oracle CEO. Hurd described the dramatic impact cloud, exploding data volumes, social, business, consumer sophistication, and mobility are having on IT organizations, and then interviewed executives about the individual challenges faced by each, including Yael Cosset of dunnhumby; Stephen Little of Xerox; Jamie Miller of GE; Filippo Passerini of P&G; Kimberly S. Stevenson of Intel Corporation; and Tim Theriault of Walgreen Co.

Posted September 29, 2014

Ron Bodkin founded Think Big Analytics to help organizations gain value from big data. Before that, he was vice president of engineering at Quantcast where he led the data science and engineering teams deploying Hadoop and NoSQL for batch and real-time decision making. In this interview, Bodkin who is CEO of Think Big, now a Teradata company, discusses the challenges organizations face in getting big data projects off the ground and what they need to consider when they embark on projects to leverage data from the Internet of Things and social media.

Posted September 25, 2014

In what is being heralded as an epic change at Oracle, Larry Ellison, the charismatic CEO of Oracle, is stepping down and will be succeeded by Safra Catz and Mark Hurd who will share the CEO position. Despite the widespread pronouncements that this is the end of an era and that Ellison is relinquishing his membership in the exclusive club of rock-star tech leaders that has included Apple's Steve Jobs and Microsoft's Bill Gates, some industry insiders believe that is simply not the case. Ellison, they say, will remain a force to be reckoned with at Oracle.

Posted September 25, 2014

A new survey of data managers and professionals who are subscribers to DBTA finds that, over the next 10 years, about two-fifths of respondents expect to leave the field - with about half of those people retiring and the other half moving on to other jobs in management or self-employment.

Posted September 25, 2014

The expectations for technology in business have never been higher—thanks to the consumerization of IT, disruptions across the tech stack, and the emergence of new technology-based business models. And the pressures facing corporate IT departments been greater. So, how can business leaders and CIOs maximize the value of their technology investments to drive competitive business advantage? By tackling three important areas.

Posted September 25, 2014

Pages
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15

Sponsors