Newsletters




Data Warehousing

Hardware and software that support the efficient consolidation of data from multiple sources in a Data Warehouse for Reporting and Analytics include ETL (Extract, Transform, Load), EAI (Enterprise Application Integration), CDC (Change Data Capture), Data Replication, Data Deduplication, Compression, Big Data technologies such as Hadoop and MapReduce, and Data Warehouse Appliances.



Data Warehousing Articles

At OpenWorld, Oracle's annual conference for customers and partners, John Matelski, president of the IOUG, and CIO for Dekalb County, Georgia, gave his perspective on the key takeaways from this year's event. Matelski also described the user group's efforts to help the community understand the value of Oracle's engineered systems and deal with the broad implications of big data, and how the IOUG is supporting Oracle DBAs in their evolving roles.

Posted December 12, 2012

Tervela Turbo is now certified on CDH4 (Cloudera's Distribution Including Apache Hadoop Version 4). Introduced in October, Tervela Turbo, a high-performance data movement engine, helps Cloudera customers implement mission-critical Hadoop systems with reliable data capture, high-speed data loading into HDFS, disaster recovery for Hadoop, and ETLT data warehousing. Tervela has also joined the Cloudera Connect Partner Program.

Posted December 12, 2012

Amazon Web Services Inc. has announced the limited preview of Amazon Redshift, a managed, petabyte-scale data warehouse service in the cloud, which aims to enable customers to increase the speed of query performance when analyzing data sets using the same SQL-based BI tools they use today. "Over the past 2 years, one of the most frequent requests we've heard from customers is for AWS to build a data warehouse service," says Raju Gulabani, vice president of Database Services, AWS.

Posted November 28, 2012

Jeff West, president of Quest International Users Group, joined by Jonathan Vaughn, Quest's executive director, talked with DBTA at Oracle OpenWorld about what's ahead for 2013. The group has launched smaller, product-concentrated events to support JD Edwards and PeopleSoft users' specific areas of interest, and expanded its range of online offerings for users who may not be able take advantage of in-person conferences. Plans are underway to help members learn about PeopleSoft 9.2 coming in March and to prepare for the looming end of support for JD Edwards World. As always, says West, Quest continues to help get information to members from Oracle and their peers. "It is always about return on investment and aligning IT with the business. That is always on the top of people's minds."

Posted November 27, 2012

MapR Technologies, Inc., provider of the MapR Distribution for Hadoop, has formed a partnership with Hadapt, which offers a data analytics platform for natively integrating SQL with Apache Hadoop. The partnership enables customers to leverage MapR's Hadoop distribution in conjunction with Hadapt's Interactive Query capabilities to analyze all types of data, structured, semi-structured and unstructured, in a single, enterprise platform. Partnerships such as the one with Hadapt enable a broad community of users to have access to Hadoop data while also leveraging the existing skill sets of those users, Jack Norris, vice president of MapR, tells 5 Minute Briefing.

Posted November 15, 2012

Cloudera, provider of Apache Hadoop-based software and services, announced the first big data management solution that allows batch and real-time operations on any type of data within one scalable system. Cloudera Enterprise Real-Time Query (RTQ), powered by Cloudera Impala, improves the economics and performance of large scale enterprise data management, allowing organizations to process data at petabyte scale and interact with that data in real time all on the same system.

Posted November 06, 2012

Open source software vendor Talend announced that it has added big data profiling for Apache Hadoop and support for NoSQL databases in the upcoming release of its integration platform, Talend v5.2. Data profiling, the process of evaluating the character and condition of data stored across the enterprise, is a critical step toward gaining control over organizational data, and is emerging as a big data best practice. "Profiling allows you to understand what you have in your Hadoop cluster and how this data can be used for your big data integration and management project," Yves de Montcheuil, Talend's vice president of marketing, tells 5 Minute Briefing.

Posted November 06, 2012

Attunity Ltd., a provider of information availability software solutions, is partnering with Teradata to offer Attunity Replicate for Teradata, a big data replication solution designed to enable loading of heterogeneous data to Teradata with high performance, efficiency and ease-of-use.

Posted October 25, 2012

Kognitio is allowing companies to download a fully functional copy of its software at no charge, and with no time restrictions. The company, which made the announcement at O'Reilly Strata and Hadoop World conference, said it is offering a full-featured, perpetual use license of up to 128 gigabytes without an expiration period or other limited functions normally found in "trialware." This capability gives companies the ability to do in-memory analytics on, for example, more than 500 million customer records at once.

Posted October 25, 2012

At SAP TechEd 2012 in Las Vegas, SAP unveiled its plans for SAP HANA Cloud, a next-generation cloud platform based on in-memory technology. As part of SAP HANA Cloud, the company also announced the general availability of SAP NetWeaverCloud, an open standards-based application service, and SAP HANA One, a deployment of SAP HANA certified for production use on the Amazon Web Services (AWS) Cloud, as the first offerings based on SAP HANA Cloud.

Posted October 24, 2012

The opportunities and challenges presented by big data are addressed in a new report summarizing the results of a survey of data managers and professionals who are part of the Independent Oracle Users Group. The survey was underwritten by Oracle Corporation and conducted by Unisphere Research, a division of Information Today, Inc. Key highlights from the survey include the finding that more than one out of 10 data managers now have in excess of a petabyte of data within their organizations, and a majority of respondents report their levels of unstructured data are growing.

Posted October 24, 2012

Survey respondents to the IOUG Big Data survey were entered into a drawing to win an iPad by providing their email addresses. The winner of the iPad in the recent IOUG Big Data study sweepstakes drawing was Thomas F. Lewandowski, an independent Oracle DBA.

Posted October 24, 2012

SAP AG announced the availability of the SAP Precision Retailing solution, which is designed to help companies reach consumers at the moment of decision. The enterprise solution delivers personalized offers and suggestions in real time to consumers across multiple channels, including mobile devices, in-store kiosks and websites. "The solution we just launched bridges the gap between the consumer at the point of decision who needs to make a purchase and the ocean of data - and we extract the needle in the haystack - that piece of information that will help the consumer make a decision," Herve Pluche, vice president, Retail Consumer Mobile Initiative, SAP Labs, tells 5 Minute Briefing.

Posted October 11, 2012

Attunity Ltd., a provider of information availability software solutions, has introduced Attunity Replicate for Exadata, a new data replication solution designed to enable quick loading of big data into and out of Oracle Exadata. The solution further expands Attunity's reach into the big data market and features targeted optimizations that result in high performance and rapid implementation.

Posted October 10, 2012

Are today's data systems — many of which were built and designed for legacy systems of the past decade — up to the task of moving information to end users at the moment they need it? And is this information timely enough? In many cases, there's a lot of work that still needs to be done before real-time information, drawn from multiple sources, becomes a reality. A new survey of 338 data managers and professionals who are subscribers to Database Trends and Applications reveals that real-time data access is still a distant pipe dream for at least half of the companies represented in the survey. The survey, conducted by Unisphere Research, a division of Information Today, Inc., in partnership with Attunity in March of 2012, finds that close to half of the survey respondents, 48%, report that relevant data within their organizations still take 24 hours or longer to reach decision makers. This suggests that much data is still batch-loaded overnight.

Posted October 09, 2012

The opportunities and challenges presented by big data are examined in a new survey of 298 data managers and professionals who are part of the Independent Oracle Users Group. The survey was underwritten by Oracle Corporation and conducted by Unisphere Research, a division of Information Today, Inc. According to this new survey, more than one out of 10 data managers now have in excess of a petabyte of data within their organizations, and a majority of respondents report their levels of unstructured data are growing. Fewer than one out of five respondents feel their IT infrastructure will be ready to handle this incoming surge of data. Protecting data overall is important, but unstructured data gets low priority at this time.

Posted September 26, 2012

Data management vendor Terracotta, Inc. has released BigMemory Go, the latest innovation in the BigMemory line that allows customers to put as much data in memory as desired to speed application performance at big data scale. The product is being offered via a free 32GB per instance production license that can be deployed on as many servers as desired.

Posted September 25, 2012

Infobright has introduced a database appliance purpose-built for the analysis of machine-generated data called Infopliance. Integrating Infobright's analytic database with dedicated server and storage hardware, the new appliance is intended to provide customers with a cost-effective, scalable solution for extracting insight from growing volumes of real-time and historical information generated by web logs, network logs, call detail records, security data, and sensor data. "We have seen a growing need among customers who are going from one terabytes to two terabytes, to 15 terabytes, to 30 terabytes. They are experiencing very explosive data growth, and we began to realize that there would be a market for delivering a purpose-built machine-generated data appliance," Don DeLoach, president and CEO of Infobright, tells 5 Minute Briefing.

Posted September 24, 2012

The first computer program I ever wrote (in 1979, if you must know) was in the statistical package SPSS (Statistical Package for the Social Sciences), and the second computer platform I used was SAS (Statistical Analysis System). Both of these systems are still around today—SPSS was acquired by IBM as part of its BI portfolio, and SAS is now the world's largest privately held software company. The longevity of these platforms—they have essentially outlived almost all contemporary software packages—speaks to the perennial importance of data analysis to computing.

Posted September 19, 2012

In recent years, the networks of developers, integrators, consultants, and manufacturers committed to supporting database systems have morphed from one-on-one partnerships into huge ecosystems in which they have become interdependent on one another, and are subject to cross-winds of trends and shifts that are shaping their networks. Nowhere is this more apparent than the huge ecosystem that has developed around Oracle. With Oracle's never-ending string of acquisitions, new functionality, and widespread adoption by enterprises, trends that shape this ecosystem are certain to have far-reaching effects on the rest of the IT world. Concerns that percolate through the ecosystem reflect — and influence — broad business concerns. New paradigms — from cloud computing to big data to competing on analytics — are taking root within the Oracle ecosystem long before anywhere else.

Posted September 19, 2012

SAP AG introduced a new solution to help organizations gain real-time insights into market trends and customer sentiment. The SAP rapid-deployment solution for sentiment intelligence with SAP HANA is intended to allow users to analyze customer sentiment from social networking sites, communities, wikis, blogs and other sources, and combine the information with CRM data. Customers that have had success getting started with big data analytics are the ones that have set out to solve a very specific use case or set out to solve a specific problem, David Jonker, director of marketing for database and technology at SAP, tells DBTA. "The rapid deployment solution for sentiment intelligence does exactly that."

Posted August 23, 2012

Pentaho's Business Analytics 4.5 is now certified on Cloudera's latest releases, Cloudera Enterprise 4.0 and CDH4. Pentaho also announced that its visual design studio capabilities have been extended to the Sqoop and Oozie components of Hadoop. "Hadoop is a very broad ecosystem. It is not a single project," Ian Fyfe, chief technology evangelist at Pentaho, tells DBTA. "Sqoop and Oozie are shipped as part of Cloudera's distribution so that is an important part of our support for Cloudera as well - providing that visual support which nobody else in the market does today."

Posted August 23, 2012

Symantec Corp. has partnered with Hortonworks to introduce the new Symantec Enterprise Solution for Hadoop, providing a scalable, resilient data management solution for handling big data workloads. The add-on solution for Symantec's Cluster File System enables Symantec customers to run big data analytics on their existing storage infrastructure.

Posted August 14, 2012

Data warehousing is undergoing the most radical transformation seen since it was first conceived in the 1970s, and brought to market in the late 1980s and 1990s. One reason for this transformation is that data warehouses are on the front lines of the big data explosion. Findings from a new survey of IOUG members indicate that while most companies have well-established data warehouse systems, adoption is still limited within their organizations. This survey, underwritten by Oracle Corporation and conducted by Unisphere Research, a division of Information Today, Inc., included input from 421 data managers and professionals.

Posted August 08, 2012

Syncsort, a global leader in high-performance data integration solutions, has certified its DMExpress data integration software for high-performance loading of Greenplum Database. Syncsort has also joined the Greenplum Catalyst Developer Program. Syncsort DMExpress software delivers extensive connectivity that makes it easy to extract and transform data from nearly any source, and rapidly load it into the massively parallel processing (MPP) Greenplum Database without the need for manual tuning or custom coding. "IT organizations of all sizes are struggling to keep pace with the spiraling infrastructure demands created by the sheer volume, variety and velocity of big data," says Mitch Seigle, vice president, Marketing and Product Management, Syncsort.

Posted July 25, 2012

Datameer has announced a new release of its big data analytics solution, which combines data integration, analytics and visualization of any data type in one application. The new capabilities offered in Datameer 2.0 are in two main categories, Joe Nicholson, vice president of marketing, Datameer, tells DBTA. One is adding new functionality and the other is bringing Hadoop to the desktop with Hadoop natively embedded in two of three new editions of the application.

Posted June 28, 2012

Lucid Imagination, a developer of search, discovery and analytics software based on Apache Lucene and Apache Solr technology, has unveiled LucidWorks Big Data, a fully integrated development stack that combines advantages of multiple open source projects including Hadoop, Mahout, R and Lucene/Solr to provide search, machine learning, recommendation engines and analytics for structured and unstructured content in one solution available in the cloud. "With more and more companies being challenged by the explosive growth of information, as has been widely reported, the vast majority of that content is unstructured or semi structured text, and traditional business intelligence or traditional analytics methodologies don't come close to addressing the vast percentage of content," Paul Doscher, CEO of Lucid Imagination, tells DBTA.

Posted June 28, 2012

Data analytics vendor Teradata and information management software provider Kalido have introduced a new joint solution that they say will allow customers to build or expand a data warehouse in 90 days or less, providing deeper analytics to users for improved business decision-making. This solution combines the Teradata Data Warehouse Appliance with the Kalido Information Engine, providing customers with a streamlined data consolidation tool that aggregates disparate data into a single unified platform.

Posted June 28, 2012

MapR Technologies will make its distribution for Hadoop available on Google Compute Engine. The combination of the new Google service and the MapR distribution is intended to enable customers to quickly provision large MapR clusters on demand and to take advantage of the scalability of a cloud-based solution. "Off-premise, on-demand computing is an important part of the future for Hadoop," says John Schroeder, CEO and co-founder of MapR Technologies. "MapR is solidifying that future by partnering with Google and leveraging their cost-effective, high performance and scale-out infrastructure."

Posted June 28, 2012

SAP marked the 1-year anniversary of the SAP HANA platform becoming generally available. To celebrate the occasion, SAP AG announced the launch of the SAP HANA Distinguished Engineer program. The new program is focused on promoting SAP HANA expertise in the market and is intended to support a new group of community-driven, hands-on HANA technical professionals.

Posted June 27, 2012

Connotate, Inc., a provider of solutions that help organizations monitor and collect data and content from the web, is partnering with Digital Reasoning, which enables unstructured data analytics at scale, to provide a solution that creates actionable intelligence from fact-based analysis of big data.

Posted June 26, 2012

IBM has introduced a new analytics appliance that is intended to allow organizations to analyze up to 10 petabytes of data in minutes, helping them uncover patterns and trends from large data sets, while meeting compliance mandates. The new IBM Netezza High Capacity Appliance addresses a growing challenge: Banks, insurance companies, healthcare organizations and communications services providers are required by industry regulators to retain massive amounts of data - in some cases up to a decade. And, as data retention laws continue to evolve, organizations are faced with the need to store and analyze ever-expanding "big data" sets that may not be directly related to daily operations, yet still hold potential business value.

Posted June 26, 2012

Companies are scrambling to learn all the various ways they can slice, dice, and mine big data coming in from across the enterprise and across the web. But with the rise of big data — hundreds of terabytes or petabytes of data — comes the challenge of where and how all of this information will be stored. For many organizations, current storage systems — disks, tapes, virtual tapes, clouds, inmemory systems — are not ready for the onslaught, industry experts say. There are new methodologies and technologies coming on the scene that may help address this challenge. But one thing is certain: Whether organizations manage their data in their internal data centers, or in the cloud, a lot more storage is going to be needed. As Jared Rosoff, director of customer engagement with 10gen, puts it: "Big data means we need ‘big storage.'"

Posted June 13, 2012

Kalido, a provider of agile information management software and Gold level member in Oracle PartnerNetwork (OPN), has achieved Oracle Exadata Optimized status, demonstrating that Kalido Information Engine 9.0 has been tested and tuned on Oracle Exadata Database Machine to deliver speed, scalability and reliability to customers. The new optimized status signifies that Kalido's application which is certified to run on Exadata has now been optimized, meaning that it is taking advantage of the features in Exadata, including the memory, and the tuning and performance capabilities, and that as a result, Kalido is able to give customers better performance of its application running on the Exadata stack, says Bill Hewitt, Kalido president and CEO.

Posted June 13, 2012

Oracle has announced the Sun ZFS Backup Appliance, an integrated, high performance backup solution for Oracle engineered systems, including the Oracle Exadata Database Machine, Oracle Exalogic Elastic Cloud and Oracle SPARC SuperCluster T4-4. According to Oracle, the Sun ZFS Backup Appliance delivers up to 20TB per hour full backup and up to 9.4TB per hour full restore throughputs, the fastest published recovery rates among general purpose storage systems for Oracle engineered systems data protection. Oracle's Sun ZFS Backup Appliance is available in two configurations, High Performance and High Capacity, and comes pre-racked and cabled, helping to eliminate hardware tuning.

Posted June 13, 2012

Hortonworks, a commercial vendor promoting the innovation, development and support of Apache Hadoop, has announced the general availability of Hortonworks Data Platform (HDP) 1.0 which is intended to make Hadoop easy to consume and use in enterprise environments. "With the general availability of Hortonworks Data Platform 1.0, Hortonworks is delivering on its promise to make Apache Hadoop an enterprise viable data platform," says Rob Bearden, CEO of Hortonworks.

Posted June 13, 2012

Cloudera has unveiled the fourth generation of its flagship Apache Hadoop data management platform, Cloudera Enterprise. Cloudera Enterprise 4.0 combines the company's Cloudera Manager software with expert technical to provide a turnkey system for deploying and managing Hadoop in production. The company also announced the general availability of CDH4 (Cloudera's Distribution Including Apache Hadoop, version 4), resulting from the successful completion of a beta program among its enterprise customers and partner ecosystem and the contributions of Cloudera's engineering team and the greater Apache open source community.

Posted June 06, 2012

Pages
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46

Sponsors