Newsletters




Big Data

The well-known three Vs of Big Data - Volume, Variety, and Velocity – are increasingly placing pressure on organizations that need to manage this data as well as extract value from this data deluge for Predictive Analytics and Decision-Making. Big Data technologies, services, and tools such as Hadoop, MapReduce, Hive and NoSQL/NewSQL databases and Data Integration techniques, In-Memory approaches, and Cloud technologies have emerged to help meet the challenges posed by the flood of Web, Social Media, Internet of Things (IoT) and machine-to-machine (M2M) data flowing into organizations.



Big Data Articles

Helping to support enterprises' plans to implement SAP Business Suite 4 SAP HANA (SAP S/4HANA), BackOffice Associates, a provider of information governance and data modernization solutions, announced that its data management solutions are available to help enterprises orchestrate the data migration and modernization process for implementations of SAP S/4HANA.

Posted May 28, 2015

Cloudera is now offering support for Capgemini's new reference architecture for the SAP HANA platform and Cloudera Enterprise. "By bringing the power of Cloudera's enterprise data hub offering to the ecosystem in support of SAP HANA, we can enable Capgemini's clients to expand the amount of data they have within their environment in a cost-efficient manner," said Tim Stevens, vice president of corporate and business development at Cloudera.

Posted May 28, 2015

The certification enables Nimble Storage to participate in SAP's program for SAP HANA tailored data center integration using its certified solutions. Through participation in the program, customers can leverage their existing hardware and infrastructure components for their SAP HANA-based environments, providing further choice for organizations even when working in heterogeneous environments.

Posted May 28, 2015

CA Workload Automation Advanced Integration 1.0 for SAP Business Warehouse has received SAP certification. Specifically, the SAP Integration and Certification Center has certified that CA Workload Automation Advanced Integration 1.0 integrates with SAP Business Warehouse to provide a unified view for jobs running in both SAP and non-SAP applications.

Posted May 28, 2015

Embarcadero Technologies, a provider of software solutions for application and database development, has unveiled the new XE7 version of ER/Studio, its flagship data architecture suite.

Posted May 28, 2015

EMC, which enables businesses and service providers to transform their operations and deliver IT as a service, is acquiring Virtustream to form its new managed cloud services business. The all-cash transaction is expected to close in the third quarter of 2015, subject to customary closing conditions, is valued at approximately $1.2 billion. Virtustream brings to the EMC portfolio a managed cloud software and services capability — whether on or off premise — which EMC also intends to incorporate into the Federation Enterprise Hybrid Cloud Solution.

Posted May 27, 2015

Entrinsik is showcasing the new capabilities in Informer 4.5, the latest version of the company's reporting and business intelligence software, which will be released on June 8. "With 4.5 we're really building on our SQL capabilities and adding more advanced options for users that are reporting on SQL platforms," stated Tim Nicholson, director of Informer services at Entrinsik.

Posted May 27, 2015

Larry Okeson, the owner of Software Search, is retiring after 36 years of being at the helm, and will be succeeded by Caroline Okeson, his daughter-in-law and senior recruiter/branch manager for Software Search.

Posted May 27, 2015

Today, it's not enough for companies to collect as much data as possible in the hope of gleaning insight. Much of this information is useless and only serves to bog down corporate networks. Instead, companies need to focus on data integration and the ability to make the right data available to the right users, regardless of origin or platform. Here are 6 technologies to improve data integration.

Posted May 26, 2015

At Data Summit 2015 in New York City, Tony Shan, chief architect, Wipro, gave a talk on the key components of a successful big data methodology and shared lessons learned from real world big data implementations. According to Shan, there is an 8-step process for a big data framework with specific techniques and methods.

Posted May 26, 2015

Building on its native support for a variety of database platforms - including Apache Hive for Hadoop, MongoDB Enterprise, and traditional relational database management systems (RDBMS) - Embarcadero Technologies' data modeling tool, ER/Studio Data Architect version 10.0, has been certified by HortonWorks and Mongo DB Enterprise.

Posted May 21, 2015

SAP has announced the cloud edition of SAP Business Suite 4 SAP HANA (SAP S/4HANA), which adds new simplification and innovations across core business functions. In addition, the new release gives customers the opportunity to deploy real hybrid scenarios — combining on-premise and cloud solutions.

Posted May 20, 2015

Chartio, a cloud business intelligence service, is introducing a new solution called Data Stores for transforming and storing data for business intelligence in a more agile way. Data Stores will enable administrators to use Chartio's Data Pipeline to rapidly transform data and store it in the cloud, making it more useful and accessible for end users.

Posted May 20, 2015

Rosslyn Analytics, a provider of big data cloud technology, has announced it is one of the first to offer companies analytics as a service on Azure. The big data cloud analytics platform, powered by Azure, provides self-service management from source to analytics and enables business and IT users to interact with, change, and analyze data using a combination of self-service data integration, cleansing and enrichment tools and machine learning and visualization technologies.

Posted May 20, 2015

Oracle provides informational resources, including educational events, webcasts, and white papers.

Posted May 20, 2015

IOUG offers webcasts on a range of topics including security strategies, cloud computing, data sharing and many more.

Posted May 20, 2015

Oracle is collaborating with Mirantis to enable Oracle Solaris and Mirantis OpenStack users to accelerate application and database provisioning in private cloud environments via Murano, the application project in the OpenStack ecosystem.

Posted May 20, 2015

Oracle is shipping a new big data product called Oracle Big Data Spatial and Graph. Spatial and graph analytics has been available as an option for Oracle Database for more than 10 years, and with this introduction the company is bringing spatial and graph analytics to Hadoop and NoSQL.

Posted May 20, 2015

MapR Technologies, Inc., a provider of a distribution for Apache Hadoop, is including Apache Drill 1.0 in the MapR Distribution.

Posted May 19, 2015

DataStax, which provides an enterprise distribution of Apache Cassandra, has introduced DataStax Enterprise 4.7 (DSE 4.7), which the company says, is purpose-built for the stringent performance and availability demands of web, mobile and Internet of Things (IOT) applications. The new release includes advancements to integrated enterprise search, analytics, in-memory computing, and database management and monitoring to address the increasing shift toward mixed workload deployments.

Posted May 19, 2015

The shortage of skilled talent and data scientists in Western Europe and the U.S. has triggered the question of whether to outsource analytical activities. This need is further amplified by competitive pressure to reduce time to market and lower costs.

Posted May 19, 2015

As the excitement and opportunity provided by big data tools develop, many organizations find their big data initiatives originating outside existing data management policies. As a result, many concepts of formal data governance are either intentionally or unintentionally omitted as these enterprises race to ingest huge new data streams at a feverish pace in the hope of increased insight and new analytic value.

Posted May 19, 2015

Similar to the dot-com revolution, the Internet of Things is the culmination of radical advances in four core technology pillars.

Posted May 19, 2015

Google white papers have inspired many great open source projects. What has been missing until now, however, has been a way of bringing these technologies together such that any data-centric organization can benefit from the capabilities of each technology across its entire data center, and in new ways not documented by any single white paper. This is called the "Zeta Architecture."

Posted May 19, 2015

Business pressures, including cost reduction, scalability, and "just-in-time" application software implementation, are just some of the requirements prompting businesses to "cloudify" at least some aspect of their IT infrastructure.

Posted May 19, 2015

Data-driven companies continue to explore data management technologies that better unify operational, analytical, and other disparate or siloed data in a way that offers tangible business value and data management relief.

Posted May 19, 2015

The demand for effective data management is intensifying. At the same time, the database market has expanded into a wide array of solutions—from traditional relational database management systems to alternative databases such as NoSQL, NewSQL, cloud, and in-memory offerings.

Posted May 19, 2015

Data preparation is gaining considerable visibility as a distinct aspect of data management and analytics work.

Posted May 19, 2015

Just when you thought NoSQL meant the end of SQL, think again, and realize why you need to hold on to your relational database administrator like it was 1999. NoSQL has proven to be a resilient next-generation database technology for increasingly common internet-era specialized workloads. Now approaching a decade after its arrival on the scene, NoSQL is moving beyond architectural marvels to practical tools in the software development toolkit and, in that process, unveiling tried-and-true capabilities formerly known to be the scalpels of the enterprise relational database. Let's go back to the future and take a look at how the DBA is becoming as relevant as ever while NoSQL evolves for the enterprise.

Posted May 19, 2015

RedPoint Global was founded in 2006 by Dale Renner, Lewis Clemmens, and George Corugedo, who previously had worked together at Accenture. Based in Wellesley, Mass., RedPoint collaborates with clients around the world in 11 different verticals. "We have always been very focused on the data, and recognize that a lot of business problems live and die by the quality of the data," says Corugedo.

Posted May 19, 2015

Hadoop is contributing to the success of data analytics. Anad Rai, IT manager at Verizon Wireless, examined the differences between traditional versus big data at Data Summit 2015 in a session titled "Analytics: Traditional Versus Big Data." The presentation, which was part of the IOUG track moderated by Alexis Bauer Kolak, education manager at the IOUG, showed how big data technologies are helping data discovery and improving the transformation of information and knowledge into wisdom.

Posted May 14, 2015

The data lake is one of the hottest topics in the data industry today. It is a massive storage reservoir that allows data to be stored in its rawest forms. Hadoop Day at Data Summit 2015 concluded with a panel on everything data lake featuring James Casaletto, solutions architect for MapR, Joe Caserta, president and founder of Caserta Concepts, and George Corugedo, CTO with RedPoint Global Inc.

Posted May 14, 2015

With the influx of big data solutions and technologies comes a bevy of new problems, according to Data Summit 2015 panelists Miles Kehoe, search evangelist at Avalon Consulting, and Anne Buff, business solutions manager for SAS best practices at the SAS Institute. Kehoe and Buff opened the second day of Data Summit with a keynote discussion focusing on resolving data conundrums.

Posted May 14, 2015

To transform data into value, IT must move from thinking about what it does to data, and instead focus on business outcomes and what can be done with the data to advance the business, according to Edd Dumbill, vice president, strategy, Silicon Valley Data Science, who gave the welcome keynote at Data Summit 2015.

Posted May 14, 2015

If used correctly, machine data can provide a company a significant advantage in terms of understanding user and machine behavior, fighting cyber security risks and fraudulent behavior, service levels and customer behavior. In his talk at Data Summit 2015, Dejan Deklich, vice president, engineering platform and cloud at Splunk, discussed issues around machine data analysis and showcased some prominent use cases.

Posted May 13, 2015

In order to break down barriers in creating and storing data, understanding the modern data architecture is key. That was the focus of Mike Lamble, CEO at Clarity Solution Group, and Ron Huizenga's, product manager at Embarcadero Technologies, presentation at Data Summit 2015.

Posted May 12, 2015

Capgemini is extending its long-standing strategic partnership with SAP, allowing Capgemini to act as a single point of contact for customers globally, and delivering SAP products and support services through one consolidated framework. By signing a global value-added reseller (VAR) agreement with SAP, Capgemini is among a select group of global SAP partners that are part of the global program, which has specific entry requirements that include global reach, reseller capabilities and revenue targets.

Posted May 12, 2015

Splice Machine is partnering with Talend to enable customers to simplify data integration and streamline data workflows on Hadoop. Through this partnership, organizations building operational data lakes with Splice Machine can augment Talend's data integration technology with its data quality capabilities.

Posted May 12, 2015

Pentaho users will now be able to use Apache Spark within Pentaho thanks to a new native integration solution that will enable the orchestration of all Spark jobs. Pentaho Data Integration (PDI), an effort initiated by Pentaho Labs, will enable customers to increase productivity, reduce maintenance costs, and dramatically lower the skill sets required as Spark is incorporated into big data projects.

Posted May 12, 2015

HP has made multiple contributions to the OpenStack Kilo release, including new converged storage management automation and new flash storage technologies to support flexible, enterprise-class clouds. HP's storage contributions to the OpenStack Kilo release focus on two strategic goals.

Posted May 11, 2015

Teradata has made enhancements to the Teradata Database's hybrid row and column capabilities to provide quicker access to data stored on columnar tables and drive faster query performance. Other relational database management systems store data tables in rows or columns, and each method offers benefits, depending on the application and type of data. However, they have been mutually exclusive. Teradata's new hybrid row and column capabilities allow the best of both worlds.

Posted May 08, 2015

Software AG has made updates to its Terracotta In-Memory Data Management platform. New improvements to Terracotta Open Source Kit 4.3 include distributed storage and off-heap storage.The platform is used for boosting performance, scalability, and building real-time applications. Additionally, Terracotta helps developers leverage in-memory storage for current and emerging data workloads.

Posted May 07, 2015

Tableau's cloud analytic solution, Tableau Online, is being upgraded to version 9.0. The new release enables faster performance, and provides additional live database connection support, single sign-on, and other new features designed to help users do more with their data in the cloud. The new update brings a complete redesign of Tableau Online to deliver a faster, more scalable, resilient, and extensible platform with capabilities such as parallel queries, query fusion, vectorization and smarter query caches that will make Tableau Online as much as 10 times faster.

Posted May 07, 2015

When databases are built from a well-designed data model, the resulting structures provide increased value to the organization. The value derived from the data model exhibits itself in the form of minimized redundancy, maximized data integrity, increased stability, better data sharing, increased consistency, more timely access to data, and better usability.

Posted May 06, 2015

Pivotal has made updates to its big data suite that include upgrades to the Pivotal HD enterprise-grade Apache Hadoop distribution, which is now based on the Open Data Platform core, and performance improvements for Pivotal Greenplum Database.

Posted May 05, 2015

Pages
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159

Sponsors