NoSQL Central Articles
There is a wealth of information, connections and relationships within the terabytes and petabytes of data being collected by organizations on distributed cloud platforms. Utilizing these complex, multi-dimensional relationships will be the key to developing systems to perform advanced relationship analysis. From predictive analytics to the next generation of business intelligence, "walking" the social and professional graphs will be critical to the success of these endeavors.
Posted February 02, 2011
DataStax (formerly Riptano), has unveiled DataStax OpsCenter for Apache Cassandra. A platform for managing, monitoring and operating enterprise Cassandra applications, DataStax OpsCenter provides Cassandra users with an advanced operations environment bundled with support for their real-time, high-volume, and low-latency applications. "Our customers are finding that anywhere that scale of data and real-time responsiveness are a challenge, Cassandra is proving to be the answer," said Matt Pfeil, CEO and co-founder, DataStax. "With DataStax OpsCenter we're giving these customers the confidence and control they need to deploy Cassandra for their most important applications."
Posted February 01, 2011
Jaspersoft today released software to support a variety of Big Data sources for business intelligence reporting. With the release of more than a dozen connectors as part of its open source "Big Data Reporting" project, as well as beta connectors for selected Big Data proprietary databases, Jaspersoft delivers native reporting for Hadoop, NoSQL and MPP analytic databases.
Posted January 25, 2011
The NoSQL acronym suggests it's the SQL language that is the key difference between traditional relational and newer non-relational data stores. However, an equally significant divergence is in the NoSQL consistency and transaction models. Indeed, some have suggested that NoSQL databases would be better described as "NoACID" databases - since they avoid the "ACID" transactions of the relational world.
Posted January 07, 2011
Karmasphere has introduced Karmasphere Analyst, productivity software for data professionals working with massive data sets. "Karmasphere is focused on providing software for developers and analysts to work with Big Data in Hadoop clusters. Over the past 12 months, we have been introducing products specifically focused on developers who are early adopters of Hadoop and are working on Big Data in Hadoop, but what we have heard time and time again from enterprise organizations is the need to unlock access to data in Hadoop clusters for enterprise analysts," Martin Hall, CEO of Karmasphere, tells 5 Minute Briefing.
Posted December 21, 2010
Because any database that does not support the SQL language is, by definition, a "NoSQL" database, some very different databases coexist under the NoSQL banner. Massively scalable data stores like Cassandra, Voldemort, and HBase sacrifice structure to achieve scale-out performance. However, the document-oriented NoSQL databases have very different architectures and objectives.
Posted November 30, 2010
Sybase, Inc., an SAP company, has announced the release of Sybase IQ 15.3 to beta customers. The new release introduces the PlexQ Distributed Query Platform, a massively parallel processing (MPP) architecture that accelerates highly complex queries by distributing work to many computers in a grid configuration. GA is planned for the first half of 2011.
Posted November 17, 2010
Calpont Corporation has announced InfiniDB Enterprise 2.0. According to Calpont, InfiniDB is a powerful and reliable platform for data professionals who need rapid and easy access to data, scalable data capacity, or the need to augment or replace their traditional RDBMS technologies. "Our technology is a pure 100% columnar architecture and we have scale out capabilities which means we support a massively parallel processing paradigm and those are really the two key things that make this new generation of analytic databases so powerful," Nick Ochoa, vice president of marketing, Calpont, tells 5 Minute Briefing.
Posted November 16, 2010
There has been a lot of interest lately in NoSQL databases and, of course, many of us have strong backgrounds and experience in traditional relational "SQL" databases. For application developers this raises questions concerning the best way to go. One recurring truth that eventually surfaces with all new software technologies is that "one size does not fit all." In other words, you need to use the right tool for the job, as each has its own strengths and weaknesses. In fact, a danger of many new architectural approaches is one of "over-adoption" - using a given tool to address a wide array of situations when originally they were designed for the specific problem domain in which they excel.
Posted November 09, 2010
Informatica Corporation, an independent provider of data integration software and Cloudera, a provider of Apache Hadoop-based data management software and services, yesterday announced at Informatica World 2010 that the two companies are partnering to provide customers with the solutions needed to address the challenges associated with managing large-scale data, including structured, complex and social data. Together, Informatica and Cloudera say they intend to bring the productivity benefits of the Informatica Platform to the data-intensive distributed computing capability of Hadoop.
Posted November 02, 2010
MarkLogic Corporation has introduced a major new version of its flagship product, MarkLogic Server, a purpose-built database for unstructured information targeted at customers in three verticals: media/publishing, government, and financial services segments.
Posted November 02, 2010
The relational database - or RDBMS - is a triumph of computer science. It has provided the data management layer for almost all major applications for more than two decades, and when you consider that the entire IT industry was once described as "data processing," this is a considerable achievement. For the first time in several decades, however, the relational database stranglehold on database management is loosening. The demands of big data and cloud computing have combined to create challenges that the RDBMS may be unable to adequately address.
Posted October 12, 2010
InterSystems Corporation has rolled out a new version of its Caché high-performance object database. The new release targets the growing demand by CIOs for economical high availability by introducing database mirroring, while also addressing Java developers' need for high-volume high-performance processing combined with persistent storage for event processing systems. Robert Nagle, InterSystems vice president for software development, recently chatted with DBTA about the release and the new features it offers. Commenting on the growing interest in NoSQL databases, Nagle observes that many of the beneficial characteristics people see in NoSQL are in fact true of Caché - a flexible data model and zero DBA cost. "But for us, what is unique is that it is not NoSQL, it is that it needs to be SQL - without the overhead of relational technology - because I think SQL is extremely important for almost every class of application that is deployed."
Posted October 12, 2010
In Greek mythology, Cassandra was granted the gift of prophesy, but cursed with an inability to convince others of her predictions - a sort of unbelievable "oracle," if you like. Ironically, in the database world, the Cassandra system is fast becoming one of the most credible non-relational databases for production use - a believable alternative to Oracle and other relational databases.
Posted October 12, 2010
NoSQL - probably the hottest term in database technology today - was unheard of only a year ago. And yet, today, there are literally dozens of database systems described as "NoSQL." How did all of this happen so quickly? Although the term "NoSQL" is barely a year old, in reality, most of the databases described as NoSQL have been around a lot longer than the term itself. Many databases described as NoSQL arose over the past few years as reactions to strains placed on traditional relational databases by two other significant trends affecting our industry: big data and cloud computing.
Posted August 10, 2010
Objectivity, Inc., a provider of data management solutions, has released the first version of its enterprise-ready distributed graph database product, following a successful beta program which began earlier this year. InfiniteGraph enables large-scale graph processing, data analytics and discovery, and supports the leading requirements of organizations seeking valuable connections in data and information, and building advanced systems and services around social networking, business intelligence, scientific research, national security and similar efforts.
Posted August 03, 2010
Quest Software, Inc. has introduced a beta program for Toad for Cloud Databases, a new data access and management tool for non-relational data stored in cloud databases, also known as NoSQL databases. Toad for Cloud Databases is intended to help users unlock data stored in the cloud by using the familiar SQL language or Toad's popular visual query and data access capabilities. Users can query and report on non-relational data, migrate data in both cloud and relational databases from one to the other, and create queries that combine the two.
Posted July 07, 2010
VoltDB, LLC, has begun shipping the VoltDB OLTP database management system that is intended to offer faster transaction processing capabilities due to lower overhead. VoltDB, developed under the leadership of Postgres and Ingres co-founder, Mike Stonebraker, is a next-generation, open source DBMS that, according to the company, has been shown to process millions of transactions per second on inexpensive clusters of off-the-shelf servers. The VoltDB design is based on an in-memory, distributed database partitioning concept that is optimized to run on today's memory-rich servers with multi-core CPUs. Data is held in memory (instead of on disk) for maximum throughput, which eliminates buffer management. VoltDB distributes data - and a SQL engine to process it - to each CPU core in the server cluster. Each single-threaded partition operates autonomously, eliminating the need for locking and latching. Data is automatically replicated for intra-cluster high availability, which eliminates logging.
Posted June 07, 2010
CodeFutures Corporation has introduced the latest version of its dbShards, which provides cost-effective database scalability and performance using database sharding. The company also announced that Familybuilder, a fast-growing family application on the internet is using dbShards to keep up with the demands placed on its high-volume MySQL database, while simultaneously scaling to meet anticipated future growth.
Posted April 20, 2010
If you spend any time at all reading IT trade journals and websites, you've no doubt heard about the NoSQL movement. In a nutshell, NoSQL databases (also called post-relational databases) are a variety of loosely grouped means of storing data without requiring the SQL language. Of course, we've had non-relational databases far longer than we've had actual relational databases. Anyone who's used products like IBM's Lotus Notes can point to a popular non-relational database. However, part and parcel of the NoSQL movement is the idea that the data repositories can horizontally scale with ease, since they're used as the underpinnings of a website. For that reason, NoSQL is strongly associated with web applications, since websites have a history of starting small and going "viral," exhibiting explosive growth after word gets out.
Posted April 07, 2010
The concept of database sharding has gained popularity over the past several years due to the enormous growth in transaction volume and size of business-application databases. Database sharding can be simply defined as a "shared-nothing" partitioning scheme for large databases across a number of servers, enabling new levels of database performance and scalability. If you think of broken glass, you can get the concept of sharding—breaking your database down into smaller chunks called "shards" and spreading them across a number of distributed servers.
Posted August 14, 2009
Database sizes have grown exponentially, with more than half of all databases in use globally projected to exceed 10TB by 2010, according to The Data Warehouse Institute. But as the size of data warehouses has exploded and business requirements have forced companies to conduct more ad hoc queries on those warehouses, response times have slowed, requiring increasingly expensive database hardware investments.
Posted May 15, 2008
The data explosion driving data warehouse equipment purchases in the last few years has just begun. Equipment proliferation already pressurizes data center energy requirements. Fortunately, a column-based analytics server can help companies with both kinds of green - the environment and money - by offering enormous energy and cost reductions while significantly boosting performance.
Posted February 15, 2008