Infobright 4.0 with DomainExpert™ Super-Charges Web and Machine-Generated Data Analytics
Jun 14, 2011
Analytic Database Delivers Huge Gains in Ad-hoc Query Performance and Load Speed for Today’s Fastest-Growing Category of Big Data
Toronto, Ontario – June 14, 2011 – Infobright today announced Infobright 4.0, the industry’s first database with built-in intelligence for near-real time analysis of machine-generated data. Representing the fastest-growing category of Big Data, machine-generated data includes sources ranging from web, telcom network and call-detail records, to data from online gaming, social networks, sensors, computer logs, satellites, financial transaction feeds and more.
Infobright 4.0 is built on the latest breakthrough in analytics: DomainExpert technology. Developed exclusively by Infobright, this technology can slash query response time by using specific intelligence about machine-generated data to automatically optimize how data is stored and how queries are processed. In addition, users can add their own domain knowledge to extend this capability to match their specific needs.
"Infobright's DomainExpert technology delivers new capabilities that allow Pentaho’s data discovery users to analyze their data much more quickly," said Richard Daley, founder and CEO of Pentaho. "Since our technology is already tightly integrated with Infobright, these new capabilities will be available immediately for customers to generate data analytics faster, easier and more affordably.”
With its new Rough Query feature, Infobright 4.0 can also speed up query response time by a factor of 20 when users are looking for “needles in the haystack” within a large volume of data. In combination, Infobright’s very high rate of data compression and its Rough Query capability allow companies to store far more data history, yet drill down into the data in a fraction of the time of other databases.
“I really like Infobright's Rough Query idea because investigative analytics is often iterative. So the longer you can work just in RAM, the better off you are,” said Curt Monash, president of Monash Research and editor and publisher of the DBMS 2 blog.
Also introduced is the new Distributed Load Processor (DLP), an add-on product for Infobright® Enterprise Edition (IEE). DLP scales load speed linearly across multiple servers and provides connectivity to a Hadoop cluster, enabling users to more quickly transform raw data into analytic intelligence. The product is designed for companies needing near-real time loading of very large volumes of data. With DLP, load speeds of over 2TB per hour into a single database table can be achieved.
Infobright 4.0 – New Capabilities
• DomainExpert technology: By adding intelligence about a particular data domain—such as web, financial services or telecom—the Infobright database is automatically optimized for analytic performance. In release 4.0, online data types such as email addresses, URLs and IP addresses are included, and users can also easily add their own domain intelligence to meet their unique needs.
• Rough Query: Rough Query provides near-instantaneous results for users who need to drill down into very large volumes of data. Rather than execute a long-running query to find a specific answer, Rough Query enables a user to narrow down the results in an iterative manner, with sub-second response time, before the full query is run. The total query time can be reduced by 20 times using this capability.
• Distributed Load Processor (DLP): DLP delivers load speeds that scale linearly as it processes and compresses data using multiple remote servers, then transfers the compressed data to the IEE database. Load performance gains result because CPU-intensive data compression is distributed across multiple machines, and query performance is enhanced by offloading this work from the database server.
• Hadoop Connector: Users looking to leverage Hadoop’s large-scale distributed batch processing benefits with Infobright’s fast ad-hoc analytic capabilities can do so easily using the Hadoop connector within the DLP. This tool provides a simple way to extract data from a Hadoop cluster and load it into IEE at very high speeds.
“Infobright’s goal is to give users immediate access to their data, the flexibility to do any kind of analysis without IT or DBA intervention, and the scalability to sustain high performance whether they are dealing with information within their own enterprise or Big Data in the Cloud,” said Don DeLoach, CEO of Infobright. “Our 4.0 release is another major step forward, as it extends the intelligence built into the system to specifically address the challenges companies of all sizes are facing in extracting useful and timely insight from the data now relentlessly churned out 24/7 by smart devices, sensors, real-time logs and a variety of other sources.”
Infobright Release 4.0 will be generally available within the next 30 days. The company will host two webinars on release 4.0 on June 22. For more information, go to http://www.infobright.com/Event/webinar_iee_4.0_super-charges_data_analytics/.
Infobright’s high-performance database is the preferred choice for applications and data marts that analyze large volumes of “machine-generated data” such as Web data, network logs, telecom records, stock tick data and sensor data. Easy to implement and with unmatched data compression, operational simplicity and low cost, Infobright is being used by enterprises, SaaS and software companies in online businesses, telecommunications, financial services and other industries to provide rapid access to critical business data. For more information, please visit http://www.infobright.com or join our open source community at http://www.infobright.org.
All names referred to are trademarks or registered trademarks of their respective owners.
Kickstart Consulting for Infobright
“Each month we process and analyze data generated by 20 billion online transactions. We are pleased by Infobright’s performance and the fact that we now can get answers to questions…
A New Approach
The Analytic Data Warehouse
Traditional data warehouse products put a tremendous burden on IT in order to create and maintain an environment that will allow users to query against large volumes of data.