Tuesday, March 22, 2011

Thoughts on Big Data and Data Virtualization

Big Data Analysis in Relationship to Queplix Data Virtualization Solution
On the plus side for obtaining IT and business alignment, more companies are beginning to combine business and information management responsibilities in a single role, carried out by a single person, rather than a “business and IT partnership” with two people, two hierarchies and two sets of reporting relationships. Gartner expects 20 percent of companies to employ business information managers by 2013, compared with 5 per cent in 2009.
-       Massive Data News in the report from April 2010
 
Here are the next ten things you should know about big data:
1.    Big data means the amount of data you’re working with today will look trivial within five years.
2.    Huge amounts of data will be kept longer and have way more value than today’s archived data.
3.    Business people will covet a new breed of alpha geeks. You will need new skills around data science, new types of programming, more math and statistics skills and data hackers…lots of data hackers.
4.    You are going to have to develop new techniques to access, secure, move, analyze, process, visualize and enhance data; in near real time.
5.    You will be minimizing data movement wherever possible by moving function to the data instead of data to function. You will be leveraging or inventing specialized capabilities to do certain types of processing- e.g. early recognition of images or content types – so you can do some processing close to the head.
6.    The cloud will become the compute and storage platform for big data which will be populated by mobile devices and social networks. 
7.    Metadata management will become increasingly important.
8.    You will have opportunities to separate data from applications and create new data products.
9.    You will need orders of magnitude cheaper infrastructure that emphasizes bandwidth, not iops and data movement and efficient metadata management.
10.  You will realize sooner or later that data and your ability to exploit it is going to change your business, social and personal life; permanently.
 David Vellante in Big Data on February 16, 2011

Queplix VDM solution provides a data management solution continuum, starting from data integration of multiple disperse data sources to Master Data Management. All in a single “dashboard” view. We have an automated NoSQL, object oriented representation of business objects, abstracted from multiple sources and described as metadata repository. QVDM is a persistent solution that operates with minimum disruption to the sources in automated fashion.

QVDM offers today enhanced Data Alignment, Data Quality and Data Enrichment, proactive Data Stewardship interface and Global Data Dictionary. As such, QVDM is a full-spectrum data management solution. One of our strengths is in our ability to intelligently identify and described business objects from a variety of data sources, using our Application Software Blades™. In doing so, we eliminate the need to deal with proprietary data storage formats and the need to copy large amounts of data in order to make it available for analytics.

The Big Data industry has been developing rapidly recently, even though the technology was created years back (Google Big Table, etc.) The goal of Big Data is to be able to store and processlarge amount of data for analytical purposes using Map/Reduce technology. The big data technology by itself does not provide analytical engine, but rather enables it to operate on large volumes of data. The big data storage vendors (i.e. Hadoop) provide flat non-relational storage facility and the multi-processing engine to address BI queries for large data volumes. It is not possible to achieve the same performance using traditional RDBMS.

The most obvious synergy between QVDM and Big Data technologies is in the noSQL approach to data management. Big Data vendors pursue common goal which is to enable BI solutions to work on large volumes. Let’s consider an example of Jaspersoft BI:
“Jaspersoft’s vision goes well beyond Big Data. Our modern architecture and agnostic data source support is tailored for the cloud, from IaaS to PaaS, either public or private variations. In particular, NoSQL support puts us in the driver’s seat to become the de facto embedded standard for reporting and analysis within PaaS cloud environments.”
— Brian Gentile, CEO of Jaspersoft.
Jaspersoft’s BI engine can be deployed on top of Hadoop and in order to work effectively with large data sets it needs to utilize abstraction of business entities. Queplix Virtual Data Manager is built on the noSQL architecture and can provide the abstraction required fro Jaspersoft and other BI vendors today.

QVDM architecture works natively with Big Data engine by abstracting the data from the siloed sources and can optionally be used to migrate the data to the Big Data storage like Hadoop. With QVDM it is now possible to create the abstracted metadata layer and then use it to recreate the business objects in Hadoop in order to copy/move the data and therefore enable BI to work on Big Data engines. In other words, through Queplix Hadoop application Blade, we can now enable BI vendors is to virtualize Big Data repositories and provide persistence. 

New NoSQL, Hadoop and Big Data Webinar!

http://campaign.r20.constantcontact.com/render?llr=ve7cot47&v=001QhuZciaLI3WEy4uhQOfPFH99H8ehg-2W5nMFMhLZuBu68BVt84cOSzI2K2Cb3gTsQauqh3wvJSQotBCwhVo50seC-SUKMVAa-xDqHXcYTJcqTtthraKsb809f6gzL3l8OWJYZVrQ2o4q3sReu7Cmq4Xq8R8T2cBIoSQ6dmIly0hNL2FJG-8q7N3qQHpGFsgGzDUi86nC1D6gup7-Yru91brAYl-LwyHjcDbNytmSVj1_wxXZgZToYrFxGHRkec1Pt9ZSDDFeJlGae8FXmrzQYzhOCAMF_5LEOrvRdEGz3tw%3D

Thursday, February 3, 2011

The Chief Executive Officer of Queplix, Mark Cashman, shares his view and insights on the state of Queplix and trends within technology and industry that are shaping them.

The Chief Executive Officer of Queplix, Mark Cashman, shares his view and insights on the state of our business and trends within technology and industry that are shaping them.

A Market Point of View

Written by Mark Cashman Tuesday, February 01, 2011 11:11 PM
Some things should make you stand up and take notice; data virtualization is one of those things.  Certainly virtualization as a technology or category is no longer new.  But data virtualization is one of the emerging segments in virtualization that has now reached critical mass.  All of the same drivers make it compelling - lower cost, lower risk and higher return on investment.
Data virtualization is the process of identifying, packaging, moving and ultimately storing your data without physically making additional copies throughout the process.  Data virtualization creates a layer between applications and their data so that you can use the data, to your benefit.  This virtual layer allows you to unlock your data from proprietary applications to get the utility you need.  Data virtualization separates the applications from your data such that you can use the data in new ways.
Data virtualization has been cast by some as a replacement to ETL, EAI or EII.  It is far more substantial - data virtualization is enabling you to do things you have not been able to do before without a tremendous amount of work, and with greater return on investment and lower risk.
As a proof point I would suggest doing a little research on which companies are now marketing themselves as data virtualization vendors.  Here you  you will find some household names from the technology company roster. These vendors understand the challenges the enterprise face and want to cash in on the action.  But, of course, it is often the case of finding a wolf in sheep’s clothing.  Their primary business model is wrapped around legacy technology.  For some, the complexity of their approach brings so much customization that there is more revenue in the services than in the software product.  Surely the model is upside down - we can help you fix it.
So, do your research and consider taking on change.
Stand up and take notice of the challenges you have faced to date and ask yourself does the existing way of managing my data really work? Can I achieve my goals and objectives using existing tools and technology, and how long will it take? The comfortable or easy bet didn’t get you to where you are today. Reach out explore other opportunities and realize another evolution surrounds you. Don’t become extinct like the dinosaurs!  The nimble market leaders across all industries know how to use new and innovative technologies that offer competitive advantage.  Faster development cycles, lower risk and greater return on investment are more important than ever.  Data virtualization can bring you all of this across applications for data discovery, business intelligence, data integration, data management, governance, quality, compliance and much more
In my next blog post I will explore how this new approach streamlines many of the tasks you perform today simply and more cost effectively.

Queplix Data Virtualization and Hadoop - marriage made in heaven

Recently, there have been a lot of news and development covering advances in parallel processing frameworks, such as Hadoop. Some innovative data warehouse software vendors are increasingly starting to research new development strategies the parallel processing offers. So far majority of the efforts were targeted at the improving the performance and optimization maps of the queries within the traditional physical data warehouse architectures. For example, traditional data warehouse vendors like Teradata joining the Hadoop movement and applying parallel processing to their physical DW infrastructures. Companies like Yahoo and Amazon are also spearheading map/reduce Hadoop adaption for large data scale analytics.

I had been monitoring advances in the Hadoop front in particular, as I believe it will provide a convergence grounds for our products and a new development direction for Queplix Data Virtualization. Data virtualization and Hadoop are born out of the same premise – provide data storage scalability and ease of information access and sharing and I see how the two technologies complement each other perfectly.

Hadoop’s data warehouse infrastructure (Hive) is what we are researching now to integrate with Queplix Data Virtualization products. Hive is a data warehouse infrastructure built on top of Hadoop that provides tools to enable easy data summarization, adhoc querying and analysis of large datasets data stored in Hadoop files. It provides a mechanism to put structure on this data and it also provides a simple query language called Hive QL. Queplix Data Virtualization will soon utilize the flexibility of its object-oriented data modeling combined with massive power of the Hadoop parallel processing to build virtual data warehouse solutions. Imagine the analytical performance of such virtual data warehouse solution created by using the Virtual Metadata Catalog and Virtual Entities in its base as organizational and hierarchal units (instead of traditional tables and columns and SQL-driven access).  Such “virtual” data warehouse solution would be a perfect fit for large scale operational and analytical processing, data quality and data governance projects with full power of Queplix heuristic and semantic data analysis. Today, data virtualization solutions are deployed by many larger enterprises to gain the visibility into the disperse application data silos without disrupting the original sources and applications; in the near future Data Virtualization and Hadoop-based virtual data warehouse solutions will be deployed in tandem to implement the full spectrum data management enterprise solutions ranging from larger-scale data integration projects (i.e. massive application data store mergers as a result of M&A between large companies) all the way to Virtual Master Data Management pioneered by Queplix. Such solutions will not only provide a better abstraction and continuity of business for the enterprise applications but also will utilize full power of parallel processing and will provide immense scalability to Queplix semantic data analytics and data alignment products.

Here are some of new and exciting ideas about Queplix is working on now: utilizing Hadoop for Virtual CEP (Complex Event Processing) within Queplix Virtual Metadata Catalog; generating “data steward” real time alerts using predictive data lineage analysis actually before the data quality problems start affect your enterprise applications; implementing Hadoop-based virtual data warehouse solutions to provide High Availability for large application stores that require massive analytics and semantic data processing; large scale Virtual Master Data Management initiatives involving enterprise-wide Customer or Product Catalog building; large-scale Business Intelligence projects based on Queplix Virtual Metadata Catalog.
Watch this blog for new developments and advances of Queplix technology to integrate Hadoop and Data Virtualization as we make announcements throughout this year!

Thursday, January 27, 2011

Social Networks are coming to the Enterprise!

I am happy to write that Queplix just announced our social Blades for Facebook, Linked In and Google Contacts: Read more. What we are trying to do here is to bridge the social network information with the enterprise data stores. When we started working on that we had a lot of skepticism internally about the potential use case for  the social information within the enterprise. Having worked within large corporate structures and implemented hundreds of large scale secure Enterprise information systems - I had my own doubts. It all changed in a heartbeat when we got our justification from our potential customer. One of the leading US banks approached Queplix and requested.. social blades. Yes.. THE large F-500 BANK!! I guess the time for this is now and forward looking companies are starting to recognize the value of the information contained in social ranking, chatter and people's opinions spread out across the social universe but still very much inter-connected. Enterprises are becoming increasingly sensitive to their public image, the perceptions of them by their customers and employees. The delicate point of balance here of course is privacy. One thing is to share your personal information and what you had for breakfast this morning with the members of your family and your friends; it completely changes when you know someone at your company will look at it. So the lines must be drawn around what can be seen, shared and used by the enterprises by getting access to the social networks information. While this work is still in progress and we are threading very carefully here. In reality - we always rely on the specific social network privacy laws and enforce them within our Blades. You, as the information owner have to specifically approve the Queplix application "Blade" to be connected to you and further you can see what exactly is being shared and how.  I think the biggest value right now for the enterprise is to monitor their own social images within social networks, and not look at people. There is a lot of information out there about the company itself and its products, this is where we focus first when connecting the enterprise with the social information. And this is what Queplix social blades are all about.

Hot off the wire!

QUEPLIX ANNOUNCES NEW DATA VIRTUALIZATION WEBINAR SERIES

Free Webinars in Q1 2011 Highlight the Applications and Benefits of Data Virtualization for Data Quality, Data Integration, Data Governance and Master Data Management


SUNNYVALE, Calif., January 19, 2011 – Queplix™ Corp., a leader in data virtualization, today announced that it will hold a series of free webinars in Q1 2011 that explore the applications and benefits of data virtualization for a variety of data-related initiatives, from data integration to master data management. Featuring experts from Queplix as well as independent industry experts, the webinars will provide a broad perspective on issues associated with data integration and management.

Free Queplix Webinars Q1 2011
•    January 28, 2011 at 10 a.m. PST: Thinking of Master Data Management?  Don't Do It.  Find Out Why.
•    February 4, 2011 at 10 a.m. PST: Data Integration Technologies, Architectures and Differentiation
•    February 11, 2011 at 10 a.m. PST: Data Quality Essentials for Data Integration
•    February 18, 2011 at 10 a.m. PST: Automating Data Governance
•    February 25, 2011 at 10 a.m. PST: Introducing Virtual Master Data Management™
•    March 4, 2011 at 10 a.m. PST: Supercharge Your ETL Team with Virtual Data Manager™
•    March 11, 2011 at 10 a.m. PST: Integrating Netsuite® with Salesforce® - A Case Study
•    March 18, 2011 at 10 a.m. PST: Fast and Easy Data Migration with Virtual Migration Manager™

"People really want to know how data virtualization can help them improve performance and address the issues that are triggered as their organizations move to align business-critical customer, product and vendor data across their applications," said Michael Zuckerman, chief marketing officer for Queplix. "The series of webinars we hosted in the fourth quarter of 2010 attracted more than 1,000 registrants. We expect our next series of webinars will be equally popular, since they address basic issues such as data quality and data governance that are of prime importance to any organization contemplating data integration or master data management."  

The new webinars are the continuation of a series hosted by Queplix that discusses data virtualization technology and the advantages and business value it brings to many different applications and business functions. To register for the Queplix webinar, visit www.queplix.com.

Data virtualization enables companies to automatically and securely integrate cloud, SaaS and on-premise applications with speed and simplicity. This benefits application areas like customer relationship management (CRM), where cloud-based solutions such as SalesForce.com® require integration and ongoing data harmonization with on-premise applications – PeopleSoft®, Siebel®, SAP® enterprise resource planning (ERP) and many others. Queplix solutions can be implemented to enable consistent views and manageability of this vital information in days versus months. As a result, customers achieve far greater agility, cost savings and rapid return on their investment than with alternative technologies.

Tweet This
: Queplix Kicks Off New Year with Data Virtualization Webinar Series

Follow Queplix on Twitter at
: http://www.twitter.com/QueplixCorp

Resource library: For more information about data virtualization and application integration, visit Queplix Resource Library.

About Queplix Corp.
Queplix is a leader in data virtualization. Data virtualization enables our customers to automatically and securely integrate cloud, SaaS and on-premise applications and data with speed and simplicity. Uniquely, Queplix brings powerful data management to automatically harmonize the integrated applications and data for higher business value. Our customers benefit from greater cost savings and rapid return on investment. Queplix has thousands of users around the world, serving customers such as The Home Depot, Homesite and Sony Ericsson. Please visit http://www.queplix.com for more information.

# # #

Queplix and the Queplix logo are trademarks of Queplix Corp. All other brands, products, or service names are or may be trademarks or service marks of their respective owners.