April 15, 2020 By Virginie Grandhaye
Bharath Chari
3 min read

When people dream about becoming a baker or a pastry chef, they often think about the delicious pastries they’ll create, delighting their patrons with towering cakes wrapped in impossibly smooth fondant. But very rarely does anyone start off by thinking about the preparation involved in baking… Without being able to use freshly milled flour for baking, for example, you would actually never be able to eat a good piece of cake or a crusty loaf of bread. To produce those delicious pastries, a lot of preparation must happen before the actual baking process begins.

The same parallel can be made between AI and Data Integration. Let me explain:

The business challenge:

As an example, let’s examine a regional U.S. retailer who recently decided to modernize its supply chain management, including supply chain availability, fulfillment, and online cart. To accomplish this objective, the retailer decided to implement the Onera Decision Engine in Google Cloud Platform (GCP). The Onera Decision Engine is the cognitive operating system that harnesses AI to power the modern commerce supply chain, using advanced forms of cloud computing and machine learning technology to predict real-world behavior, generating nuanced insights and decisions. However, SaaS analytical systems won’t deliver maximum value if you can’t move, prepare, and deliver the right data in real time with high levels of throughput and performance. The retailer needed to publish 9 million messages per hour to the Google pub-sub messaging service during normal periods and 21 million messages per hour during peak periods.

That brings me to the point of accessing the data. Without Data, no AI….

Data everywhere…so where do you begin?

Data is typically spread across many systems:

  • Open data, on the Cloud.
  • Collected data (from social networks, or connected devices). Those can be on the public cloud, or local to your infrastructure.
  • Internal historical data of your company (customers data, historical orders…) Those are usually stored on a private Cloud, or on traditional storage systems, behind your firewall.

On top of this complexity, data can take many forms. They can be:

  • On traditional storage (Relational Databases).
  • Streamed data (for real time use cases)
  • Inside data lakes or data warehouses.
  • Inside corporate applications (like SAP).

This is a real issue: the most time-consuming task (80 percent) when driving an analytics project is “Collecting and organizing the data.”

Data Integration: Scalable data architecture for AI in the age of Covid-19

DataOps is an answer for shortening the cycle of making data available to the data scientist, where data integration capabilities (data transformation or extract, transform and load (ETL), data replication and data virtualization) play a vital role in providing access to high-quality data.

New SaaS analytical systems will not succeed without a robust and scalable data integration infrastructure for data movement, data integrationdata quality, and data governance that works across on premises and public and private clouds for AI. Similar to the fact that you will never bake a good piece of cake if you don’t have the right ingredients ahead of time

With the COVID-19 crisis, data integration is more critical than ever. With the defining moment we are all going through, companies need to think even more about their digital transformation – taking into account consumers’ behavioral changes, transforming their business models, and operationalizing AI and cloud first applications by transforming their infrastructure. Taking these  behavioral changes into account is a must for companies, should they wish to sustain and grow.

IBM DataStage: Deliver real time data for AI at scale and at high throughput

Going back to the example discussed: the US retailer attempted to implement their analytics and AI system using a vendor’s data integration product which was touted to be “built on cloud” but not able to exceed 1.2 million messages published to GCP per hour. In comparison, IBM’s multi-cloud data integration solution, IBM DataStage, is built on a massively parallel processing architecture was able to meet and exceed the client requirements. Moreover, IBM also demonstrated that the degree of parallel execution can be changed easily without having to make any changes to the job and can achieve higher rates of throughput simply by adding more hardware. The same job can publish 100 million records per hour or more just by running on additional cloud computing infrastructure. This DataStage example can be applied to many classes of SaaS analytical applications that require feeding decision engines in real time with unprecedented levels of throughput and performance.

If you are considering using AI in your business for other use-cases such as Enterprise Resource Planning (ERP), Customer Relationship Management (CRM), Human Resource Management (HRM) in retail, distribution, manufacturing and financial services, you should consider Data Integration as a mandatory step, to extract, load, transform, and deliver trusted data in real time for AI.

Read this Gartner report to find out how IBM is addressing these needs and has been positioned as a Leader in the Magic Quadrant for Data Integration tools for more than a decade.

Learn more about InfoSphere DataStage here.

Accelerate your journey to AI.

Was this article helpful?
YesNo

More from Analytics

IBM acquires StreamSets, a leading real-time data integration company

3 min read - We are thrilled to announce that IBM has acquired StreamSets, a real-time data integration company specializing in streaming structured, unstructured and semistructured data across hybrid multicloud environments. Acquired from Software AG along with webMethods, this strategic acquisition expands IBM's already robust data integration capabilities, helping to solidify our position as a leader in the data integration market and enhancing IBM Data Fabric’s delivery of secure, high-quality data for artificial intelligence (AI).  According to a Forrester study conducted on behalf of…

Fine-tune your data lineage tracking with descriptive lineage

4 min read - Data lineage is the discipline of understanding how data flows through your organization: where it comes from, where it goes, and what happens to it along the way. Often used in support of regulatory compliance, data governance and technical impact analysis, data lineage answers these questions and more.  Whenever anyone talks about data lineage and how to achieve it, the spotlight tends to shine on automation. This is expected, as automating the process of calculating and establishing lineage is crucial to…

Reimagine data sharing with IBM Data Product Hub

3 min read - We are excited to announce the launch of IBM® Data Product Hub, a modern data sharing solution designed to accelerate data-driven outcomes across your organization. Today, we're making this product generally available to our clients across the world, following its announcement at the IBM Think conference in May 2024. Data sharing has become the lifeblood of modern organizations, fueling growth and driving innovation. But traditional approaches to data sharing can often be a bottleneck constricting the seamless sharing of data.…

IBM Newsletters

Get our newsletters and topic updates that deliver the latest thought leadership and insights on emerging trends.
Subscribe now More newsletters