For generations, enterprises have used IBM data management products to store and manage transactional and analytical data. They rely on IBM database products like DB2 Mainframes, IBM Informix, and DB2 on LuW to store mission critical data. A huge amount of analytical data is also managed by IBM analytical databases like DB2 and IBM Netezza.
With Netezza moving into its sunset phase, Databricks is a modern, state-of-the-art alternative for enterprises. The challenge is to ensure the operational data travels in an easy, effective, and seamless way to the cloud. Databricks Lakehouse helps companies consolidate ETL, streaming, BI and data science workloads onto the same platform, by combining the performance and governance of the data warehouse with the flexibility of the data lake.
In this webinar, learn from our experts how to architect modern data pipelines to consolidate data from multiple IBM data sources into Databricks Lakehouse, using the state-of-the-art replication technique—Change Data Capture (CDC).
They will walk you through what CDC is, the challenges involved in building scalable and reliable CDC pipelines, why enterprises need CDC, and how to implement production-level CDC pipelines from IBM Mainframe, Netezza, and Informix to your Databricks Lakehouse. This webinar also features a live demo on how to move data from IBM DB2 to Databricks Lakehouse.
Join the discussion with Databricks and Arcion to glean valuable insights into:
- Adoption drivers and use cases for the Lakehouse Architecture
- The role of CDC in migrations and ongoing replications with IBM Mainframe data
- Must-have elements for a scalable & reliable CDC pipeline
- A demo of real-time CDC from IBM DB2 to Databricks Lakehouse
- Migration Strategy for DataStage ETL and Stored Procedure modernization
Can’t make it? Sign up anyway to receive the replay in your inbox.
Meet the speakers
Soham Bhatt is a Solutions Architect leading the EDW and ETL modernization practice at Databricks. Before Databricks he worked at Toyota Motors on building their next-generation Big Data Platform. Prior to that, his background was in building Enterprise Data Warehouses for Fortune 100 companies with Inmon and Kimball methodologies. In his current role, he loves guiding his customers with best practices as they migrate their EDWs to Data Lakehouses.
Rajkumar Sen is the founder and chief technology officer at Arcion Labs, the only cloud-native, CDC-based data replication platform. In his previous role as director of engineering at MemSQL, he architected the query optimizer and the distributed query processing engine. Raj also served as a principal engineer at Oracle, where he developed features for the Oracle database query optimizer, and a senior staff engineer at Sybase, where he architected several components for the Sybase Database Cluster Edition. He has published over a dozen papers in top-tier database conferences and journals and is the recipient of 14 patents.