LEARNing resources

Database Replication and Migration Articles

Explore powerful data mapping tools that simplify the process of transforming, visualizing, and integrating data across diverse systems. Choose the right data mapping software for your needs.
24
 min read

14 Best Data Mapping Tools for 2023

Explore powerful data mapping tools that simplify the process of transforming, visualizing, and integrating data across diverse systems. Choose the right data mapping software for your needs.
Explore our curated selection of the top 10 ETL (Extract, Transform, Load) tools for 2023. Find the perfect solution to streamline your data integration and management processes, ensuring efficiency and reliability in your data-driven operations.
23
 min read

10 Best ETL Tools of 2023

In this blog, we will take a look at the inner workings of ETL tools and explored the top 10 ETL tools of 2023. These tools ranged from enterprise-grade solutions like Arcion and IBM Infosphere DataStage to more accessible, user-friendly platforms like Blendo and Integrate.io.
Stay ahead in the world of data analytics with our list of the 10 best big data tools for 2023. Explore the latest technologies and platforms that empower you to harness the potential of big data for actionable insights.
28
 min read

10 Best Big Data Tools of 2023

In this comprehensive guide, we've explored the definition of Big Data, the range of tools available, their functionalities, key use cases, and tips for selecting the best fit for your needs.
Looking to migrate from Postgres to Oracle? Discover the seamless and efficient transition with our expert guide. Get step-by-step instructions, best practices, and key considerations for a successful migration.

Postgres To Oracle: Load Data In Minutes

This article looked at two popular and powerful database management systems - PostgreSQL and Oracle Database. You were also shown their key features and what makes them stand out amongst database systems. We also explored two approaches for migrating data from Postgres to Oracle.
Looking to migrate from Sybase to Mongodb? Discover the seamless and efficient transition with our expert guide. Get step-by-step instructions, best practices, and key considerations for a successful migration.

Sybase To MongoDB: Load Data In Minutes

Looking to migrate from Sybase to Mongodb? Discover the seamless and efficient transition with our expert guide. Get step-by-step instructions, best practices, and key considerations for a successful migration.
 Looking to migrate from Informix to MySQL? Discover the seamless and efficient transition with our expert guide. Get step-by-step instructions, best practices, and key considerations for a successful migration.

Informix to MySQL: 2 Easy Ways

Looking to migrate from Informix to MySQL? Discover the seamless and efficient transition with our expert guide. Get step-by-step instructions, best practices, and key considerations for a successful migration.
As we discussed in the introduction, this write-up is going to explore two methods that can be used to move data from MySQL to BigQuery. Below is a detailed step-by-step explanation of these methods that can be quickly followed and implemented to create a data pipeline between the two platforms.

How To Load Data From MySQL To BigQuery

This step-by-step blog is going to explore two methods that can be used to move data from MySQL to BigQuery: using Arcion log-based CDC solution vs. manual transfer, both comes with code snippet that you can copy and paste. We'll also go over the pros and cons of both approaches.
SQL Server Replication is a powerful feature that allows you to distribute data to multiple servers and keep them synchronized. Learn about the benefits and best practices of SQL Server Replication with our comprehensive guide.

SQL Server Replication: Everything You Need To Know

SQL Server Replication is a powerful feature that allows you to distribute data to multiple servers and keep them synchronized. Learn about the benefits and best practices of SQL Server Replication with our comprehensive guide.
Unlock the potential of real-time streaming data with our comprehensive guide. Discover how this cutting-edge technology enables businesses to gather and analyze data in the moment, harnessing actionable insights for faster decision-making.

Real-Time Streaming Data: A Complete Guide

Unlock the potential of real-time streaming data with our comprehensive guide. Discover how this cutting-edge technology enables businesses to gather and analyze data in the moment, harnessing actionable insights for faster decision-making.
Migrating data from various data sources, especially cloud solutions, is a common theme in modern analytics. As such, this blog post is going to explain how to connect Snowflake to PostgreSQL so you can seamlessly transfer your data between both platforms.
18
 min read

Snowflake to Postgres: 2 Easy Methods

Migrating data from various data sources, especially cloud solutions, is a common theme in modern analytics. As such, this blog post is going to explain how to connect Snowflake to PostgreSQL so you can seamlessly transfer your data between both platforms.
Looking to migrate your data from MySQL to Databricks? Our expert guide will walk you through the process, helping you to seamlessly transfer your data and enjoy the benefits of cloud-based analytics. Read more.

MySQL to Databricks: Migration Guide and Best Practices

Looking to migrate your data from MySQL to Databricks? Our expert guide will walk you through the process, helping you to seamlessly transfer your data and enjoy the benefits of cloud-based analytics.
Looking to migrate your data from Oracle to Databricks? Discover the benefits of making the switch and learn how to transfer your data with expert tips and guidance seamlessly.

Oracle to Databricks: The Definitive Guide

Looking to migrate your data from Oracle to Databricks? Discover the benefits of making the switch and learn how to transfer your data with expert tips and guidance seamlessly.
The term real-time data integration is made up of two main phrases, real-time and data integration. You will look at each phrase in isolation before a holistic view is presented. Data integration refers to the process of combining data from multiple source systems into a single unified view for processing and gaining analytical insights.

Real-Time Data Integration: The Definitive Guide

In this blog, we'll dive into what is real-time data integration, why it's important, and three integration processes: capture data, transform data, and load data. We'll also go over the components needed for a real-time data integration architecture.
Real-time data analytics helps with the making of better decisions in a faster time frame. This is through the reliance on fresh data and fast queries.
19
 min read

Real-time Data Analytics: The Ultimate Guide

In this article, we’ll introduce the concept and types of real-time data analytics, a detailed breakdown of the advantages and benefits of real-time data analytics, and real-life examples of how it benefits different industries. Finally, we review the challenges and limitations to watch out for during the implementation.
Looking for the best data integration tools? We have done the groundwork and listed the 10 best data integration tools of 2023.
20
 min read

Top 10 Best Data Integration Tools of 2023

Looking for the best data integration tools? We have done the groundwork and listed the 10 best data integration tools of 2023.
An easy way of enabling Kafka CDC is to utilize the Kafka Streams library. Streams in Kafka are defined as an unbounded set of data that is continuously updating and can be seen as an immutable sequence of records where each data record is characterized as a key-value pair.

Kafka CDC: The Definitive Guide

In this post, we will look at implementing CDC with Apache Kafka, including the advantages and disadvantages. We will explore different methods of implementing CDC, including using Arcion to quickly build a scalable replication pipeline.
In this blog, you will learn to use SSMA to migrate data from Sybase to SQL Server through a series of steps. We’ll create a new SSMA project, connect to SAP ASE and the SQL Server instance, then map SAP ASE schema to SQL Server.
18
 min read

Migrating Sybase ASE to SQL Server: 2 Easy Methods

In this post, we will cover the basics of Change Data Capture in SQL Server, including native tools that can assist with data migration between the platforms. Using SSMA and SybaseToSQL, we will look at the step-by-step process of migrating from Sybase to SQL Server natively. Lastly, we will look at how to implement a CDC-enabled pipeline with Arcion to replicate data from Sybase ASE to Microsoft SQL Server.
In this blog, we are going to look at five Oracle replication tools that you should consider when you want to carry out your data replication.
17
 min read

5 Best Oracle Replication Tools

In this article, we will look at what Oracle is, what replication is, and the benefits of using a data replication tool. Lastly, we’ll look at 5 Oracle replication tools that could be used to implement data replication.
Change Data Capture can be performed in the Sybase ecosystem through the CDC Provider Sybase Replication Server component which is available to the Sybase Adaptive Server Enterprise (ASE).

Sybase CDC (Change Data Capture): The Definitive Guide

In this post, we will cover the basics of Change Data Capture, how it can be applied to Sybase ASE through native tools, and how to implement a CDC-enabled pipeline with Arcion. In the example, we go over a step-by-step example of how to replicate data from Sybase ASE to Snowflake.
Oracle GoldenGate is an extension of Oracle's data offering that enables you to replicate data from one database to another. Its modular architecture gives you the flexibility to extract and replicate selected data records, transactional changes, and changes to data definition language (DDL) across a variety of topologies.
17
 min read

Oracle GoldenGate Replication: A Step-by-Step Guide

Oracle GoldenGate is an extension of Oracle's data offering that enables moving data from one database to another. Its modular architecture gives you the flexibility to extract and replicate selected data records, transactional changes, and changes to data definition language (DDL) across a variety of topologies. In this post we will look at the benefits and limitations of the tool as well as how to implement it.
Oracle Change Data Capture (CDC) is a concept or mechanism that is used to identify and record data changes that occur in a datastore to other datastores in real time. It is used to synchronize data across data stores ensuring data consistency for applications and systems by moving data from one data repository to another. Using Kafka with Oracle is a common way to implement CDC within Oracle databases.

Oracle CDC to Kafka - Load data in Minutes

Oracle CDC, or Oracle Change Data Capture, is a technology that efficiently identifies data that has been added, updated, or removed from an Oracle database. Using Kafka with Oracle is a common way to implement CDC within Oracle databases. Explore how to implement a CDC pipeline using Oracle CDC and Kafka.
This article will look at the process involved in the migration of data from Oracle to Postgres. The stages to be discussed include assessment, schema migration, functional testing, performance testing, and data migration.

Oracle to Postgres Migration: Complete Guide

In this article, we will take a look at the process involved in the migration of data from Oracle to Postgres. We’ll discuss different stages including assessment, schema migration, functional testing, performance testing, and data migration.
Data can be replicated in PostgreSQL either synchronously or asynchronously. This can be done using a single-master or multi-master model, and a logical or physical replication model.

PostgreSQL Streaming Replication: 2 Easy ways

In this blog, we’ll walk you through step-by-step how to set up PostgreSQL streaming replication with no-code platform Arcion and use PostgreSQL built-in replication for streaming.
Oracle replication is an integrated feature of the Oracle server with its set of tools to replicate and integrate Oracle with other databases. Oracle replication uses distributed database technology to share data between various sites and with the right Oracle replication tool.
12
 min read

What is Oracle Replication? Methods and Guide

In this guide, we’ll walk you through what Oracle replication is all about and offer you four easy methods to replicate an Oracle database: full dump and load, incremental approach (table differencing), trigger-based approach, and change data capture.
There are multiple CDC methods available for use with the PostgreSQL database, these methods are similar but require a unique way to detect and specify changes made to your database and create a means of sending the changes to a target system.

PostgreSQL CDC: A Comprehensive Guide

There are multiple CDC methods available for use with the PostgreSQL database, these methods are similar but require a unique way to detect and specify changes made to your database and create a means of sending the changes to a target system.
Data Lakehouse is a new data solution concept that combines elements of a data warehouse with those of a data lake. It is a hybrid architecture with the fundamental purpose of combining the best characteristics of both the data warehouse and data lake.

Data Lakehouse: Everything You Need To Know

Data Lakehouse is a new data solution concept that combines elements of a data warehouse with those of a data lake. In this post, we’ll walk you through what is a Data Lakehouse, its benefits, the technology behind it, and how to use CDC to move data into a Data Lakehouse.
MongoDB Change Data Capture (CDC) is a concept or mechanism that is used to identify and record data changes that occur in a datastore to other datastores in real-time. It is used to synchronize data across the datastores ensuring data consistency for applications and systems by moving data from one data repository to another.

MongoDB CDC - Load Data in Real time

MongoDB Change Data Capture (CDC) is a concept or mechanism that is used to identify and record data changes that occur in a datastore to other datastores in real-time. It is used to synchronize data across the datastores ensuring data consistency for applications and systems by moving data from one data repository to another.
With CDC activated in MySQL, it becomes easier to keep track of data changes within MySQL as MySQL data is captured in real or near real-time and also allows you to recover from failures since data may be extracted just before the failure occurred.

What is MySQL CDC: 3 Easy Methods To Stream Data

The comprehensive guide covers how to set up MySQL CDC, the three methods of setting up MySQL CDC: use MySQL Triggers, binary logs, and Arcion. We also discussed pros & cons for each method. Follow along with the step-by-step guide.
Streaming SQL is a fundamental component of any modern data architecture as many companies now rely on their ever-increasing data production to make data-driven decisions not minding where the data resides.

What Is Streaming SQL? The Definitive Guide

Streaming SQL is a fundamental component of any modern data architecture as many companies now rely on their ever-increasing data production to make data-driven decisions not minding where the data resides.
A data pipeline is a series of data processing steps that involve the movement of data from one place, usually the source to a destination. Along the way, the data is transformed and optimized through filtering, cleaning, aggregating, and enriching, before arriving in a state that can be used for analysis to garner business insights.
16
 min read

Building Data Pipelines - Ultimate Guide

A data pipeline is a series of data processing steps that involve the movement of data from one place, usually the source to a destination. In this blog, we’ll walk you through what is a data pipeline, the elements of a data pipeline, how to design a data pipeline, and the important factors for sustainable data pipelines.
Data ingestion can be defined as the process of moving data from one or more sources into a target site for immediate use such as queries and analysis or storage. The data sources may include IoT devices, data lakes, databases, on-premise databases, SaaS applications, and so on from which data is ingested into data warehouses, data marts, or document stores.
12
 min read

What Is Data Ingestion? A Complete Guide

Data ingestion can be defined as the process of moving data from one or more sources into a target site for immediate use such as queries and analysis or storage. The data sources may include IoT devices, data lakes, databases, on-premise databases, SaaS applications, and so on from which data is ingested into data warehouses, data marts, or document stores.
A data pipeline is a combination of data processing steps used to move data from a source to a destination also known as a sink. It is a process in which each step’s output provides the input for the next step and along the way, the data is transformed and optimized to a required state that meets the specification of the destination.
12
 min read

What Is A Data Pipeline? Guide With Examples

A data pipeline is a combination of data processing steps used to move data from a source to a destination also known as a sink. In this guide, we’ll show you the difference between data pipelines and ETL, the elements of a data pipeline, different types of data pipelines, data pipeline architecture examples and more.
Using Kafka is a popular way to build a pipeline to stream data in your SQL Server database. We’ll go over a step-by-step approach to show how this can be done.

Kafka CDC SQL Server: A Detailed Guide

Apache Kafka is the most popular streaming tool. Using Kafka is a very common way to build a pipeline with SQL Server to stream data in your SQL Server database. In this overview, we will go over a step-by-step approach to show how this can be done. We will also look at how an equivalent pipeline can also be created with Arcion even more simply.
Database replication is the process of copying data from a source database to one or more target databases. Learn everything you need to know in this complete guide.

A Complete Guide on Database Replication

Database replication is the process of copying data from a source database to one or more target databases. Learn everything you need to know in this complete guide.
What is a data pipeline and what are the best data pipeline tools you need? Read all this and more covered inside our latest blog post on the matter.

What Are the Best Data Pipeline Tools? A Deep Dive Into Selecting the Most Efficient Solution

What is a data pipeline and what are the best data pipeline tools you need? Read this comprehensive guide to see a detailed comparison on the popular choices in the market.
In this guide, you will learn how to enable and implement CDC with Salesforce, steps to set up & subscribe to Salesforce Change Data Capture, and more.

Implementing CDC with Salesforce (Everything You need to know)

Data in Salesforce can be synchronized with external systems through Change Data Capture (CDC). This can be beneficial for analytics or to store updated data in a data warehouse.
Cloud data migration costs can vary widely from project to project. Read our latest post to learn how to prepare and efficiently manage your migration costs.

How to Manage Your Data Migration Costs While Moving to the Cloud

Cloud data migration costs can vary widely from project to project. Read our latest post to learn how to prepare and efficiently manage your migration costs.
In this comprehensive guide, we will show you how to load data from MySQL to MemSQL in real-time with Change Data Capture.

MySQL CDC to MemSQL (Now SingleStore): Load Data in Real-Time With CDC

In this comprehensive guide, we will show you how to load data from MySQL to MemSQL (now SingleStore) in real-time with Change Data Capture.
Microsoft SQL Server is one of the most popular RDBMS in the world. Adding Change Data Capture to MS SQL Server is a popular request for many organizations.

SQL Server CDC (Change Data Capture) Comprehensive Guide

Microsoft SQL Server is one of the most popular RDBMS in the world. Adding Change Data Capture to MS SQL Server is a popular request for many organizations. In this blog, we will introduce what Change Data Capture is as relates to SQL Server and its application as an audit solution. We’ll walk you through how to enable CDC to migrate data in SQL Server both at the database and table level. Finally, some advantages and disadvantages of SQL Server Change Data Capture to enable you to have a well-rounded understanding if faced with choosing SQL Server as a Change Data Capture solution.
Change Data Capture is a method used to detect changes in data and apply those changes to a different data store. When a change occurs in a source system, such as a database, action is taken by a process or another system to store or replicate those changes.

What Is Change Data Capture ? Everything You Need to Know

‍This article will focus on the data engineering concept of Change Data Capture, often shortened to CDC. In this article you will be introduced to the concept of Change Data Capture, how CDC fits into a data pipeline, CDC’s role within Extract, Load, Transform (ETL) techniques, how CDC works in real-world scenarios, and the benefits to businesses or organizations that adopt CDC. By the end of the article, you should have a firm understanding of Change Data Capture, its components, and the role it can play in managing rapidly changing data.

Sign up for the Arcion newsletter

Join the waitlist for Arcion Cloud (beta)

Fully managed, in the cloud.