I am using Google cloud Dataflow to setup a pipeline between MySQL to PostgreSQL using Datastream to SQL template.
First the data is migrated from MySQL DB to Cloud Bucket using Datastream in Avro format.
after this step the Dataflow template Datastream to SQL works and collects the data froom AVRO files and migrate this data to PostgreSQL Database.
Now One Case arises that I need to alter my MySQL table and need to add a column to it I have made same changes to my PostgreSQL table. But My new data is coming to the PostgreSQL database table but the data in the column which i added is not getting added to the target table. What should I do ?
Solved! Go to Solution.
The core issue likely stems from the schema mapping within your data migration pipeline becoming outdated due to the table alteration. This outdated schema mapping means the pipeline does not recognize or know how to handle the newly added column. Here are some steps you can take to try to resolve the issue:
1. Pipeline and Schema Management Understanding
2. Addressing Schema Mismatch
3. Data Backfill Strategy
Given the potential for existing data to lack values for the new column, a backfill strategy may be necessary:
Important Considerations
The core issue likely stems from the schema mapping within your data migration pipeline becoming outdated due to the table alteration. This outdated schema mapping means the pipeline does not recognize or know how to handle the newly added column. Here are some steps you can take to try to resolve the issue:
1. Pipeline and Schema Management Understanding
2. Addressing Schema Mismatch
3. Data Backfill Strategy
Given the potential for existing data to lack values for the new column, a backfill strategy may be necessary:
Important Considerations