I have large table in CloudSQL that needs to be updated every hour, and I'm considering Airflow as a potential solution. What is the best way to update a large amount of data in a CloudSQL database from Airflow?
The constrain are:
- The table need still be readable while the job is running 
- The table need to be writable in case one of the job runs overtime and 2 jobs end up running at the same time 
Some of the ideas I have:
- Load data needs to update into a pandas framework and run - pd.to_sql
- Load data into a csv in Cloud Storage and execute - LOAD DATA LOCAL INFILE
- Load data in memory, break it into chunks, and run a multi-thread process that each update the table chunk by chunk using a shared connection pool to prevent exhausting connection limits 
 
     
     
    