For smaller-scale migrations involving thousands or tens of thousands of assets, running the script from your local machine should suffice.
For larger migrations with hundreds of thousands of assets, it's advisable to run the script from a virtual machine (VM).
- If you're planning to run the script on a VM via an SSH connection, consider using a terminal multiplexer like
screen
ortmux
.- Failing to use a multiplexer may result in the termination of the migration process if your SSH connection closes and your VM session is terminated.
- The script is not resource-intensive as most of the heavy lifting is done by Cloudinary's back-end systems.
- For example, an AWS
t2.micro
VM should suffice.
- For example, an AWS
- A general rule of thumb for estimating migration log file size is approximately 300MB per 100,000 assets migrated. This can increase if you're transferring extensive taxonomy data.
- Each log file record contains the following data for troubleshooting:
- All fields from the corresponding input CSV file record.
- The Cloudinary API payload it was converted to.
- The API response.
- Each log file record contains the following data for troubleshooting:
- It's advisable to start by forking this repository so you can maintain any customizations and keep a history of changes.
- Clone the forked repository onto your VM.
- Node.js can be easily installed using the Node Version Management shell script.
- Open a terminal shell and navigate to the folder where you've cloned the repository. You should see a
package.json
file in that folder.- Run
npm install --production
to fetch and install required packages for the migration script (omitting dev dependencies).
- Run