You need to be running Flagsmith version
2.28.2 or greater to be able to export and import your Flagsmith data.
Flagsmith offers some helpful utilities if you want to migrate your application from your self hosted instance to our cloud offering.
The process looks something like:
- Step 1. Contact Flagsmith support to confirm you would like to migrate from self hosted to cloud
- Step 2. Generate a JSON file from your self hosted instance (more information below)
- Step 3. Send the JSON file to Flagsmith support
- Step 4. Flagsmith support will import the JSON file into our cloud offering
- Step 5. Register and re-add your users and passwords (Flagsmith support will need to assign at least one organisation administrator to the newly imported organisation)
What is exported?
We will export the following entities:
We will not export the following entities:
- Flagsmith Users that log into the Dashboard and manage Flagsmith
- Audit logs
- Change requests
- Scheduled flag changes
The export process
The export process involves running a command from a terminal window. This must either be run from a running container in your self hosted deployment or, alternatively, you can run a separate container that can connect to the same database as your deployed fleet of flagsmith instances. To run the command, you'll also need to find the id of your organisation. You can do this through the django admin interface. Information about accessing the admin interface can be found here. Once you've obtained access to the admin interface, if you browse to the 'Organisations' menu item on the left, you should see something along the lines of the following:
The id you need is the one in brackets after the organisation name, so here it would be 1.
Once you've obtained the id of your organisation, you're ready to export the organisation as a JSON file. There are 2 options as to where to output the organisation export JSON file. Option 1 - local file system, Option 2 - S3 bucket. These options are detailed below.
Option 1 - Local File System
python manage.py dumporganisationtolocalfs <organisation-id> <local-file-system-path>
python manage.py dumporganisationtolocalfs 1 /tmp/organisation-1.json
Since this will write to your local file system, you may need to attach a volume to your docker container to be able to obtain the file afterwards. There is an example docker-compose file provided below to help guide you to do this.
Option 2 - S3 bucket
The command you will need to run for this is slightly different as per the following.
python manage.py dumporganisationtos3 <organisation-id> <bucket-name> <key>
python manage.py dumporganisationtos3 1 my-export-bucket exports/organisation-1.json
This requires the application to be running with access to an AWS account. If you're running the application in AWS,
make sure whichever role you are using to run you container has access to read from and write to the given S3 bucket.
Alternatively, you can provide the
AWS_SECRET_ACCESS_KEY environment variables to refer to an
IAM user that has access to the S3 bucket.