Any questions, problems or suggestions with this guide? Ask a question in our community or contribute the change yourself at https://gitlab.com/baserow/baserow/-/tree/develop/docs .
Docker version 19.03 is the minimum required to use Baserow. It is strongly advised however that you install the latest version of Docker available. Please check that your docker is up-to date by running
docker -v
.
This guide assumes you already have Docker installed and have permissions to run Docker containers. See the Install on Ubuntu for an installation from scratch.
Run the command below to start a Baserow server running locally listening on port 80
.
You will only be able to connect to Baserow from the machine running the server via
http://localhost
.
docker run \
-d \
--name baserow \
-e BASEROW_PUBLIC_URL=http://localhost \
-v baserow_data:/baserow/data \
-p 80:80 \
-p 443:443 \
--restart unless-stopped \
baserow/baserow:1.30.1
BASEROW_PUBLIC_URL
to https://YOUR_DOMAIN
or http://YOUR_IP
to enable
external access. Ensure that this address matches the one you enter in your browser’s
URL bar. Any different address will be considered a published application.-e BASEROW_CADDY_ADDRESSES=:443
to enable
automatic Caddy HTTPS.-e DATABASE_URL=postgresql://user:pwd@host:port/db
to use an external
Postgresql.-e REDIS_URL=redis://user:pwd@host:port
to use an external Redis.There is a security flaw with docker and the ufw firewall. By default docker when exposing ports on 0.0.0.0 will bypass any ufw firewall rules and expose the above container publicly from your machine on its network. If this is not intended then run with the following ports instead:
-p 127.0.0.1:80:80 -p 127.0.0.1:443:443
which makes your Baserow only accessible from the machine it is running on. Please see https://github.com/chaifeng/ufw-docker for more information and how to setup ufw to work securely with docker.
The baserow/baserow:1.30.1
image by default runs all of Baserow’s various services in
a single container for maximum ease of use.
This image is designed for simple single server deployments or simple container deployment services such as Google Cloud Run.
If you are instead looking for images which are better suited for horizontal scaling (e.g. when using K8S) then please instead use our baserow/backend and baserow/web-frontend images instead which deploy each Baserow service in its own container independently.
A quick summary of its features are:
/baserow/data
folder inside the container.DATABASE_URL
or the DATABASE_...
variables to disable the internal postgres
and instead connect to an external Postgres. This is highly recommended for any
production deployments of this image, so you can easily connect to the Postgres
database from any other services or processes you might have.REDIS_URL
or the REDIS_...
variables to disable the internal redis and instead
connect to an external Postgres.BASEROW_CADDY_ADDRESSES
to https://YOUR_DOMAIN.com
and it will
automatically enable https
for
you and store the keys and certs in /baserow/data/caddy
.docker stop baserow
docker run
command you usually use to run your
Baserow and start up a brand-new container:# We haven't yet deleted the old Baserow container so you need to start this new one
# with a different name to prevent an error like:
# `response from daemon: Conflict. The container name "/baserow" is already in use by
# container`
docker run \
-d \
--name baserow_version_REPLACE_WITH_NEW_VERSION \
# YOUR STANDARD ARGS HERE
baserow/baserow:REPLACE_WITH_LATEST_VERSION
docker logs -f baserow_version_REPLACE_WITH_NEW_VERSION
[BASEROW-WATCHER][2022-05-10 08:44:46] Baserow is now available at ...
WARNING: If you have not been using a volume to persist the
/baserow/data
folder inside the container this will delete all of your Baserow data stored in this container permanently.
docker rm baserow
On November 2023 PostgreSQL released a final update for version 11 of the database together with an end-of-life notice for this version. This means, that PostgreSQL 11 will no longer receive security and bug fixes.
If you are using an embedded PostgreSQL database (an embedded one is when you do not provide POSTGRESQL_*
environment variables when launching Baserow, as opposed to an external one, where you provide connection details to your external PostgreSQL instance), and if you restart or try to run a new Baserow instance, if your data was initialized with PostgreSQL version 11, you’ll notice that it doesn’t start up anymore and raises an error because you need to upgrade your data directory to be compatible with PostgreSQL version 15. Baserow provides an image to automatically upgrade your data directory to PostgreSQL version 15, which is now the version officially supported by Baserow.
If you don’t want to upgrade at this point in time, jump to Legacy PostgreSQL version section below. Although, be aware, that we will only support PostgreSQL 11 for a limited amount of time and that this version won’t receive official updates from PostgreSQL anymore.
To upgrade your data directory to be compatible with PostgreSQL 15, follow these steps:
CAUTION: before doing this, make sure to Back up your Baserow instance to avoid potential data loss.
docker ps
. If Baserow is running, stop the container with docker stop baserow
.docker run \
--name baserow-pgautoupgrade \
# ALL THE ARGUMENTS YOU NORMALLY ADD TO YOUR BASEROW INSTANCE
--restart no \
baserow/baserow-pgautoupgrade:1.30.1
Baserow provides an image which runs a legacy PostgreSQL 11 version which can be run if you don’t want to upgrade to PostgreSQL 15 at this point. Note, that we will only support PostgreSQL 11 for a limited amount of time to help transitioning between database versions. Also, be aware, that PostgreSQL 11 is not receiving official security updates and bug fixes anymore.
To run Baserow image which uses legacy PostgreSQL 11 version, run:
docker run \
--name baserow-pg11 \
# ALL THE ARGUMENTS YOU NORMALLY ADD TO YOUR BASEROW INSTANCE
--restart unless-stopped \
baserow/baserow-pg11:1.30.1
See Configuring Baserow for more detailed information on all the other environment variables you can configure.
If you have a domain name and have correctly configured DNS then you can run the following command to make Baserow available at the domain with automatic https provided by Caddy.
Append
,http://localhost
to BASEROW_CADDY_ADDRESSES if you still want to be able to access your server from the machine it is running on using http://localhost. See Caddy’s Address Docs for all supported values for BASEROW_CADDY_ADDRESSES.
docker run \
-d \
--name baserow \
-e BASEROW_PUBLIC_URL=https://www.REPLACE_WITH_YOUR_DOMAIN.com \
-e BASEROW_CADDY_ADDRESSES=:443 \
-v baserow_data:/baserow/data \
-p 80:80 \
-p 443:443 \
--restart unless-stopped \
baserow/baserow:1.30.1
docker run \
-d \
--name baserow \
-e BASEROW_PUBLIC_URL=https://www.yourdomain.com \
-v baserow_data:/baserow/data \
-p 80:80 \
--restart unless-stopped \
baserow/baserow:1.30.1
docker run \
-d \
--name baserow \
-e BASEROW_PUBLIC_URL=https://www.yourdomain.com:3001 \
-v baserow_data:/baserow/data \
-p 3001:80 \
--restart unless-stopped \
baserow/baserow:1.30.1
docker run \
-d \
--name baserow \
-e BASEROW_PUBLIC_URL=https://www.yourdomain.com \
-e DATABASE_HOST=TODO \
-e DATABASE_NAME=TODO \
-e DATABASE_USER=TODO \
-e DATABASE_PASSWORD=TODO \
-e DATABASE_PORT=TODO \
-v baserow_data:/baserow/data \
-p 80:80 \
-p 443:443 \
--restart unless-stopped \
baserow/baserow:1.30.1
docker run \
-d \
--name baserow \
-e BASEROW_PUBLIC_URL=https://www.yourdomain.com \
-e REDIS_HOST=TODO \
-e REDIS_USER=TODO \
-e REDIS_PASSWORD=TODO \
-e REDIS_PORT=TODO \
-e REDIS_PROTOCOL=TODO \
-v baserow_data:/baserow/data \
-p 80:80 \
-p 443:443 \
--restart unless-stopped \
baserow/baserow:1.30.1
docker run \
-d \
--name baserow \
-e BASEROW_PUBLIC_URL=https://www.yourdomain.com \
-e EMAIL_SMTP=True \
-e EMAIL_SMTP_HOST=TODO \
-e EMAIL_SMTP_PORT=TODO \
-e EMAIL_SMTP_USER=TODO \
-e EMAIL_SMTP_PASSWORD=TODO \
-e EMAIL_SMTP_USE_TLS= \
-v baserow_data:/baserow/data \
-p 80:80 \
-p 443:443 \
--restart unless-stopped \
baserow/baserow:1.30.1
This is assuming you are using the postgresql server bundled by ubuntu. If not then you will have to find the correct locations for the config files for your OS.
sudo ls /etc/postgresql/
/etc/postgresql/YOUR_PSQL_VERSION/main/postgresql.conf
for editing as root# listen_addresses
line.listen_addresses = '*' # what IP address(es) to listen on;
/etc/postgresql/YOUR_PSQL_VERSION/main/pg_hba.conf
for editing as roothost all all 172.17.0.0/16 md5
sudo systemctl restart postgresql
sudo less /var/log/postgresql/postgresql-YOUR_PSQL_VERSION-main.log
docker run \
-d \
--name baserow \
--add-host host.docker.internal:host-gateway \
-e BASEROW_PUBLIC_URL=http://localhost \
-e DATABASE_HOST=host.docker.internal \
-e DATABASE_PORT=5432 \
-e DATABASE_NAME=YOUR_DATABASE_NAME \
-e DATABASE_USER=YOUR_DATABASE_USERNAME \
-e DATABASE_PASSWORD=REPLACE_WITH_YOUR_DATABASE_PASSWORD \
--restart unless-stopped \
-v baserow_data:/baserow/data \
-p 80:80 \
-p 443:443 \
baserow/baserow:1.30.1
The DATABASE_PASSWORD
, SECRET_KEY
and REDIS_PASSWORD
environment variables can
instead be loaded using a file by using the *_FILE
variants:
echo "your_redis_password" > .your_redis_password
echo "your_secret_key" > .your_secret_key
echo "your_pg_password" > .your_pg_password
docker run \
-d \
--name baserow \
-e BASEROW_PUBLIC_URL=http://localhost \
-e REDIS_PASSWORD_FILE=/baserow/.your_redis_password \
-e SECRET_KEY_FILE=/baserow/.your_secret_key \
-e DATABASE_PASSWORD_FILE=/baserow/.your_pg_password \
-e EMAIL_SMTP_PASSWORD_FILE=/baserow/.your_smtp_password \
--restart unless-stopped \
-v $PWD/.your_redis_password:/baserow/.your_redis_password \
-v $PWD/.your_secret_key:/baserow/.your_secret_key \
-v $PWD/.your_pg_password:/baserow/.your_pg_password \
-v baserow_data:/baserow/data \
-p 80:80 \
-p 443:443 \
baserow/baserow:1.30.1
If you want to directly access the embedded Postgresql database then you can run:
docker run -it \
--rm \
--name baserow \
-p 5432:5432 \
-v baserow_data:/baserow/data \
baserow/baserow:1.30.1 \
start-only-db
# Now get the password from
docker exec -it baserow cat /baserow/data/.pgpass
# Finally connect on your host machine to the Baserow postgres database at port 5432
# the password above with the username `baserow`.
The build in Caddy server is configured to automatically handle additional application
builder domains. Depending on the environment variables, it will also automatically
fetch SSL certificates for those domains. Note that the BASEROW_CADDY_ADDRESSES
environment variable must be :80
or :443
to allow multiple domains. If you have set
a URL there, it won’t work.
By default, it will accept requests of any domain over the http protocol, which is
perfect if you have a proxy in front of Baserow. If BASEROW_CADDY_ADDRESSES
starts
with https
protocol or is :443
, then it will redirect http requests to https, and
will handle the SSL certificate part automatically. This is recommended when the
container is directly exposed to the internet.
If you want to run a one off backend command against your Baserow data volume without
starting Baserow normally you can do so with the backend-cmd-with-db
argument like so:
docker run -it \
--rm \
--name baserow \
-v baserow_data:/baserow/data \
baserow/baserow:1.30.1 \
backend-cmd-with-db manage dbshell
This image can also be configured to deploy Baserow in a horizontally scalable way.
We recommend you first consider using our baserow/backend
and baserow/web-frontend
single service per container images
on K8S/Helm.
However, if you just want to easily horizontally scale Baserow on something like
AWS ECS or Google Cloud Run then the baserow/baserow
can be used.
To deploy this image in a horizontally scalable way you need to ensure all state is stored externally and not inside containers or volumes. To do this you will need a:
With this image we recommend you set the following environment variables to scale it horizontally.
See our Configuration Docs for more details on the following environment variables.
DATABASE_*
env vars need to be set to point Baserow at an external
postgresREDIS_*
env vars need to be set to point Baserow at an external redisDISABLE_VOLUME_CHECK=yes
needs to be set as this image has a
check for less technical users that /baserow/data
is mounted to an external
volume on startup. Instead, you want your containers to be stateless and so this
check needs to be disabled.AWS_*/GC_*/AZURE_*
env vars need to be set connecting Baserow to an external
file storage service.
This image has the following “horizontal scaling” environment variables:
BASEROW_AMOUNT_OF_GUNICORN_WORKERS
this controls the number of REST API
workers (the things that do most of the API work) per container.BASEROW_AMOUNT_OF_WORKERS
controls the number of background task celery
runners, these run realtime collaboration tasks, cleanup jobs and other slow tasks
like big file exports/imports.
BASEROW_RUN_MINIMAL=yes
AND BASEROW_AMOUNT_OF_WORKERS=1
.
Baserow stores all of its persistent data in the /baserow/data
directory by default.
We strongly recommend you mount a docker volume into this location to persist Baserows
data so you do not lose it if you accidentally delete your Baserow container.
The backup and restore operations discussed below are best done on a Baserow server which is not being used.
Note, that this only works if you’re not using an external PostgreSQL server. This will backup:
Otherwise if you remove the Baserow container you will lose all of your data.
The command below assumes you have been running Baserow with the
-v baserow_data:/baserow/data
volume. Please change this argument accordingly if you
have mounted the /baserow/data folder differently.
# Ensure Baserow is stopped first before taking a backup.
docker stop baserow
docker run --rm -v baserow_data:/baserow/data -v $PWD:/backup ubuntu tar cvf /backup/backup.tar /baserow/data
# Ensure Baserow is stopped first before taking a backup.
docker stop baserow
docker run --rm -v baserow_data:/baserow/data -v $PWD:/backup ubuntu tar cvf /backup/backup.tar /baserow/data
docker run --rm -v new_baserow_data_volume:/results -v $PWD:/backup ubuntu bash -c "mkdir -p /results/ && cd /results && tar xvf /backup/backup.tar --strip 2"
# Now launch Baserow using the new data volume with your normal run command:
docker run -v new_baserow_data_volume:/baserow/data .....
Please ensure you only back-up a Baserow database which is not actively being used by a running Baserow instance or any other process which is making changes to the database.
Baserow stores all of its own data in Postgres. To backup just this database you can run the command below.
# First read the help message for this command
docker run -it --rm -v baserow_data:/baserow/data baserow/baserow:1.30.1 \
backend-cmd-with-db backup --help
# Stop Baserow instance
docker stop baserow
# The command below backs up Baserow to the backups folder in the baserow_data volume:
docker run -it --rm -v baserow_data:/baserow/data baserow/baserow:1.30.1 \
backend-cmd-with-db backup -f /baserow/data/backups/backup.tar.gz
# Or backup to a file on your host instead run something like:
docker run -it --rm -v baserow_data:/baserow/data -v $PWD:/baserow/host \
baserow/baserow:1.30.1 backend-cmd-with-db backup -f /baserow/host/backup.tar.gz
When restoring Baserow you must ensure you are restoring into a brand new Baserow data volume.
# Stop Baserow instance
docker stop baserow
# Restore Baserow backup from a new volume containing the backup:
docker run -it --rm \
-v old_baserow_data_volume_containing_the_backup_tar_gz:/baserow/old_data \
-v new_baserow_data_volume_to_restore_into:/baserow/data \
baserow/baserow:1.30.1 backend-cmd-with-db restore -f /baserow/old_data/backup.tar.gz
# Or to restore from a file on your host instead run something like:
docker run -it --rm \
-v baserow_data:/baserow/data -v \
$(pwd):/baserow/host \
baserow/baserow:1.30.1 backend-cmd-with-db restore -f /baserow/host/backup.tar.gz
The Dockerfile already defines a HEALTHCHECK command which will be used by software that supports it. However if you wish to trigger a healthcheck yourself on a running Baserow container then you can run:
docker exec baserow ./baserow.sh backend-cmd backend-healthcheck
# Run the below to see all available healthchecks
docker exec baserow ./baserow.sh backend-cmd help
You can run management commands on an existing Baserow container called baserow by running the following to see the available commands:
docker exec baserow ./baserow.sh backend-cmd manage
# For example you could migrate the database of a running Baserow using:
docker exec baserow ./baserow.sh backend-cmd manage migrate
Baserow will automatically source any .sh
files found in /baserow/supervisor/env/
or
/baserow/data/env/
on startup. Use this to create a single config file to configure
your Baserow like so:
custom_baserow_conf.sh << EOF
export BASEROW_PUBLIC_URL=todo
export BASEROW_CADDY_ADDRESSES=todo
# You can perform any Bash logic required in here to setup Baserow conditionally.
EOF
docker run \
-d \
--name baserow \
-e BASEROW_PUBLIC_URL=http://localhost \
-v $PWD/custom_baserow_conf.sh /baserow/supervisor/custom_baserow_conf.sh \
-v baserow_data:/baserow/data \
-p 80:80 \
-p 443:443 \
--restart unless-stopped \
baserow/baserow:1.30.1
Or you can just store it directly in the volume at baserow_data/env
meaning it will be
loaded whenever you mount in this data volume.
FROM baserow/baserow:1.30.1
# Any .sh files found in /baserow/supervisor/env/ will be sourced and loaded at startup
# useful for storing your own environment variable overrides.
COPY custom_env.sh /baserow/supervisor/env/custom_env.sh
# Set the DATA_DIR environment variable to change where Baserow stores its persistent
# data. At startup Baserow will attempt to chown and setup this folder correctly.
ENV DATA_DIR=/baserow/data
# This image bakes in its own default user with UID/GID of 9999:9999 by default. To
# Set this to change the user Baserow will run its Caddy, backend, Celery and
# web-frontend services as. However be warned, the default entrypoint needs to be run
# as root so using USER may break things.
ENV DOCKER_USER=baserow_docker_user