awsbck

This utility lets you compress a folder and upload it to a AWS S3 bucket, once or periodically.


## Disclaimer This software is in a beta stage and, although it has not caused any problems in testing, I wouldn't recommend it for production use. Use at your own risks! The CLI will certainly change, but any breaking change should mean an increase in the minor version number as per semver , until it reaches `1.0.0`. New features that are backwards-compatible and bug fixes will lead to patch number bumps until then. ## Usage ``` Usage: awsbck [OPTIONS] --bucket --id --key Arguments: Path to the folder to backup [env: AWSBCK_FOLDER=] Options: -c, --cron Specify a cron espression to run the backup on a schedule [env: AWSBCK_CRON=] -f, --filename The name of the archive that will be uploaded to S3, without extension (optional) [env: AWSBCK_FILENAME=] -r, --region The AWS S3 region [env: AWS_REGION=] [default: us-east-1] -b, --bucket The AWS S3 bucket name [env: AWS_BUCKET=] --id The AWS S3 access key ID [env: AWS_ACCESS_KEY_ID=] -k, --key The AWS S3 secret access key [env: AWS_SECRET_ACCESS_KEY=] -h, --help Print help (see more with '--help') -V, --version Print version ``` CLI arguments take precedence over environment variables. The cron expression is parsed by the [`cron`](https://github.com/zslayton/cron) crate, with the following format (year is optional): ```rust // sec, min, hour, day of month, month, day of week, year let expression = "0 30 9,12,15 1,15 May-Aug Mon,Wed,Fri 2018/2"; ``` The `--filename` option accepts ASCII alphanumeric characters and `!-_.*'()/`. Other characters will be discarded. ### Example ```shell # The .env file in the current directory is read by awsbck $ cat .env AWS_REGION="eu-central-1" AWS_ACCESS_KEY_ID="YOUR_KEY_ID" AWS_SECRET_ACCESS_KEY="yoursecret" $ awsbck -c "@hourly" -b my_bucket /my_folder ``` ### Docker example ``` $ export AWS_REGION="eu-central-1" $ export AWS_ACCESS_KEY_ID="YOUR_KEY_ID" $ export AWS_SECRET_ACCESS_KEY="yoursecret" $ docker run \ --rm \ --mount type=bind,src="$(pwd)"/target,dst=/target,readonly \ -e AWS_REGION -e AWS_ACCESS_KEY_ID -e AWS_SECRET_ACCESS_KEY \ ghcr.io/beeb/awsbck:latest \ -c "15 */10 * * * *" -b my_bucket /target ``` ## Installation ### Prebuilt binaries Check out [the releases](https://github.com/beeb/awsbck/releases) for prebuilt binaries. ### Cargo ```shell $ cargo install awsbck ``` ### Nix Available through nixpkgs on the unstable channel. ```shell $ nix-env -iA nixpkgs.awsbck ``` ### Docker This utility is available as a [docker image `ghcr.io/beeb/awsbck`](https://github.com/beeb/awsbck/pkgs/container/awsbck). There are two tag variants, one running as a non-root user (`latest`) and one as a root user (`root-latest`). This image is particularly useful to backup named volumes in docker. If you encounter problems where the `awsbck` logs report a permissions problem, then you can try to switch to the `root-latest` tag. Below an example of using it with `docker compose`. In order to make sure the backup happens properly, we can't just copy the db data, as it might be in the middle of a write or other operation. Thus we send the `pg_dumpall` command and store the resulting dump to a separate volume that we can backup to S3. ```yml --- version: '3.2' volumes: # the first volume is to persist the database raw data database: # this volume will be used to share the dump file with awsbck database-backup: services: postgresql: image: postgres:14 restart: unless-stopped volumes: - type: volume source: database target: /var/lib/postgresql/data/ - type: volume source: database-backup target: /backup # this service will send a dump command to the postgres container periodically (here 6h) # and store the resulting file in the `database-backup` volume mounted at `/backup` postgres-backup: image: docker:cli container_name: postgres_backup volumes: - type: bind source: /var/run/docker.sock target: /var/run/docker.sock command: [ '/bin/sh', '-c', 'while true; do sleep 21600; docker exec -t postgres pg_dumpall -c -U postgres > /backup/dump_database.sql; done' ] # we mount the backup volume as read-only and back up the SQL dump daily at 3.12am awsbck: image: ghcr.io/beeb/awsbck:latest restart: unless-stopped volumes: - type: volume source: database-backup target: /database read_only: true environment: AWSBCK_FOLDER: /database AWSBCK_CRON: '0 12 3 * * * *' AWS_REGION: eu-central-1 AWS_BUCKET: my_bucket AWS_ACCESS_KEY_ID: $AWS_ACCESS_KEY_ID AWS_SECRET_ACCESS_KEY: $AWS_SECRET_ACCESS_KEY ```