mirror of
https://github.com/garraflavatra/docker-volume-s3-backup.git
synced 2025-06-27 21:35:10 +00:00
Compare commits
3 Commits
3653df72f4
...
044ddcdbda
Author | SHA1 | Date | |
---|---|---|---|
044ddcdbda | |||
cc24a7d123 | |||
b462c54de6 |
@ -22,6 +22,6 @@ ENV BACKUP_KEEP_DAYS ''
|
|||||||
ADD src/run.sh run.sh
|
ADD src/run.sh run.sh
|
||||||
ADD src/env.sh env.sh
|
ADD src/env.sh env.sh
|
||||||
ADD src/backup.sh backup.sh
|
ADD src/backup.sh backup.sh
|
||||||
# ADD src/restore.sh restore.sh -- not ready yet
|
ADD src/restore.sh restore.sh
|
||||||
|
|
||||||
CMD ["sh", "run.sh"]
|
CMD ["sh", "run.sh"]
|
||||||
|
14
README.md
14
README.md
@ -42,22 +42,19 @@ services:
|
|||||||
|
|
||||||
### Restore
|
### Restore
|
||||||
|
|
||||||
> [!NOTE]
|
|
||||||
> The restore command is yet to be ported from the original restore script from [`eeshugerman/postgres-backup-s3`](https://github.com/eeshugerman/postgres-backup-s3). Contributions are welcome :)
|
|
||||||
|
|
||||||
> [!CAUTION]
|
> [!CAUTION]
|
||||||
> DATA LOSS! All folder content will be dropped and re-created.
|
> All existing files will be replaced by the backup content!
|
||||||
|
|
||||||
#### ... from latest backup
|
#### From latest backup
|
||||||
|
|
||||||
```sh
|
```sh
|
||||||
docker exec <container_name> sh restore.sh
|
docker exec <container_name> sh restore.sh
|
||||||
```
|
```
|
||||||
|
|
||||||
> [!NOTE]
|
> [!NOTE]
|
||||||
> If your bucket has more than a 1000 files, the latest may not be restored -- only one S3 `ls` command is used
|
> If your bucket has over 1000 files, the latest may not be restored – the S3 `ls` command is issued only once.
|
||||||
|
|
||||||
#### ... from specific backup
|
#### From specific backup
|
||||||
|
|
||||||
```sh
|
```sh
|
||||||
docker exec <container_name> sh restore.sh <timestamp>
|
docker exec <container_name> sh restore.sh <timestamp>
|
||||||
@ -65,7 +62,8 @@ docker exec <container_name> sh restore.sh <timestamp>
|
|||||||
|
|
||||||
## Development
|
## Development
|
||||||
|
|
||||||
### Run a simple test environment with Docker Compose
|
You can use the Docker Compose file in the root with your own S3 bucket for development and testing.
|
||||||
|
|
||||||
```sh
|
```sh
|
||||||
cp .env.example .env
|
cp .env.example .env
|
||||||
# fill out your secrets/params in .env
|
# fill out your secrets/params in .env
|
||||||
|
@ -9,7 +9,7 @@ echo "Creating backup..."
|
|||||||
tar -czf dump.tar.gz /data
|
tar -czf dump.tar.gz /data
|
||||||
|
|
||||||
timestamp=$(date +"%Y-%m-%dT%H:%M:%S")
|
timestamp=$(date +"%Y-%m-%dT%H:%M:%S")
|
||||||
s3_uri_base="s3://${S3_BUCKET}/${S3_PREFIX}/${BACKUP_NAME}_${timestamp}.dump"
|
s3_uri_base="s3://${S3_BUCKET}/${S3_PREFIX}/${BACKUP_NAME}_${timestamp}.tar.gz"
|
||||||
|
|
||||||
if [ -n "$PASSPHRASE" ]; then
|
if [ -n "$PASSPHRASE" ]; then
|
||||||
echo "Encrypting backup..."
|
echo "Encrypting backup..."
|
||||||
|
@ -8,18 +8,18 @@ source ./env.sh
|
|||||||
s3_uri_base="s3://${S3_BUCKET}/${S3_PREFIX}"
|
s3_uri_base="s3://${S3_BUCKET}/${S3_PREFIX}"
|
||||||
|
|
||||||
if [ -z "$PASSPHRASE" ]; then
|
if [ -z "$PASSPHRASE" ]; then
|
||||||
file_type=".dump"
|
file_type=".tar.gz"
|
||||||
else
|
else
|
||||||
file_type=".dump.gpg"
|
file_type=".tar.gz.gpg"
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if [ $# -eq 1 ]; then
|
if [ $# -eq 1 ]; then
|
||||||
timestamp="$1"
|
timestamp="$1"
|
||||||
key_suffix="${POSTGRES_DATABASE}_${timestamp}${file_type}"
|
key_suffix="${BACKUP_NAME}_${timestamp}${file_type}"
|
||||||
else
|
else
|
||||||
echo "Finding latest backup..."
|
echo "Finding latest backup..."
|
||||||
key_suffix=$(
|
key_suffix=$(
|
||||||
aws $aws_args s3 ls "${s3_uri_base}/${POSTGRES_DATABASE}" \
|
aws $aws_args s3 ls "${s3_uri_base}/${BACKUP_NAME}" \
|
||||||
| sort \
|
| sort \
|
||||||
| tail -n 1 \
|
| tail -n 1 \
|
||||||
| awk '{ print $4 }'
|
| awk '{ print $4 }'
|
||||||
@ -27,18 +27,16 @@ else
|
|||||||
fi
|
fi
|
||||||
|
|
||||||
echo "Fetching backup from S3..."
|
echo "Fetching backup from S3..."
|
||||||
aws $aws_args s3 cp "${s3_uri_base}/${key_suffix}" "db${file_type}"
|
aws $aws_args s3 cp "${s3_uri_base}/${key_suffix}" "dump${file_type}"
|
||||||
|
|
||||||
if [ -n "$PASSPHRASE" ]; then
|
if [ -n "$PASSPHRASE" ]; then
|
||||||
echo "Decrypting backup..."
|
echo "Decrypting backup..."
|
||||||
gpg --decrypt --batch --passphrase "$PASSPHRASE" db.dump.gpg > db.dump
|
gpg --decrypt --batch --passphrase "$PASSPHRASE" dump.tar.gz.gpg > dump.tar.gz
|
||||||
rm db.dump.gpg
|
rm dump.tar.gz.gpg
|
||||||
fi
|
fi
|
||||||
|
|
||||||
conn_opts="-h $POSTGRES_HOST -p $POSTGRES_PORT -U $POSTGRES_USER -d $POSTGRES_DATABASE"
|
|
||||||
|
|
||||||
echo "Restoring from backup..."
|
echo "Restoring from backup..."
|
||||||
pg_restore $conn_opts --clean --if-exists db.dump
|
tar -xzf dump.tar.gz -C /data --strip-components 1
|
||||||
rm db.dump
|
rm dump.tar.gz
|
||||||
|
|
||||||
echo "Restore complete."
|
echo "Restore complete."
|
Reference in New Issue
Block a user