-
Notifications
You must be signed in to change notification settings - Fork 179
Database backup-fetch failing with GCS storage #733
Comments
Just to confirm which version of Workflow are you running? I know there was some GCS-specific fixes in wal-e v1.0.1 which landed in v2.10.0 IIRC. |
This issue looks related to wal-e/wal-e#270. Can you try wal-e/wal-e#270 (comment) and see if that resolves the issue for you? |
Re. the version of Workflow I'm on, I'm not sure how to tell -- I installed it via the suggested path using helm which I assume installs the latest? When I tried setting
I'm looking around for a solution to that now. |
wal-e appears to be v1.0.1 though:
|
The issue I linked to seems to remain as an open issue so my assumption of it being fixed in v1.0.1 is incorrect. I'm going to close this as a duplicate of deis/postgres#154 so we can track efforts in one ticket. Thank you for the report though! |
Thanks. |
Hi,
I have set up a deis cluster with a default values.yml and everything works fine. When I configure
global/storage
to begcs
(and fill in thegcs
section withkey_json
and the three bucket names) I getIt appears that the service account has access to the bucket as I see data being written to the bucket (and I see the successful write in the log above) but the
backup-fetch
process appears to be failing.I also tried setting WALE_GS_PREFIX and GOOGLE_APPLICATION_CREDENTIALS and running
strace -f /usr/bin/python3 /usr/local/bin/wal-e backup-fetch /var/lib/postgresql/data LATEST
inkubectl exec
and the relevant bit seems to be:It looks as if perhaps file descriptor 17 was closed and then and attempt is made to F_SETPIPE_SZ which causes the problem. It looks like this is just the pipe to
lzop
and in fact I see ~5000lzop
processes running in the pod where I've been experimenting with this:they all look the same, FWIW:
Any pointers would be greatly appreciated.
The text was updated successfully, but these errors were encountered: