Over just 14 days our physical disk usage has increased from 52% to 59%. That’s approximately 1.75 GB of disk space being gobbled up for unknown reasons.
At that rate, we’d be out of physical server space in 2 -3 months. Of course, one solution would be to double our server disk size which would double our monthly operating cost.
The ‘pictrs’ folder named ‘001’ is 132MB and the one named ‘002’ is 2.2GB. At first glance this doesn’t look like it’s an image problem.
So, we are stumped and don’t know what to do.
Looks like docker log files. We merged a commit to lemmy-ansible for a fix a while ago: https://github.com/LemmyNet/lemmy-ansible/pull/49
Will this fix be in the next release?
If you’re using lemmy-ansible, follow the instructions there for upgrading.
it’s just a change to the docker-compose.yml file. so depending on how your instance is setup (using ansible or docker-compose), you could just make the change yourself.
I tried changing the docker-compose.yml file and it didn’t work. It just threw some vague error.
What are those errors specifically, and could you post your docker-compose.yml?
Here’s the part that I’ve added to the docker-compose.yml:
Here’s the error:
it looks like from the error message that you maybe don’t have the correct amount of whitespace. here’s a snippet of mine:
i also added the 4 logging lines for each service listed in my docker-compose.yml file. hope this helps!
Yes. This did help and thank you.
I have no idea where you are getting that from, but it doesn’t match lemmy-ansible, or the PR I linked.
Also please do not screenshot text, just copy-paste the entire file so I can see what’s wrong with it.
I looked more closely at the lemmy-ansible code and now our docker-compose.yml is functioning properly. Thanks for the heads up!