I also noticed the logs were growing and taking up quite some gb’s. I’ve used truncate
to make the syslog smaller and journalctl
vacuum
to keep that smaller
I also noticed the logs were growing and taking up quite some gb’s. I’ve used truncate
to make the syslog smaller and journalctl
vacuum
to keep that smaller
i’m trying to find documentation for this config but can’t find any. Do you have a link maybe ?
thanks for your detailed how-to ! Will try this, hopefully it holds 2 more weeks I’m leaving for holidays now :)
oh it’s visible because I’m admin/ mod ?
actually I did delete the server (after creating a snapshot of it) a week or so ago. But this morning I wanted to check lemmyfly.org, couldn’t load the page. Checking my Hetzner dashboard I noticed CPU was spiked at 200%?! It did drop again though, but apparently had last for 2-3 minutes. But prometheus was down, so no graphs apart from the hetzner ones. I doesn’t relate to network traffic spikes, so I don’t know what caused it. I’ve started the prometheus server again (that snapshot was really useful :) ) and will leave it on for a couple of months now.
current system consumption:
I might need to get an extra volume for storage, Lemmy is starting to eat up the root filesystem… Does anyone know how I re-configure Lemmy to look at a different volume for storage ?
For me it were the last 30 or something entries in local_user. They all didn’t have email_verified. I access the db through Postico. Also removed all other older accounts that didn’t verify their email address
Mine too I see now. You can only ban a user when they have posted something right ? There is no list for admins of all users registered ? Maybe I’ll do it direct in the database…
http_response metrics are also up, using https://github.com/martin-helmich/prometheus-nginxlog-exporter
would be nice to see the numbers go up :)
Let us know your findings when you did!
i’m now looking into logging more specific http request data. Maybe https://github.com/martin-helmich/prometheus-nginxlog-exporter?
I use namecheap’s email service privateemail since a couple of years. Google workspace was great, but don’t need the other features anymore and this is fifth of the price https://www.namecheap.com/hosting/email/
Looking into the database, it contains many thousands of posts. I’m assuming this is stored in the local db for serving it to instance members. So when you open a post from instance B on instance A, A fetches post-data from B, stores it in A database, then serve the content from db A to the browser
The purging of older content is good news I didn’t know that.
every instance is sharing in the traffic to browse the fediverse. Not one service is responsible for serving content, you (the instance admin) are only serving for your members.
The downside of this is there is a huge amount of replicated data stored everywhere. Content of popular communities will be scraped by and stored on many many servers, filling up servers and increasing storage and bandwith bills for all those servers
no, a random image I just now check was 3.3mb even. But you’re right, the image source does point to the origins server location… Then why are all those images stored on the server ?
Thanks for mentioning your provider, their prices are a lot cheaper then DO !
When I look on my server in the volumes/pictrs folder, there are a LOT of images that I recognize from my feed. They are not from posts posted to my server but belong to posts of communities that I have subscribed to
Thats the way the decentralized fediverse works (I think!) let me know if I’m wrong
only pictures and posts that the instance-member are subscribed to, and from the moment they interacted with it. But yes, I’m curious to see where this goes.
I noticed it when I was looking in the database and disk on my instance, where there are already thousands of posts and more then a thousand images - while my own instance only has a couple of posts and images
pictrs isn’t even taking up that much space, postgres takes up most !
# du -sch volumes/* 8.0K volumes/lemmy-ui 3.4G volumes/pictrs 8.2G volumes/postgres 12G total