this post was submitted on 12 Jun 2023
28 points (96.7% liked)

Selfhosted

40042 readers
920 users here now

A place to share alternatives to popular online services that can be self-hosted without giving up privacy or locking you into a service you don't control.

Rules:

  1. Be civil: we're here to support and learn from one another. Insults won't be tolerated. Flame wars are frowned upon.

  2. No spam posting.

  3. Posts have to be centered around self-hosting. There are other communities for discussing hardware or home computing. If it's not obvious why your post topic revolves around selfhosting, please include details to make it clear.

  4. Don't duplicate the full text of your blog or github here. Just post the link for folks to click.

  5. Submission headline should match the article title (don’t cherry-pick information from the title to fit your agenda).

  6. No trolling.

Resources:

Any issues on the community? Report it using the report flag.

Questions? DM the mods!

founded 1 year ago
MODERATORS
 

For those self-hosting a lemmy instance, what hardware are you using? I am currently using a small Hetzner VPS. It has 2 vCPU, 2GB RAM and 40GB SSD storage. My instance is currently just in testing with me as the only user, but I plan to use it for close friends or family that may want to try this out, but might not want to sign up for a different instance. My CPU and RAM usage is great so far. My only concern is how large the storage will balloon to over time. I’ve been up for ~20 hours and it’s grown to 1.5G total volume since.

top 38 comments
sorted by: hot top controversial new old
[–] [email protected] 4 points 1 year ago (1 children)

shower thoughts... and still on my first cup of coffee to more just musing than anything ...

if storage is the concern wonder if the lemmy roadmap might one day include an option to use cloud based storage?

azure storage at .06/gb per month is likely cheaper and more redundant than local storage - even if you factor in calls to the blob which could be lowered via caching.

cloud storage potentially might one day lead to a option for smaller self hosters to opt into a shared blob instance where the and cost is shared.

in this scenario security to ensure the cloud blob couldn't be deleted would need to be thought through (maybe splitting the password among multiple admins with each having one part of the whole?) but might be one way to better encourage more self hosting for them compute side of things.

[–] [email protected] 5 points 1 year ago

If I self hosted my own Lemmy on my home server, just for myself and I posted / uploaded images on it, when another user from another instance views my image, they cache it, would this mean later down the line if I deleted to free up server space, if someone else on that instance was to come across my image after deletion, because it was previously cached, the image would still show?

Wondering if rolling storage is possible eventually, where an archive of posts older than 2 years is performed and data deleted.

[–] [email protected] 4 points 1 year ago (2 children)

Using Oracles "Always free" instances.

4 vCPU (ARMv8, not sure about the speed) 24 GB RAM 200 GB Flash storage

[–] [email protected] 1 points 1 year ago (1 children)
[–] [email protected] 2 points 1 year ago

Until everyone starts doing it, Oracle can be...fun to deal with

[–] [email protected] 1 points 1 year ago

I am concerned about scalability, I want to offer an option for many users but I would like to have a parameter hardware vs user counting

[–] [email protected] 3 points 1 year ago (2 children)

If I can eventually get it to work, I'll be hosting my personal Lemmy instance on a Hetzner CX21 VPS (2 CPU, 4GB RAM). Unfortunately after about a dozen attempts, with many variables tweaked to try to narrow down the issue, I still get nothing but a spinning button when I try to login after installation. I find it so odd that some people have no problem getting it up-and-running, and it simply won't work for others.

[–] [email protected] 1 points 1 year ago

I had this issue and the cause was the email server. If you're using smtp you'll have to make sure your vps opens port 25.

However you can also just register without an email and it should work

[–] [email protected] 1 points 1 year ago

I experienced this too. If you're logging in as an admin with email verification turned on, there is a bug that prevents you from logging in (since technically your email isn't verified yet). I was able to get logged in by following the steps here (https://lemmy.click/comment/973) which included modifying the database table to mark your account as verified.

[–] [email protected] 2 points 1 year ago* (last edited 1 year ago)

Currently a 1CPU/2GB RAM Linode instance for 26 users. Linode's pricing gets insane as you scale up though so I will definitely be looking elsewhere if I need to scale much bigger. I think I could get away with 1gb of RAM at half the cost right now, but I'm also hosting a Matrix homeserver on this VPS and Synapse is a hungy boy.

[–] [email protected] 2 points 1 year ago (1 children)

From what I've heard (take this with a huge grain of salt) is that the posts themselves shouldn't take up much of your storage. The biggest thing that could take up your storage are images, but they are only stored on the instances where the community in which they were posted in is.

[–] [email protected] 1 points 1 year ago (1 children)

Do you have any suggestions for what the 1.5GB is then?

[–] [email protected] 3 points 1 year ago* (last edited 1 year ago) (2 children)

It's likely the Docker images, and maybe the Docker build cache if they built from source instead of using the Docker Hub image.

I've been up for about a day longer than OP, and my Lemmy data is still under 800MB. OP either included non Lemmy data in that math, or is subscribed to way more communities than me. My storage usage has been growing much faster today with all the extra activity, but I won't have to worry about storage space for about a month even at this rate.

And that's assuming Lemmy doesn't automatically prune old data. I'm not sure if it does or not. But if it doesn't, I imagine I'll see posts in about 2-3 weeks talking about Lemmy's storage needs and how to manage it as an instance admin.


EDIT: Turns out ~90% of my Lemmy data is just for debugging and not needed:

https://github.com/LemmyNet/lemmy/issues/3103#issuecomment-1631643416

[–] [email protected] 2 points 1 year ago (1 children)

It would be cool to hear back from you guys hosting in about a week or so, so see if it'll just grow linearly, or if slows down at some point.

[–] [email protected] 0 points 1 year ago

Sure, I'm curious too. I'll keep an eye on the usage!

[–] [email protected] 1 points 1 year ago (1 children)

I used the ansible route to get going. I am subbed to ~150 communities currently. Some of those won't stay, but for now I am subbing to almost anything to see how that affects disk usage. I am interested to see how, or if, it levels off over time and what a week or two out looks like. I expect by then we will all have many more tips for each other as we trial and error our way through.

Here's my current usage:

[–] [email protected] 0 points 1 year ago* (last edited 1 year ago) (1 children)

Ahhhh, image posts are where your usage is going! Makes sense, my instance is just for my account and I don't submit anything. Your postgres size is more or less in line with where mine was at your uptime. I'm using Docker Compose so I'm only considering the size of the volumes in my metrics, not the image sizes or anything.

EDIT: Turns out I accidentally firewalled my pictrs service, it couldn't reach external sites to create thumbnails for external links. That's why my storage usage was so much lower. Whoops.

[–] [email protected] 2 points 1 year ago (1 children)

Yeah, images are where the main bulk of the storage is going. Interestingly, my instance is also just for my account presently and I have not submitted any images until my screenshot above. So these images are just those that are being pulled from other instances. I was under the impression that images were hosted from their respective instance and not saved locally, so I am curious to see how this plays out long term.

[–] [email protected] 1 points 1 year ago (1 children)

Thumbnails are stored locally, I believe.

[–] [email protected] 2 points 1 year ago

Confirmed. Investigated that earlier.

[–] [email protected] 2 points 1 year ago (1 children)

1 vCPU, 1GB Ram, 50GB storage using the smallest x86_64 compute instances on Oracle Cloud. Qualifies for always free which is nice while I'm simply testing out a personal server. It's working just fine within those constraints. For now, at least.

Like you, I'm worried about storage. I would like to run it from home, but I live in the woods and my internet isn't reliable enough.

[–] [email protected] 3 points 1 year ago

If you run from home I would say to take precautions, like using a cloudflare tunnel.

[–] [email protected] 2 points 1 year ago (1 children)

4vcpu (Ryzen), 8GB RAM, 256gb disk (which will be expanded when it gets to like 60% full). Not too worried about storage unless I get a bunch of image-happy users, text all comes in as json and goes straight to Postgres so it’s not a concern.

[–] [email protected] 1 points 1 year ago (1 children)

I am going to host an instance as well and I am worried about disk space...

This server you are using is from an VPS provider? how much they cost?

[–] [email protected] 1 points 1 year ago* (last edited 1 year ago)

I've got a baremetal server with OVH running VMware, so it's just a VM that I manage. I'm paying more for it than I'd like, but it's running far more than just Lemmy. If I wind up ditching it in the future, it's just a quick vMotion off to another machine + DNS updates.

Here's a current output of my storage about a week into hosting the instance. It's growing slower than I expected, and I do have plans to move volumes/pictrs up to an s3 bucket whenever I start running low on local storage.

[jon@lemmy lemmy]# du -sh volumes/*
2.5G    volumes/pictrs
2.2G    volumes/postgres

I would recommend locking down SSH on your Lemmy server, I have mine restricted to allow logins from VPN only. Otherwise you'll get probed 24/7 with a public server.

[–] [email protected] 2 points 1 year ago* (last edited 1 year ago)

I host it on a low power PC in my closet. I also limit the resource usage to 2 cores and 2GB of RAM in Kubernetes (RKE2) config. My ISP won't give me any static IP, so I routed the traffic via a $3 VPS over tailscale network.

[–] [email protected] 2 points 1 year ago

3vcpu,4gb ram, 80ssd. But I'm just running a personal servers, so it's plenty for my needs.

[–] [email protected] 1 points 1 year ago* (last edited 1 year ago)

I'm using a Ramnode VPS since I had some unused credit I wanted to use up. 2 vcpu, 1 GB ram, and 35 GB ssd.

Seems to be working well enough so far, but right now it's just me. If I open up to more users, I might need to upgrade, but we'll cross that bridge when we get there.

Edit: I may have spoke too soon; had to reboot the server due to low memory. Hopefully a swap file will alleviate that a bit, but I might have to upgrade the RAM on this server. We'll see.

[–] [email protected] 1 points 1 year ago

It's definitely overkill, but right now I'm hosting it on my nomad cluster. It only has 4cpu/8gb allocated at the moment but will autoscale (vertically) if needed. I already have a separate postgres cluster used for other things so I'm just borrowing that for now too. I haven't tried running multiple instances yet but I'll probably test that out this week to see how/if it works.

[–] [email protected] 1 points 1 year ago

I selfhost on my own homeserver. At the moment, I've spared it 2 cores, 1GB of RAM and 32GB of NVMe storage mirrored.

[–] [email protected] 1 points 1 year ago* (last edited 1 year ago) (1 children)

I have it running on my microk8s single node cluster. It's a dual xeon (40 cores total) with unfortunately only 64gb ram. The motherboard's max. I got a das with 72tb storage, currently in btrfs mirrored. Hoping btrfs raid-like configs become more usable in the future. I was using zfs but I always ran into issues.

[–] [email protected] 1 points 1 year ago

Interesting - I always ran into issues with btrfs so now I am using ZFS exclusively :D

[–] [email protected] 1 points 1 year ago

I'm running it on an LXC container that lives on a proxmox cluster.

2 vCPU at 2.6Ghz. 2GB of RAM (it's LXC so I can allocate more if needed...) and 40GB of SSD-backed CEPH storage. I actually just upped this to 150GB because I can see the velocity of data I'm storing for this. I have about 2 more TB of storage on the CEPH cluster before I need to order a few more SSDs.

I have terrible internet, but I do have a static address. And they're installing fiber in my neighborhood right now. So that will change soon too.

Based on what I've seen thus-far, I suspect I can handle about a hundred users on this without much issue.

[–] [email protected] 0 points 1 year ago (1 children)
  • 1 vCPU 2.9ghz
  • 1 GB DDR4 Memory
  • 25 GB NVMe/SSD Storage

5~ USD a month. Working great for personal use and I'd imagine a handful of users. Hosted in a data center that is very close to me.

Also fwiw: 4 days of lemmy. I am subbed to a bunch of stuff. I've only uploaded like three pictures to my instance... All that space is thumbnails from other instances.

692M    ./postgres
8.0K    ./lemmy-ui
499M    ./pictrs
1.2G    .
1.2G    total
[–] [email protected] 1 points 1 year ago (1 children)

There's my current disk usage. I've gone wild subscribing to just about every community I come across to see how the storage adds up. Right now I've got ~150 communities subbed. We'll see how it goes and when I'll need to expand the storage.

[–] [email protected] 0 points 1 year ago (1 children)

Not too bad... How long has your instance been up? Next thing I want to investigate is the postgrea database itself. On my to-do list.

[–] [email protected] 1 points 1 year ago

Been running ~22 hours at this point

[–] [email protected] -1 points 1 year ago

I have thought about running my own, following this for the info.

load more comments
view more: next ›