Self-Hosted Alternatives to Popular Services

6 readers
1 users here now

A place to share, discuss, discover, assist with, gain assistance for, and critique self-hosted alternatives to our favorite web apps, web...

founded 1 year ago
MODERATORS
51
 
 
The original post: /r/selfhosted by /u/The-Little-Tinkerer on 2024-11-11 01:05:19.

Hello,

I wonder if it exists any free open source selfhosted alternative to freemius ?

Could be awesome to integrate with WordPress plugin. The only one I found is LicenseBox but isn’t free.

Thanks a lot

52
 
 
The original post: /r/selfhosted by /u/vaquishaProdigy on 2024-11-11 01:01:06.
53
 
 
The original post: /r/selfhosted by /u/dakazze on 2024-11-11 00:52:59.

Looks like it is finally my time to ask a question after participating here for quite some time ^^

I currently have an issue with jellyfin and syncplay:

Using the local jellyfin IP on my webOS TV works fine, except for syncplay. Syncplay only works when all clients in the session connect via my cloudflare domain jelly.domain.x which NginX (websockets configured according to jellyfin docs) redirects to the local jellyfin IP. I tried setting up a hostname in my openWRT router (jelly.local pointing to NginX) and added that to my jelly.domain.x redirect in NginX. This lets me access jellyfin but syncplay still does not work for whatever reason.

Disabling the cf proxy, disabling "force SSL" in NginX and letting the other client connect via http:// too does not help either.

I would just stick to the external domain for my local WebOS TV but my upload bandwidth is a limiting factor and so I would prefer to keep the traffic local if possible.

I read about NAT-loopback but I am a complete noob when it comes to networking stuff so I had to give up after researching on google for about an hour. On top of that I dont even know if that is the right solution...

My setup currently looks like this:

jelly.domain.x --> Cloudflare proxy + SSL (tried without both) --> openWRT router forward --> NginX redirect --> jellyfin server

Another crazy thing is, that syncplay works when my laptop is using the jelly.local address and the TV is using jelly.domain.x but not the other way around. To me this suggests that there is something weird going on with the webOS app but maybe we can find a workaround that lets me use the external domain while still keeping the streaming traffic local?

I would appreciate any help!

EDIT:

Okay, I found out that my Jellyfin traffic stays local when I disable the cloudflare proxy, even when using the external domain on my TV AND syncplay works. I would have preferred to keep the cf proxy but if thats the worst case I can live with it. Maybe someone still knows a solution for that.

54
 
 
The original post: /r/selfhosted by /u/-rch- on 2024-11-11 00:21:58.

Update to the latest version (1.32.4) as soon as possible:

This release has fixed some CVE Reports reported by a third party security auditor and we recommend everybody to update to the latest version as soon as possible. The contents of these reports will be disclosed publicly in the future.

55
 
 
The original post: /r/selfhosted by /u/themasterplan69 on 2024-11-10 23:45:39.

Now that ZFS and TrueNAS Scale can extend pools I'm struggling to think of a compelling reason to recommend Unraid over TrueNAS Scale.

They both can

  • all kinds of shares, SMB, NFS, iSCSI
  • add drives to expand pools
  • support for VMs
  • support for Docker
  • support for Apps

TrueNAS Scale is free (FOSS) while Unraid requires a subscription. TrueNAS also doesn't have the USB drive install requirement and it can be easily virtualized, e.g. on Proxmox.

So it has to be asked, why would one not use or recommend TrueNAS Scale in late 2024? If you are setting up a new NAS in late 2024, and not using TrueNAS, what are you choosing and for what reasons?

56
 
 
The original post: /r/selfhosted by /u/L2jelly on 2024-11-10 23:45:14.

Hi, I have my own email infrastructure SAAS & I need Contabo hosting. I know the VPS servers constantly go down but how about VDS servers or bare metal? Would really really appreciate some advice as I'm on a time crunch.

57
 
 
The original post: /r/selfhosted by /u/leokrDE on 2024-11-10 22:40:59.

Hi guys,

my situation is as following: I am working as a freelance photographer.

Most of my customers have different needs regarding the images I deliver to them, speaking of image size and metadata. From time to time, it happens that I need to deliver the same image to different customers or two the same customer in different kinds (e.g. a newspaper needs small resolution and metadata is very important, but sometimes they need a bigger version of one of the images with different metadata.

Ideally I don't want to re-export that image out of Lightroom, because that would lead to two exported copies of the images. Additionally, this requires me having access to my laptop AND access to the RAW file. While I keep the most recent images on my SSD for that reason, it happens quite often that the particular image needed is already transferred to my HDD Archive and its Backups.

I imagine a web based solution, that receives high quality exports of all images, not customized for any customer. It reads metadata from that image to build a basic set of Metadata. From there, I can generate the copies needed for the particular customer, which are automatically deleted after a week or two. If I need to have them again, a single klick would re-generate them from the high res JPEG. If needed, I can generate multiple versions of the same Image, having different resolutions and different metadata, just coexisting as database entries on how to generate them, whilst the generated images just have a short lifespan on the server.

Currently I use a Nextcloud to share images to my customers, living with the limitations mentioned above.

I am shure knowing that there won't be an out-of-box solution for my problem. However I would like to know, if anyone of you has a good suggestion, which open source projects could be a good base to start upon, so I wouldn't need to start from scratch.

I don't need fancy features like image recognition, GPS maps, auto-generated memories videos or stuff like that, since that solution is not meant for my personal images, but for professional use.

Cheers, Leo

58
 
 
The original post: /r/selfhosted by /u/phillip-un on 2024-11-10 22:25:38.

I've been using Proxmox in my home lab for a good four years. I'm thinking about switching to VMware. So the question is, which hypervisor do you use?

59
 
 
The original post: /r/selfhosted by /u/Cheap-Picks on 2024-11-10 22:17:53.
60
 
 
The original post: /r/selfhosted by /u/3X0karibu on 2024-11-10 21:19:55.
61
 
 
The original post: /r/selfhosted by /u/Routine_Librarian330 on 2024-11-10 19:06:15.

I've recently tried out a music server to stream my collection on the LAN (and potentially anywhere through a VPN tunnel into said LAN). Still, I really don't see the point: what genuine upsides? The alternatives just seem better:

  • carrying my most frequently played titles with the on my phone. Sure, that'll cost me some storage space, but modern devices have plenty of that, and as an added bonus, I'm not producing any traffic in doing so - no added power usage on my server, no traffic counting towards my data rate at my mobile provider. And I'm leaving zero data trail to be gobbled up and (ab)used by any large platforms - it's the perfect solution privacy-wise.
  • streaming everything via commercial providers (e.g. Spotify or YouTube). Yes, that will use some of my bandwidth, but so does streaming stuff off the homeserver. On the upside, these platforms have got way more titles than both of my other options - dumping music onto my phone or streaming self-hosted music. Obviously, there are some privacy costs (both platforms track and profile you).

Have I overlooked anything? Are these all the arguments there are? Or did I miss a killer feature in self-hosted music?

62
 
 
The original post: /r/selfhosted by /u/Zamorio2 on 2024-11-10 18:35:36.

I've been trying to get AIO to run for days. I've followed these instructions, done everything and gotten to install everything but when I go to my domain this error appears

https://preview.redd.it/swmyjm8sb40e1.png?width=840&format=png&auto=webp&s=4a9ab65e3f08721a244b8657b7c6a97f2c48cc18

Apache logs say this:

https://preview.redd.it/h6ssqyhvb40e1.png?width=1760&format=png&auto=webp&s=b06b63fd822e869a31ca880dc0b6678228313049

I've been trying to fix it with these instructions but I'm afraid my technical knowledge is not that good.

On Friday I got an installation running and I was able to open Nextcloud (I had to redo that installation because I mounted datadir to the incorrect samba share and I haven't been able to make it work again) so I know it's not a ports problem or IP problem or anything like that.

I also know it's not a multiple certificates problem because I've created a new domain in duckDNS for this install. It might be something in the "Authorization failed: HTTP 400 urn:ietf:params:acme:error:connection" section but I don't know what or how to fix it as it's a DuckDNS domain and I don't understand everything it says there.

This is a completely clean install in a completely new VM (same MAC adress as the previous ones but that shouldn't affect). What can I do?

63
 
 
The original post: /r/selfhosted by /u/GoldenFalcon on 2024-11-10 17:54:41.

I have been trying for hours about how to get Kavita to work outside my network so I can access comics while out of town. But after installing Kavita, getting CDisplayEX on my android device on the server's network.. I can't move beyond that. I tried NOIP.com for reverse proxy, installing caddy, tailscape, ubooquity, doing tons of stuff in command line and powershell. I'm really frustrated with this process. I don't know what I am doing wrong. This all seems very clunky for something everyone keeps saying is "easy".

64
 
 
The original post: /r/selfhosted by /u/Bill_Guarnere on 2024-11-10 16:38:00.

It's been a while since people seems to prefer miniPC to ARM SBC as home servers, and honestly I don't really understand this trend, ARM SBCs still are relevant and in most case are the best solutions imho.

I live in a country where electricity is not cheap at all, and I think my use case can be extended to many other people in the same continent (EU), and because we're talking about a system with 24/7 uptime power consumption is a priority at the same level as decent performance.

For a fair comparison I will consider only NEW hardware.

As miniPC platform I think we all agree that today the most interesting one is N100, while a complete idle N100 system can absorb around 6W, a more realistic setup with things running on it will absorb around 14-20W. But N100 prices are no joke, at least in my country:

  • an N100 motherboard cost is between 120 and 140 €
  • +20 € for 8GB of DDR4
  • +20-30 € for an external PSU or a cheap ATX PSU

At the end of the day you'll spend at least 160 €, and I'm not considering the cost for a case.

As SBC ARM platform I still consider Raspberry PI as the reference board, the reason is quite simple, its support and its reliability still are the best imho, but as we know there's plenty of different produces and platform at lower costs.

  • RPi5 8GB can be easily found for 85 € in EU (or 80$ in the USA)
  • +6 € for the official cooler+fan
  • +13 € for the official PSU

The total cost starts from around 104 €

Now let's take a look to a real RPi5 8GB power consumption, included a USB SATA SSD, as you can see we're under 5W

https://preview.redd.it/odlvxj6fo30e1.png?width=881&format=png&auto=webp&s=b84fd1437dcc2262b2db88d672e964c30b54fc74

You may think this is a completely idle system, let me show you what I'm running constantly on this RPi5:

  • Authentik (+ dedicated Redis + dedicated Cloudflare daemon + dedicated PostgreSQL)
  • Bookstack (+ dedicated MySQL)
  • Gitea (+ dedicated MySQL)
  • Grafana
  • Prometheus
  • Got Your Back instance 1
  • Got Your Back instance 2
  • Got Your Back instance 3
  • Home Assistant
  • Immich (+ ML + dedicated PostgreSQL + dedicated Redis)
  • Jellyfin
  • PhpIPAM (+ dedicated MySQL + Cron application)
  • Pihole
  • Roundcube
  • Syncthing instance 1
  • Syncthing instance 2
  • Syncthing instance 3
  • Ubiquiti Unifi Network Application (+ dedicated MongoDB)
  • Vaultwarden (+ dedicated Cloudflare daemon)
  • Watchtower
  • Wireguard
  • Wordpress website 1 (+ dedicated MySQL + dedicated Cloudflare daemon)
  • Matomo website (+ dedicated MySQL + dedicated Cloudflare daemon)
  • Wordpress website 2 (+ dedicated MySQL + dedicated Cloudflare daemon)
  • Wordpress website 3 (+ dedicated MySQL + dedicated Cloudflare daemon)
  • Nagios

On top of that my RPi5 act as:

  • nas server for the whole family (samba and nfs)
  • backup server repository for the whole family (+ night sync on a 2nd nas server turned on via wake on lan and immediately turned off after sync + night sync on Backblaze B2)
  • Collectd server
  • frontend webserver for all the other services with Apache httpd

You may think performance is terrible... well

This is an example of SMB transfer rate from and to the RPi5 while running all the things I listed before.

https://preview.redd.it/uy9ug0lpp30e1.png?width=897&format=png&auto=webp&s=4cfe71ef2e58988e3adda3476ec85d9a386a8bb5

The websites and services response rate is... how can I say... perfect.

Previously I used VPS from OVH, from Hetzner, from other service providers, and honestly my websites performance were way worst, moving those sites to docker containers on RPi5 was a huge upgrade in terms of performance.

Considering the average cost of the electricity in my country:

  • a RPi5 will cost around 5,36 €/year
  • a N100 will cost 16 €/year for 15W of absorbed power, 21,43 €/year for 20W

This may not seems a lot of difference, but if you consider that in this scenario these two systems have no real performance difference, the power cost is very significant imho.

Some will argue the N100 can be easily expanded, fine but we're still talking about a single RAM slot with 2 SATA ports, and a single PCIe slot, in case of a RPi5 we have a PCIe expansion with plenty of hat boards (and also a 5 sata slots hat board available on the market), so the expandability argument is less and less significant imho.

Even the RAM expandability of a miniPC platform is not such a strong argument considering this kind of usage, 8GB is a good amount of RAM.

Just to have a comparison this is the RAM consumption of all the stuff I'm constantly running over my RPi5 I reported before, and as you can see from the sw list I'm not doing any optimization or service consolidation (any service requiring a database has it's own database instance, same for cloudflared)

https://preview.redd.it/cf98t9aip30e1.png?width=897&format=png&auto=webp&s=ceef81f2cc508207ac0559a4412a47fd16061547

As you can see at the end of the day a good old RPi can still be a strong contender as a home server:

  • it's easily available almost everywhere (luckily the shortage phase is ended a long time ago)
  • it's not as expensive as many people think
  • its performance are perfectly in line with a miniPC platform as home server
  • it's much more compact and easy to place everywhere in your home, and thanks to its power consumption you can place it even in a drawer if you want
  • it's way more flexible in terms of expandability compared to previous generations SBCs

Imho we have to be more honest and don't exclude ARM SBCs as home server platforms, in most case they're still the best solution imho.

65
 
 
The original post: /r/selfhosted by /u/mitousa on 2024-11-10 15:56:56.
66
 
 
The original post: /r/selfhosted by /u/Dan6erbond2 on 2024-11-10 15:53:35.

Hey everyone! I just released GoMailMark, a Go-based web application that allows users to generate email signatures using Templ and integration with S3 to upload assets needed in your signatures.

GoMailMark is perfect for teams or organizations that want a unified email signature design but also value the flexibility to customize. With a reusable template, you can easily maintain a consistent look across your team’s signatures. For users with some design experience, the HTML template at views/components/signature.templ is fully customizable to match your company’s brand or design needs.

Once you've modified the template, you can launch GoMailMark with go run main.go and head to http://localhost:8000 where you will be prompted to enter information about yourself such as name, role and email that will be injected into the email template.

https://preview.redd.it/bnvw8cioi30e1.png?width=1155&format=png&auto=webp&s=84ced7e1bb1c8e0937e018afafc38c44edbafc9b

If you want to use GoMailMark in other applications, it can generate signatures programmatically. Just send a multi-part form HTTP POST request to the / endpoint, and GoMailMark will return the HTML signature as the response.

I'm still working on this project, and will try to find an easy way to deploy it using Docker. However, due to its customizable nature it will be recommended to build your own image.

GoMailMark also supports configuration via environment variables as well as JSON and YAML configs, which is where it will pull the S3 config and default branding/company information from. Check out the README in the repo for more details!

67
 
 
The original post: /r/selfhosted by /u/Dry-Horror-5022 on 2024-11-10 15:47:23.

Im just the normal user who switch from portainer, and I think it’s a really good switch/alternative.

So Im shout out to dokploy after I spent 2 days to learn everything.

It was a bit slow for me as I use nginx mostly, here’s it’s using traefik so it’s the biggest challenge for me. Ask me questions if you want.

68
 
 
The original post: /r/selfhosted by /u/DensePineapple on 2024-11-10 15:07:16.

I've tried moving away from Gmail multiple times and it's impressive how all the alternatives make Google's abomination look decent. The most commonly suggested replacements all look like pet projects from 2000 and perform about as well. I've tried AfterLogic, Cypht, MailPile, MailCow, Mailu, Rainloop, RoundCube, SOGo, etc. and they all make me question if Thunderbird maybe wasn't so bad. Am I missing something?

69
 
 
The original post: /r/selfhosted by /u/schizovivek on 2024-11-10 14:38:13.

Edit:

The images look to have come to be too large. apologies for that

This is my current setup:

I currently have 4 devices.

  • 2 PIs
    • 1 PI is currently just being used for AdguardHome.
    • 1 PI is currently just used for RetroPie
  • 1 mini PC with a USB to 5 bay sata enclosure. (so the NAS)
    • The mini pc is the "power house". It handles everything else (emby, immich, caddy, arr suite etc)
  • PC that I work on and use to manage everything.
    • honestly now that I think about it I actually use the PC only as an interface and ssh into the miniPC to manage everything

Managing this was easy enough since all my deploys are currently on my miniPC. Recently I started thinking of splitting things up a little.

I'd like to split items into

  • light-weight with 24 hour availability (eg:)
    • changedetection
    • ntfy
    • caddy
    • adguardhome
    • prowlarr
  • high availability but needs more processing power (eg:)
    • immich
    • emby
    • photoprism
    • gitea
    • jdownloader
    • mediacms
  • emulation only
    • would like to have something like romm running on it to manage the game collection

https://preview.redd.it/vb8hprg2330e1.png?width=955&format=png&auto=webp&s=fbab402fa52d83d3d7ad704682d1b1d165f218bf

With the emulation part nothing really changes. Overall I'd like to be able to manage everything from my PC in a way that makes it so I don't have to log into each machine to deploy something on it.

Currently the compose files and config files are handled via a docker-stacks folder I have on my miniPC which I use to stow to my dockerapps folder. This way everything I need is versioned into gitea for docker configuration. Once stowed into the dockerapps folder I run an alias for the docker up command that handles the compose file being inside another folder. Eg for caddy all I do is dcf caddyand it handles the env file and the compose file

I started looking into how I can remotely deploy something and found docker contexts. In order to test it out I created a context pi on my miniPC and then in order to deploy it I had to stow the compose in my miniPC as well as my PI.

I currently handle the stowing part by creating a makefile which I've configured to stow to a specific location so that's not as much of work (now that I've figured it out :-P)

With this setup I tested for something simple as Caddy and can see that I sadly had 2 stows for the compose as well as the Caddyfile. 1 stow in the miniPC so I can run the docker compose up command and the other stow in the PI so that the Caddyfile is loaded where the app is actually deployed.

As I type, just jotting down what I feel I can do better

  • Now that I think about contexts I guess I can manage all the compose files from my PC rather than maintaining them on my miniPC.
  • I'd then need to create another repo to maintain the configuration files which I can clone and stow into the specific machines.
  • I guess I can update the makefile to handle the ssh part of the doing something on the specific boxes

I'm pretty sure there are better ways to do this and am willing to give something new a try, hence this post. I wanted to hear the thoughts of how others manage their stacks. Basically wondering how I can control the compose files and the configurations from a central location. Appdata would still have to lie in the specific devices which is fine. (apologies in advance since I'm not that great in articulating my thoughts :-|)

Screengrab of how my docker-stacks folder looks like

https://preview.redd.it/l1zjuw1f330e1.png?width=286&format=png&auto=webp&s=d583b4bd72b3cf721802910c026bcd4def0f8a4d

Screengrab of my docker-stacks makefile

https://preview.redd.it/3emrjkvw330e1.png?width=1468&format=png&auto=webp&s=390a6c50eb63840ecd50c6274dd668b05889aeb8

70
 
 
The original post: /r/selfhosted by /u/Lopsided-Painter5216 on 2024-11-10 14:03:15.

I thought I'd share the recent update here as it packs a lot of features people wanted.

  • RSS Subscriptions 🧪: You can now subscribe to RSS feeds and Hoarder will automatically pull them every hour.
  • The long awaited REST API is here.
  • AI Summarization: Hoarder now has the ability to generate summaries for the links that you hoard!
  • Video downloads: For ultimate hoarding, Hoarder now has the ability to download videos from the links that you hoard (e.g. youtube)
  • Omnivore Imports: Omnivore is shutting down. You can now import your bookmarks from Omnivore into Hoarder.
  • User Management: Admins can now reset user passwords, change their roles and even create new users
  • Image OCR: Hoarder will now attempt to read and index text in the images that you hoard!

You can see the full changelog here

71
 
 
The original post: /r/selfhosted by /u/radioturn on 2024-11-10 13:31:56.

Hey everyone! ive been on the selfhosting journey for a few years now. I was wondering if anyone had any recommendations for a wiki similar to dokuwiki that also is centered around the markdown? i like the folders structure, the plain text, and plugins. Ive tried a markdown extension for dokuwiki but its not that great. On my phone and computer i use obsidian. If theres anything simple like that i would love to hear about it!

some of the ones i already played with:

  • WikiJS
  • Dokuwiki
  • Bookstack
  • Obsidian (local)
72
 
 
The original post: /r/selfhosted by /u/ArdaOneUi on 2024-11-10 13:17:12.

Im new to this and wanted to ask if there is a way to have a self hosted cloud that will reliably backup your gallery. I have a samsung phone and OneDrive is integrated into the gallery which means it automatically syncs up all pictures/video. Is there a way to do the same on my own?

73
 
 
The original post: /r/selfhosted by /u/candle_in_a_circle on 2024-11-10 13:14:49.

I'm closing down and archiving an old Google Apps tenant and want to allow ~5 of the users to access their old emails, just a few GB each (and files and contacts, but that's a separate problem).

I have a self-hosted setup (docker server, web-facing reverse proxies, authelia) which I can use. I've seen recommendations in this sub for running Dovecot as a store and Roundcube as a webmail front-end. Is this the best way?

I do need:

  • To be able to import or natively read the Gmail All.Mail MBOX
  • A webmail interface
  • Multi-user support
    • auth via authelia (or authentik)
  • Half-decent search ability
  • The stack to run on Docker

I do not need:

  • To be able to receive emails
  • To be able to send emails
  • To be able to connect any other client to the IMAP

Given the thing I don't need, Dovecot seems overkill?

74
 
 
The original post: /r/selfhosted by /u/hanneshier on 2024-11-10 13:07:26.

Hi everyone, I just switched my homeserver to Proxmox recently and I’m planning to implement Proxmox Backup Server in the near future. But I’m wondering on any best practices regarding backups of my docker LXC.

If I understand it correctly, i can only restore the entire LXC, but since I have several docker container in one LXC, that’s not very convenient in case I just need to rollback one docker container. Do you have any best practices for such a scenario?

One idea I have is to use one (zfs) mount point per docker container and store the docker volumes for that container there. Would that be a viable way?

75
 
 
The original post: /r/selfhosted by /u/Mission-Balance-4250 on 2024-11-10 12:37:02.

Seriously, Thunderbird sucks... Apple is legendary for syncing iOS Calendar and Reminders with my Radicale CalDAV server. No desktop client is even half useful. I'm surprised nobody has built a simple web app that can synchronise with a standalone CalDAV server.

  • View/edit tasks
  • View/edit calendar events

It doesn't have to be fancy. Do I build one myself? Would anyone use it? I feel like I must be grossly missing the reason nothing exists...

view more: ‹ prev next ›