I know for many of us every day is selfhosting day, but I liked the alliteration. Or do you have fixed dates for maintenance and tinkering?
Let us know what you set up lately, what kind of problems you currently think about or are running into, what new device you added to your homelab or what interesting service or article you found.
This post is proudly sent from my very own Lemmy instance that runs at my homeserver since about ten days. So far, it’s been a very nice endeavor.
I’m building services out for my family as things enshittify. Moved the family over to an immich instance, run a family blog on Wordpress (working in rolling my own since it’s over complicated and with all the Wordpress shenanigans…), plex (lifetime account, works for now). I have a number of self-built projects as well, a “momboard” like system that is integrated with my Wordpress blog for access and control, a pi based backup server that lives at my friends house and nails a VPN connection to my router and I’m playing with Meshtastic as an offline communication system for my kids scout troop when we’re camping without cell signal. Lots of home automation with home assistant as well.
I host it all on Debian servers, raspberry pi’s and esp32 devices (Meshtastic and home automation). I used to run kubernoodles but it was more complicated than needed and for my use case, docker, ansible and bash scripts manage it all just fine.
Been messing around w/ podman, and after hours of slamming my head against the wall, I decided Seafile isn’t worth it. :) It launches a bunch of stuff inside one container, and I just couldn’t figure out how to get that to work w/ quadlet (worked fine w/
podman kube play
though).I got forgejo set up and now I’m looking into setting up runners so I can finally migrate off hosted gitlab onto my own forgejo instance.
Some other things I’m planning on doing this week:
- migrate existing services to podman quadlet from docker compose - will make each existing service into a pod and play w/ pod networking
- set up technitium - tested it locally and it worked well, so just need to move it and configure it; hope to use it as the primary DNS for my house
- set up owncloud ocis - there’s a new POSIX FS option, which was my main hangup when I last looked into a nextcloud alternative (I only need storage + collabora)
- probably some kind of dashboard, because the number of services I host is getting a bit long
If I get time, I want to install openSUSE MicroOS onto my NAS and start migrating everything to it (from openSUSE Leap). I really like the idea of an immutable base OS, and my NAS is already 90% containers (pretty much just Samba left). I need to fix some permission issues anyway (keep having to
chown
my videos so samba and jellyfin can work together), and this should make things a bit more obvious.I’ll probably also start a blog about my self-hosting journey, because the info around podman is kinda sparse, especially when it comes to quadlet.
I just set up wanderer and workout-tracker. Along with installing gadgetbridge on my phone, I now have a completely self hosted fitness/workout stack with routes, equipment tracking, heatmaps, general health metrics like HRV, heart rate, etc through my Garmin watch, without having Garmin Connect installed. Awesome!
That sounds so cool! Not using any tracking/nav devices other than my phone but currently my routes just stay local without having any kind of management for them.
I’m working on my first kubernetes cluster. I’m trying to set the systems up with NixOS. I can get a kublet and a control plan running. But I’m getting permission errors when trying to use kubeclt rootless on the system running the control plane. I think I figured out which file i need to change, now I just want to record that change in my configuration.nix.
I’m curious how this goes for you. I run all my machines on NixOS except my k8s cluster which is Talos for now. I have been thinking of switching to Nix for that too.
nixos doesn’t play well with rootless containers in my experience
I need to migrate off Docker Desktop for Windows and Storage Spaces but I fear the process will be difficult due to my data volume and the stupidity of Windows. I should never have gone Windows, but I wanted to use Steam Big Picture off the media PC and didn’t want to deal with getting that functional on Linux.
But Docker Desktop for Windows keeps crashing WSL and bricking the network devices randomly, and also continuously grows memory consumption until the machine reboots. Piece of shit.
Windows Docker is so bad, I don’t even know why it’s a thing.
Some good planning might make the migration less painful. I would recommend a ZFS or other COW storage solution under the docker host so you can do snapshot backups and not have to worry about quiesing databases, etc.
Yea I’m gonna do zfs or something when I get set up properly again. I’ve got 2 16TB HDDs and Storage Spaces won’t let me pull a drive out :v
I think I’m gonna have to make a new Storage Space and slowly grow that one and shrink the other as I basically shift the extra storage budget between the two until the data is on just one of my drives without redundancy, and then I’ll pull that drive, dual boot Ubuntu or something, format, get everything prepared, and then mount, copy, start services, and then go back and kill the old storage spaces and then never run Windows for anything meaningful again.
I got a Matrix server set up with conduwuit but the problem is that none of my friends are on there so I don’t use it. The one friend I made the damn thing for so we could chat just started going through a bunch of personal stuff so now it won’t be used for a while. FML.
Cool to have it ready anyways! Does it federate? You can use all sorts of dev-support groups etc.
I started hosting audiobookshelf since Jellyfin was pretty clunky for audiobooks.
how is your experience with it? I’m considering setting up audiobook shelf as well.
Had the intention of making a hidden TOR website version for all my websites but I’m sick
Oh, sounds pretty cool, I have never looked into that.
After just about a month of hosting some things on a Raspberry Pi 4, I think it’s about time to work on repurposing this mini PC that hasn’t been doing much the last few years and keep growing my services.
To that end, can anyone point me to a good, thorough guide to getting going with Sonarr? I installed it, but then realized I needed to add a client and Prowlarr and I feel like I just started in the middle.
Search for trash guides and servarr. Both have websites that are detailed in how to set up all of the arrs apps in what ever fashion you want. I think both have Discord servers too.
Added extra disks to TrueNAS, got Seafile up and running in a Proxmox VM. Now I’m about to start fiddling with SAS to 4x Sata to get the front drive bays working. Keepin’ busy!
Pinepods 0.7.4 is out! So as the Dev I’m going through new issues and knocking them out. Smart playlists, oidc logins and notifications on release are all a thing now on the self hosted podcast platform! We’re nearing a v1 release with features on par with some of the big time podcast apps.
Hell yeah! Still got Pinepods on my to-host list.
I got a new job, and the group chat is on WhatsApp, so I’m looking into running a Synapse server with a bridge to it. I really don’t want to have to use Meta’s apps on my phone.
From what I’ve read so far, it seems like it’s going to be the most convoluted install process I’ll have encountered in my self-hosting journey. I’m excited to tackle it, but also a bit overwhelmed. Which is why I’ve been putting it off :P
It was a huge pita to get it running, but I have it.
One thing about the WA bridge is that element won’t let me give display names or look up the contact number, so the people in chatting with don’t have names, just “their number (WA)”
Try conduwuit instead of Synapse if you get stuck. For me, it was really simple to install and the dev is really nice.
Holy crap, you’re me. Except I plan on using slidge-whatsapp.
Spring break so nothing this weekend. I need to figure out backups and then common passwords/logins for my family.
switched my server from i7-870 (my ex-workstation) to Pentium G6405 (got it free). switch went without a hitch, debian with a ton of docker services (jellyfin, servarr, pihole, radicale, etc.), 8 GB RAM only. although it’s a quadcore to dualcore switch, no performance issues. I know there are better options out there, but I don’t spend money unless I really have to.
That G6405 is actually about 25% faster overall and 50% faster per thread, so performance should be better now. Not to mention much faster RAM and IO.
Core count doesn’t mean much when the CPUs are 12 years apart!
sure, that was the point - skip 10 gens and have zero issues, same software runs as-a before (signor roberto voice).
Not to mention all the extra instruction sets the newer CPU supports. The i7-870 is old enough that it doesn’t even support AES-NI, so encryption/decryption is significantly slower compared to even the lowest-end modern Intel or AMD x86 CPU.
what’s maintenance? is that when an auto-update breaks everything and you spend an entire weeknight looking up tutorials because you forgot what you did to get this mess working in the first place?
I do love how little maintenance is needed until you have to re-learn everything you forgot
I know you’re half joking. But nevertheless, I’m not missing this opportunity to share a little selfhosting wisdom.
Never use auto update. Always schedule to do it manually.
Virtualize as many services as possible and take a snapshot or backup before updating.
And last, documentation, documentation, documentation!
Happy selfhosting sunday.
I think auto update is perfectly fine, just check out what kind of versioning the devs are using and pin the part of the version that will introduce breaking changes.
I just like it when things break on scheduled maintenance and I have time to fix it or the possibility to roll back with minimal data loss, instead of an auto update forcing me spend a week night fixing it or running a broken system till I have the time.
You can have the best of both worlds - scheduled auto updates on a time that usually works for you.
With growing complexity, there are so many components to update, it’s too easy to miss some in my experience. I don’t have everything automated yet (in fact, most updates aren’t) but I definitely strive towards it.
In my experience, the more complex a system is, the more auto updates can mess things up and make troubleshooting a nightmare. I’m not saying auto updates can’t be a good solution in some cases, but in general I think it’s a liability. Maybe I’m just at the point where I want my setup to work without the risk of it breaking unexpectedly and having to tinker with it when I’m not in the mood. :)
No you just continue updating until it’s fixed again.
Yes