UI doesn’t come up until database migrations fully complete. Can take half an hour or more depending on how much content is indexed in your instance.
Lead admin for https://lemmy.tf, tech enthusiast
UI doesn’t come up until database migrations fully complete. Can take half an hour or more depending on how much content is indexed in your instance.
Your title should be “fuck subscriptions, except subscriptions from this site pulled from 1998” since everything in your guide relies on a paid debrid sub.
Uhh… if your script is subbing to 24k remote communities, those will continue to grow from then on, unless you start purging communities at some point. After one user subscribes to a community, all new content gets indexed and stored on your instance. Pict-rs can cache images short term (and eventually clear them out), but Postgres will start growing very quickly and never slow down until it fills up disks.
The lawsuit does not involve Germany in the slightest
I’ve had access to 4 for several weeks and it’s not really much better. Maybe I’m just asking too much of it though.
Strongly disagree about gpt being excellent for code, it’s extremely confident about the wrong answer most of the time. I’ve found it to be mildly useful as a Stack Overflow alternative (for asking general questions and having it point me in some direction) but it’s code outputs are garbage.
These instructions won’t work in anyone’s unraid box, even if they compile compose from source. Not sure why people think posting random chatGPT’d instructions is remotely useful.
I’m just letting mine do whatever it wants, got plenty of local storage. If/when I have storage issues I’ll add an s3 bucket, pretty easy to modify the entrypoint for pictrs to pass s3 connection info in the docker-compose deployment.
I spun up Firefly a few months ago and had about three weeks where I was actively categorizing transactions and reconciling everything and then my ADD kicked in. Really cool tool but I just need something low-maintenance for budget tracking.
From what I’ve seen and read, server to server traffic is less taxing on instances than client to server. So even if your instance is JUST you, it would be your instance talking to everything else so it would have some net benefit on the federation. But it would take a lot of users self-hosting solo instances for this to help in any noticeable way, I’d think.
There is certainly no downside to running a solo instance, if you’re even slightly interested I would say go for it!
I’m one of the other Lemmy.tf admins and I’ll share a bit. We’re currently on the docker-compose deployment from the repo, running on a VM with 4c/8gb ram/256gb disk. It’s on a baremetal VMware box at OVH with loads of resources to expand as needed.
I’m hoping we get enough users on here to force me into converting to a Helm chart and moving this to my Kubernetes cluster. Pod scaling would help address some of the issues larger instances are starting to run into, and it seems like a fun project.
As for Unraid, your best bet is to see if you can install docker-compose on it. This thread from 2020 suggests it should be possible, but the binary may not persist restarts. If you can’t use compose you would probably have to strip it apart and deploy one container at a time, and potentially work around the need for the Docker networks.
I may be interested in helping with an Unraid deployment guide if there’s heavy interest- I’m running it on my NAS at home and can tinker a bit. Feel free to DM me if you’ve got questions or need any assistance.
Edit: That Unraid forum post has a reply about using a bash alias to run docker-compose in Docker, this is the route I’d go rather than having to do jank stuff to make the binary persistent. Should be able to follow the normal docker-compose install from your root user once you have compose ready. Make sure to do your port forwarding or use Nginx Proxy Manager since SSL is mandatory to federate.
I’m interested, but I don’t know Rust and haven’t done frontend work in years. Might be able to do some work around scalability and contribute to a Kubernetes deployment guide (and/or Helm chart).
Yes, I’ve got separate subnets & vlans for a few things. My PCs/phone/tablets/etc, homelab, IoT devices (i.e. loads of Govee bulbs/ropes, gaming consoles, oven, etc), Guest (all isolated from everything else internal) and one for my roommate. I’m on a Unifi Dream Machine Pro so setting up traffic rules to allow certain traffic from PC vlan to homelab (and the other way) was pretty straightforward.
As for the VPN, yes a full tunnel would force all traffic over the VPN, but for all but my *arr stuff that’s overkill. I just join all my VMs to Zerotier and force traffic from the public LB in via their VPN IP, but the VMs can still pull yum updates and anything else they want over my WAN link.
4vcpu (Ryzen), 8GB RAM, 256gb disk (which will be expanded when it gets to like 60% full). Not too worried about storage unless I get a bunch of image-happy users, text all comes in as json and goes straight to Postgres so it’s not a concern.
I run all my lab servers/services/etc in their own /16 on my home net. Nothing is publicly routed in over my WAN IP- if I want to expose a service, it goes through Nginx Proxy Manager to my local service via a ZeroTier tunnel.
I would strongly encourage you to not expose any of the *arr services (particularly your download node) to your WAN IP. PIA’s desktop app does a pretty good job of forcing a full tunnel with a VPN kill switch, so you never have to worry about your ISP catching onto what you’re doing.
Small instances don’t seem to require anything major, I’m running mine on a VM with 4c/6gb ram/256gb disk with no issues- it’s just a few Docker pods. Just make sure you use a dynamic DNS provider if you’re hosting from home, as valid SSL is required to connect to the federation.
The vast majority of the popular accounts are not run by the women on the profile. Most of them pay friends or agencies to manage the page for them, they simply show up to photo shoots every now and then and enjoy the easy money.