

Cool, now we’ll get an influx into lemmy.world and I’ll finally have a reason to abandon this account lol
Cool, now we’ll get an influx into lemmy.world and I’ll finally have a reason to abandon this account lol
JVM and Android dalvik ART are still alive and well because if we could use clown circus Javascript to run WWW for 30 years, we sure as hell can use “My Big Fat Gabrage Collector: The Boilerplate Saga” to run all of our applications and backend infrastructure.
Yeah it’s great because even without a make plugin, you can just add your make command to the vscode actions that’ll run your makefile.
Or even better, get the plugin which will auto populate targets from the makefile lol
Believe it or not, NTFS isn’t, but Windows is to keep ye olde DOS compatibility lol.
TIL case insensitive filesystems are still a thing actually in use.
Why lol
You might want to check what the actual hardware is first. You’ll probably be fine, but client 802.11 hardware can sometimes be underwhelming for hosting because they don’t have good stuff like beefed up MuMIMO.
Although that’s assuming you will have a lot of traffic going through it, so you could always just test throughput and latency with iperf to see how well it functions.
It depends on what it is really + convenience. There are lots of morons out here running basic info sites on full beefy datacenter VMs instead of a proper cloud webhost service.
The most you’d be getting out of cloud is reliability. Self host assumes you don’t have any bottlenecks (easy enough to pass), but also 99% uptime which is impossible unless you are running with site redundancy (also possible, but I doubt how many people own multiple properties with their own distribute or private cloud solution).
if 95% uptime is acceptable, and you don’t live in an area with outage issues from weather, I’d say go for it. Otherwise, you can find some pretty cheap cloud solutions for basic websites. Even a cheapo VPS would probably work just fine.
I have run photoprism straight from mdadm RAID5 on some ye olde SAS drives with only a reduction in the indexing speed (About 30K photos which took ~2 hours to index with GPU tensorflow).
That being said I’m in a similar boat doing an upgrade and I have some warnings that I have found are helpful:
I’m personally going with the NVME scheduled backups to RAID because the caching just doesn’t seem worth it when I’m gonna be slamming huge media files around all day along with running VMs and other crap. For context, the 2TB NVME brand I have is only rated for 1200 TBW. That’s probably more then enough for a file server, but for my homelab server it would just be caching constantly with whatever workload I’m throwing at it. Would still probably last a few years no issues, but SSD pricing has just been awful these past few years.
On a related note, Photoprism needs to upgrade to Tensorflow 2 so I don’t have to compile an antiquated binary for CUDA support.
Even worse is finding a rejected pull request that would have added the exact feature you need.
I think VSCode is the only stable electron application and even then it took them like 5 years to reach passable stability lol.
Used to crash and combust all the time when I first tried it.
Hey at least it’s not JavaScript which is perpetually high on crack with Object object
Does jellyfin do untranscoded video/audio?
Haven’t used it in years but finally building up my media server again and I remember it had some funky settings for hardware encoding back then which I didn’t need because I was connecting to it via a repurposed gaming laptop that could easily handle 4k content and surround sound by itself.