• 0 Posts
  • 13 Comments
Joined 2 years ago
cake
Cake day: July 29th, 2023

help-circle
  • You might want to check what the actual hardware is first. You’ll probably be fine, but client 802.11 hardware can sometimes be underwhelming for hosting because they don’t have good stuff like beefed up MuMIMO.

    Although that’s assuming you will have a lot of traffic going through it, so you could always just test throughput and latency with iperf to see how well it functions.


  • mlg@lemmy.worldtoSelfhosted@lemmy.worldSelf host websites
    link
    fedilink
    English
    arrow-up
    4
    ·
    2 months ago

    It depends on what it is really + convenience. There are lots of morons out here running basic info sites on full beefy datacenter VMs instead of a proper cloud webhost service.

    The most you’d be getting out of cloud is reliability. Self host assumes you don’t have any bottlenecks (easy enough to pass), but also 99% uptime which is impossible unless you are running with site redundancy (also possible, but I doubt how many people own multiple properties with their own distribute or private cloud solution).

    if 95% uptime is acceptable, and you don’t live in an area with outage issues from weather, I’d say go for it. Otherwise, you can find some pretty cheap cloud solutions for basic websites. Even a cheapo VPS would probably work just fine.


  • I have run photoprism straight from mdadm RAID5 on some ye olde SAS drives with only a reduction in the indexing speed (About 30K photos which took ~2 hours to index with GPU tensorflow).

    That being said I’m in a similar boat doing an upgrade and I have some warnings that I have found are helpful:

    1. Consumer grade NVMEs are not designed for tons of write ops, so they should optimally only be used in RAID 0/1/10. RAID 5/6 will literally start with a massive parity rip on the drives, and the default timer for RAID checks on Linux is 1 week. Same goes for ZFS and mdadm caching, just proceed with caution (ie 321 backups) if you go that route. Even if you end up doing RAID 5/6, make sure you get quality hardware with decent TBW, as sever grade NVMEs are often triple in TBW rating.
    2. ZFS is a load of pain if you’re running anything related to Fedora or Redhat, and the performance implications from lots and lots of testing is still arguably inconclusive on a NAS/Home lab setup. Unless you rely on the specific feature set or are making an actual hefty storage node, stock mdadm and LVM will probably fulfill your needs.
    3. Btrfs has all the features you need but is a load of trash in performance, highly recommend XFS for file integrity features + built in data dedup, and mdadm/lvm for the rest.

    I’m personally going with the NVME scheduled backups to RAID because the caching just doesn’t seem worth it when I’m gonna be slamming huge media files around all day along with running VMs and other crap. For context, the 2TB NVME brand I have is only rated for 1200 TBW. That’s probably more then enough for a file server, but for my homelab server it would just be caching constantly with whatever workload I’m throwing at it. Would still probably last a few years no issues, but SSD pricing has just been awful these past few years.

    On a related note, Photoprism needs to upgrade to Tensorflow 2 so I don’t have to compile an antiquated binary for CUDA support.


  • mlg@lemmy.worldtoFediverse@lemmy.worldI really want to like Lemmy
    link
    fedilink
    English
    arrow-up
    15
    arrow-down
    1
    ·
    4 months ago

    Jokes on you the political content here is from the redditors who pretended to quit their award fueled addiction by also joining lemmy.

    Seriously though, compare c/Politics to c/Worldnews or c/News. There is a very large dissonance between the comments shared despite both communities posting the same news info…









  • mlg@lemmy.worldtoFediverse@lemmy.world*Permanently Deleted*
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    2
    ·
    2 years ago

    make a decentralized system that scales quickly and efficiently with usage and has a higher uptime than any server based site like reddit

    It is like the holy grail of a decentral web, and it is (currently) very hard to achieve without making a tradeoff.

    Lemmy trades off scalability and availability for federation. It’s not very effective because instances have to be selected manually, and if any large instance goes down, chaos ensues.

    It’s not even truly decentral. More like a somewhat redundant media system.

    The best I think we have right now is IPFS. It has scalability and accessibility. But it lacks speed due to DHT searches being hella slow compared to a simple DNS lookup.


  • I am insane and use bare bone LXC.

    Stupid ramblings you can probably ignore:

    spoiler

    Usually though it’s because I run most stuff bare metal anyway so LXC is for temporary or random cases where I need a weird dependency or I want to run a niche service.

    Only use docker for when I actually want faster setup like docker-osx which does all the vm stuff for running a virtual Mac for you.

    I don’t really mind docker, but for homelab I just find myself rewriting dockerfile anytime I want to change something which I don’t really need to do if I’m not publishing it or even reusing it.

    Kubernates is really more effective for actual load services, which you never need in homelab lol. It’s great to use to learn k8s cluster, but the resources get eaten fast.