• 5 Posts
  • 56 Comments
Joined 1 year ago
cake
Cake day: June 16th, 2023

help-circle


  • As someone who has owned enterprise servers for self-hosting, I agree with the previous comment that you should avoid owning one if you can. They might be cheap, but your longterm ownership costs are going to be higher. That’s because as the server breaks down, you’ll be competing with other people for a dwindling supply of compatible parts. Unlike consumer PCs, server hardware is incredibly vendor locked. Hell, my last Proliant would keep the fans ramped at 100% because I installed a HDD that the BIOS didn’t like. This was after I spent weeks tracking down a disk that would at least be recognized, and the only drives I could find were already heavily used.

    My latest server is built with consumer parts fit into a 2U rack case, and I sleep so much easier knowing I can replace any of the parts myself with brand new alternatives.

    Plus as others have said, a 1U can be really loud. I don’t care about the sound of my gaming computer, but that poweredge was so obnoxious that despite being in the basement, I had to smother it with blankets just so the fans didn’t annoy me when I was watching TV upstairs. I still have a 1U Dell Poweredge, but I specifically sought out the generation that still let you hack the fan speeds in IPMI. From all my research, no such hack exists for the Proliant line.


  • The problem with chromebooks is that the base specs are pretty shit. A lot of them have 4 GiB of RAM and maybe 16GiB of disk if you’re lucky.

    They were designed to be thin clients to connect students to the internet, and little else. Maybe they could be hacked into something useful, but I don’t think it’ll ever make a good PC. They were always destined for the landfill.

    Meanwhile, the best thinkpads were quality machines back when they came out. IMO, that’s why they’re still so versatile today. Free software can’t fix bad fundamentals.


  • I did this in my town. It was a local theater though, not a big chain

    • Went on a Wednesday night and rented access to a screening room for ~5h. We had to pay extra because they normally close earlier on a weeknight
    • Showed up early to meet the projectionist and get a technical rundown
    • Setup was a friend’s laptop plugged into HDMI running up to the projector booth
    • Nobody complained when we brought out a disk full of torrented movies
    • The theater already had a license to sell alcohol, so we had that covered
    • We brought a small bit of outside food, and nobody complained

    It was absolutely the best time I ever had at a cinema. When the evening wound down, the projectionist invited us into the back area for a tour of the projection equipment.

    I think that because we were a private event the rules about screening copyrighted materials to public audiences did not apply.


  • Not sure what motherboard you have: Most consumer boards only support “FakeRAID”, which requires a kernel driver to actually function. Good luck finding a vendor who wrote a driver for Linux.

    I’d definitely recommend software RAID instead, as you’ll have better support. I like btrfs, so I’d recommend you set up your new drives to use a btrfs RAID configuration. mdadm is another option, if you really like ext4.




  • I’d recommend BTRFS in RAID1 over hardware or mdadm raid. You get FS snapshotting as a feature, which would be nice before running a system update.

    For disk drives, I’d recommend new if you can afford them. You should look into shucking: It’s where you buy an external drive and then remove (shuck) the HDD from inside. You can get enterprise grade disks for cheaper than buying that same disk on its own. The website https://shucks.top tracks the price of various disk drives, letting you know when there are good deals.










  • For backup, maybe a blu-ray drive? I think you would want something that can withstand the salty environment, and maybe resist water. Thing is, even with BDXL discs, you only get a capacity of 100GiB each, so that’s a lot of disks.

    What about an offsite backup? Your media library could live ashore (in a server at a friend’s house). You issue commands from your boat to download media, and then sync those files to your boat when it’s done. If you really need to recover from the backup, have your friend clone a disk and mail it to you.

    Do you even need a backup? Would data redundancy be enough? Sure if your boat catches fire and sinks, your movies are gone, but that’s probably the least of your problems. If you just want to make sure that the salt and water doesn’t destroy your data, how about:

    1. A multi-disk filesystem which can tolerate at least 1 failure
    2. Regular utilities scanning for failure. BTRFS scrubs, for example.
    3. Backup fresh disks kept in a salt and water resistant container (original sealed packaging), to swap any failing disk, and replicate data from any good drives remaining.
    4. Documentation/practice to perform the aforementioned disk replacement, so you’re not googling manpages at sea.

    This would probably be cheapest and have the least complexity.



  • How about writing a script to automate the deletion, thus minimizing the chance of human error being a factor? It could include checks like “Is this a folder with .git contents? Am I being invoked from /home/username/my_dev_workspace?”

    In a real aviation design scenario, they want to minimize the bullshit tasks that take up cognitive load on a pilot so they can focus on actually flying. Your ejector seat example would probably be replaced with an automatic ejection system that’s managed by the flight computer.