traches

joined 2 years ago
[–] traches@sh.itjust.works 18 points 2 days ago

I mean if it’s worked without modification for 6 years….

[–] traches@sh.itjust.works 7 points 5 days ago

Does what I want and gets out of my way.

[–] traches@sh.itjust.works 6 points 5 days ago

I bet he trained it on /pol/

[–] traches@sh.itjust.works 29 points 5 days ago (7 children)

I wish he didn’t feel the need to be so defensive about his choices. Bazzite is perfect for this use case

[–] traches@sh.itjust.works 2 points 6 days ago

Life is too short to wash spoons with your hands.

[–] traches@sh.itjust.works 9 points 6 days ago (7 children)

Don’t pre-rinse, just scrape the bigger bits into the trash. If your dishwasher can’t handle it there’s something wrong with it.

[–] traches@sh.itjust.works 10 points 1 week ago (6 children)

but they aren’t parallel

[–] traches@sh.itjust.works 1 points 1 week ago

Yeah, I’ll probably switch eventually I’m just trying to talk myself out of it because I don’t have the time to learn right now

[–] traches@sh.itjust.works 2 points 1 week ago (2 children)

I have a desktop, laptop, and a few VMs and servery things. Dotfile manager (yadm, which is a git wrapper) to sync personal settings, everything else I just do manually. The system-level configs are either different enough that standardizing them isn’t very helpful, or no more complicated than installing packages and activating services.

[–] traches@sh.itjust.works 37 points 1 week ago (40 children)

I like the idea of nixos, but I feel like it makes a bunch of daily sacrifices in order to optimize a task I do once every few years? I hardly ever get a new computer, but I install/uninstall/update/tweak packages on my system all the time. With a dotfile manager and snapshots, I get most of the benefit without any of the drawbacks.

[–] traches@sh.itjust.works 3 points 2 weeks ago

The desktop environment is all the stuff like the taskbar, the settings menus, the application launcher, the login screen, that kind of thing. It’s the system level user interface.

You choose which one by which distro you download. Linux mint uses cinnamon, Ubuntu and fedora use gnome. There are “flavors” of Ubuntu and fedora that use KDE. That’s why I suggested ventoy: you can download a few different ones and boot into them without making a new thumb drive.

If you don’t feel like bothering with any of that, just use Linux mint. It’s good.

[–] traches@sh.itjust.works 79 points 2 weeks ago* (last edited 2 weeks ago) (10 children)
  • before you switch, sort out your apps. Look at what you use on windows, see if it runs on Linux. If not, find a replacement that does and test it out.
  • Most Linux distros can boot into a desktop from a thumb drive. You can play and test without touching your windows installation.
  • in that vein, ventoy is neat. You can make a bootable drive and drop ISOs in a folder to boot from. No messing with etcher or whatever it’s called
  • desktop environment matters as much as the distro. Check out gnome, KDE, and cinnamon.
 

I'm working on a project to back up my family photos from TrueNas to Blu-Ray disks. I have other, more traditional backups based on restic and zfs send/receive, but I don't like the fact that I could delete every copy using only the mouse and keyboard from my main PC. I want something that can't be ransomwared and that I can't screw up once created.

The dataset is currently about 2TB, and we're adding about 200GB per year. It's a lot of disks, but manageably so. I've purchased good quality 50GB blank disks and a burner, as well as a nice box and some silica gel packs to keep them cool, dark, dry, and generally protected. I'll be making one big initial backup, and then I'll run incremental backups ~monthly to capture new photos and edits to existing ones, at which time I'll also spot-check a disk or two for read errors using DVDisaster. I'm hoping to get 10 years out of this arrangement, though longer is of course better.

I've got most of the pieces worked out, but the last big question I need to answer is which software I will actually use to create the archive files. I've narrowed it down to two options: dar and bog-standard gnu tar. Both can create multipart, incremental backups, which is the core capability I need.

Dar Advantages (that I care about):

  • This is exactly what it's designed to do.
  • It can detect and tolerate data corruption. (I'll be adding ECC data to the disks using DVDisaster, but defense in depth is nice.)
  • More robust file change detection, it appears to be hash based?
  • It allows me to create a database I can use to locate and restore individual files without searching through many disks.

Dar disadvantages:

  • It appears to be a pretty obscure, generally inactive project. The documentation looks straight out of the early 2000s and it doesn't have https. I worry it will go offline, or I'll run into some weird bug that ruins the show.
  • Doesn't detect renames. Will back up a whole new copy. (Problematic if I get to reorganizing)
  • I can't find a maintained GUI project for it, and my wife ain't about to learn a CLI. Would be nice if I'm not the only person in the world who could get photos off of these disks.

Tar Advantages (that I care about):

  • battle-tested, reliable, not going anywhere
  • It's already installed on every single linux & mac PC , and it's trivial to put on a windows pc.
  • Correctly detects renames, does not create new copies.
  • There are maintained GUIs available; non-nerds may be able to access

Tar disadvantages:

  • I don't see an easy way to locate individual files, beyond grepping through snar metadata files (that aren't really meant for that).
  • The file change detection logic makes me nervous - it appears to be based on modification time and inode numbers. The photos are in a ZFS dataset on truenas, mounted on my local machine via SMB. I don't even know what an inode number is, how can I be sure that they won't change somehow? Am I stuck with this exact NAS setup until I'm ready to make a whole new base backup? This many blu-rays aren't cheap and burning them will take awhile, I don't want to do it unnecessarily.

I'm genuinely conflicted, but I'm leaning towards dar. Does anyone else have any experience with this sort of thing? Is there another option I'm missing? Any input is greatly appreciated!

 

I have a load-bearing raspberry pi on my network - it runs a DNS server, zigbee2mqtt, unifi controller, and a restic rest server. This raspberry pi, as is tradition, boots from a microSD card. As we all know, microSD cards suck a little bit and die pretty often; I've personally had this happen not all that long ago.

I'd like to keep a reasonably up-to-date hot spare ready, so when it does give up the ghost I can just swap them out and move on with my life. I can think of a few ways to accomplish this, but I'm not really sure what's the best:

  • The simplest is probably cron + dd, but I'm worried about filesystem corruption from imaging a running system and could this also wear out the spare card?
  • recreate partition structure, create an fstab with new UUIDs, rsync everything else. Backups are incremental and we won't get filesystem corruption, but we still aren't taking a point-in-time backup which means data files could be inconsistent with each other. (honestly unlikely with the services I'm running.)
  • Migrate to BTRFS or ZFS, send/receive snapshots. This would be annoying to set up because I'd need to switch the rpi's filesystem, but once done I think this might be the best option? We get incremental updates, point-in-time backups, and even rollback on the original card if I want it.

I'm thinking out loud a little bit here, but do y'all have any thoughts? I think I'm leaning towards ZFS or BTRFS.

 

Not sure about the artist, sorry

 
view more: next ›