[-] rutrum@lm.paradisus.day 1 points 16 hours ago

I learned SQL before pandas. It's still tabular data, but the mechanisms to mutate/modify/filter the data are different methodologies. It took a long time to get comfy with pandas. It wasnt until I understood that the way you interact with a database table and a dataframe are very different, that I started to finally get a grasp on pandas.

[-] rutrum@lm.paradisus.day 0 points 18 hours ago

If it works, don't fix it!

[-] rutrum@lm.paradisus.day 3 points 18 hours ago

A big feature of polars is only loading applicable data from disk. But during exporatory data analysis (EDA) you often have the whole dataset in memory. In this case, filters wont help much there. Polars has a good page in their docs about all the possible optimizations it is capable of. https://docs.pola.rs/user-guide/lazy/optimizations/

One I see off the top is projection pushdown, which only selects relevant columns for a final transformations. In pandas, if you perform a group by with aggregation, then only look at a few columns, you still perform aggregation across all the data. In polars lazy API, you would define the entire process upfront, and it would know not to aggregate certain columns, for instance.

[-] rutrum@lm.paradisus.day 13 points 1 day ago* (last edited 1 day ago)

Imo Rust already has the perfect book. I would make a resource for C developers. Especially since you know C already.

[-] rutrum@lm.paradisus.day 4 points 1 day ago

Its a paradigm shift from pandas. In polars, you define a pipeline, or a set of instructions, to perform on a dataframe, and only execute them all at once at the end of your transformation. In other words, its lazy. Pandas is eager, which every part of the transformation happens sequentially and in isolation. Polars also has an eager API, but you likely want to use the lazy API in a production script.

Because its lazy, Polars performs query optimization, like a database does with a SQL query. At the end of the day, if you're using polars for data engineering or in a pipeline, it'll likely work much faster and more memory efficient. Polars also executes operations in parallel, as well.

[-] rutrum@lm.paradisus.day 10 points 1 day ago

How do you use Godot for data science?

55

No surprise I use python, but I've recently started experimenting with polars instead of pandas. I've enjoyed it so far, but Im not sure if the benefits for my team's work will be enough to outweigh the cost of moving from our existing pandas/numpy code over to polars.

I've also started playing with grafana, as a quick dashboarding utility to make some basic visualizations on some live production databases.

60
submitted 4 weeks ago by rutrum@lm.paradisus.day to c/linux@lemmy.ml

I feel like Im dancing around perhaps the most fundamental piece of my operating system everytime I run and install software. Starting services with systemctl and checking logs with journalctl is the extent of my knowledge.

Do you know of good resources or tutorials for learning how systemd works and how to use it to run software on my desktop and servers? Thanks.

134

Typically when I'm working with photos, I'm doing graphic design type work. I've been using GIMP for this. GIMP is meant for raster graphics editing.

You could also use Inkscape for vector graphics, or Krita for more digital painting type work. But I know all these tools are very powerful and overlap on some use cases.

Do you use any AI-type tools? I use a image upscaler called Upscayl. It works really well and works entirely locally.

Do you know of any tools that can remove backgrounds? This would help with help with the type of graphic design I do.

What other tools do you like to use as it pertains to images?

51
15
submitted 3 months ago by rutrum@lm.paradisus.day to c/selfhost@lemmy.ml

I just setup my first automated and encrypted backup with borg. It's got me thinking about other chaotic events, and how to respond accordingly. I figured now is a good time to document my infrastructure: hardware, network, a files. This way if something bad happens, like my house burns down, I or a family member has instructions for how to quickly recover data and services. Examples:

  1. If my website goes down, with my nextcloud on it, what steps do I need to take to recover the data and restore service?
  2. If my harddrive fails, how do I access lost data and reimplement redundancy after a replacement is stood up?
  3. If someone important to me needs to access encrypted files, how can that access that data and get access to the passwords/encryption keys?
  4. If my phone bricks, how to recover 2fa codes?

So I'd like to have a physical printing copy that tries to cover these emergency scenarios. Of course, I'll have digital copy around as well.

I'm focusing more on digital assets, like encryption keys, personal files and media, cloud service access, accessing inaccessible machines, how to restart/recover from self hosted service if its down, etc. I understand how much wider this document can be to include physical assets, so to start I want to start with digital infrastructure.

So my big questions: what scenarios should be documented in this disaster recovery document? What should I prepare for? The nice correlary of this is that documenting a recovery plan will force me to actually stand up the backups/redundancy needed to recover.

45
submitted 3 months ago by rutrum@lm.paradisus.day to c/linux@lemmy.ml

I just got a drawing tablet, and have been wanting some software that would allow me to work out math problems, draw architecture diagrams, etc. I've seen some tools like Excalidraw, which look handy for the sharing capabilities. I also have just used plain krita, which has great feedback for the pen sensitivity, but obviously is overkill for whiteboarding.

Are there any tools you use or recommend for handwriting or picture drawing? Pen or mouse?

[-] rutrum@lm.paradisus.day 34 points 3 months ago

Big fan of the reader mode changes. I'll probabky start using it more often, not just on sites with horrendous popups.

17
submitted 4 months ago by rutrum@lm.paradisus.day to c/coffee@lemmy.world

I love coffee, but have a surplus of tea bags that I want to experiment with. Does anyone have suggestions for how to get started with tea? Or a simple recipe to use as a baseline? I'm only working with tea bags at this time, which appear to be 2g. I would also love to know how much agitation you are supposed to do with the tea bag itself.

44
35

I'm in desparate need of setting up borgmatic for borg backup. I would like to encrypt my backups. (I suppose, an unencrypted backup is better than none in my case, so I should get it done today regardless.)

How do I save those keys? Is there a directory structure I follow? Do you backup the keys as well? Are there keys that I need to write down by hand? Should I use a cloud service like bitwarden secrets manager? Could I host something?

Im ignorant on this matter. The most I've done is add ssh keys to git forges and use ssh-copyid. But I've always been able to access what I need to without keeping those (I login to the web interface.) Can you share with me best practices or what you do to manage non-password secrets?

5

If given the option, which route do you go? I have services running in both, and I'll often just do whats easier. I dont really notice a different in performance the configuration for containers is simple enough I don't mind it.

I also wish there was a nix function that parsed a docker compose and used it for the oci-container config. Then I could use my existing compose files or the ones I find in docs online.

26
submitted 5 months ago* (last edited 5 months ago) by rutrum@lm.paradisus.day to c/nix@programming.dev

This idea is inspired by nixos-mailserver. It was so easy to spin up the mailserver after changing some DNS records and putting in some settings. I thought it might be a good idea to do the same for services that need public, decentralized infrastructure to support. Some ideas include

  • Tor relay, or exit node
  • Encrypted messaging nodes. It looks like SimpleX chat relies on SMP servers to relay communication
  • Crypto miners (I know, I know, but you understand how it fits the “public contribution” usecase)
  • Search engines like searxng (I currently use a public instance)
  • Libredirect services, like proxy clients for social media

Maybe federated services, but those require more than just the software running on the public internet. Those require moderation and long term maintenance. Ideally, the services in this config would be ephemeral.

Does this sound like a good idea? Would you spin one of these up on a $10 VPS? I understand that this is the NixOS community, not necessarily the privacy community, but I figured thered be overlap.

What other services do you think would be applicable?

8
submitted 5 months ago* (last edited 5 months ago) by rutrum@lm.paradisus.day to c/nixos@infosec.pub

This idea is inspired by nixos-mailserver. It was so easy to spin up the mailserver after changing some DNS records and putting in some settings. I thought it might be a good idea to do the same for services that need public, decentralized infrastructure to support. Some ideas include

  • Tor relay, or exit node
  • Encrypted messaging nodes. It looks like SimpleX chat relies on SMP servers to relay communication
  • Crypto miners (I know, I know, but you understand how it fits the "public contribution" usecase)
  • Search engines like searxng (I currently use a public instance)
  • Libredirect services, like proxy clients for social media

Maybe federated services, but those require more than just the software running on the public internet. Those require moderation and long term maintenance. Ideally, the services in this config would be ephemeral.

Does this sound like a good idea? Would you spin one of these up on a $10 VPS? I understand that this is the NixOS community, not necessarily the privacy community, but I figured thered be overlap.

What other services do you think would be applicable?

[-] rutrum@lm.paradisus.day 58 points 7 months ago

Sometimes the app just shows a barcode that they scan. I always screenshotted the barcode and deleted the app. Better yet, save the barcode in catima https://catima.app/

[-] rutrum@lm.paradisus.day 33 points 9 months ago

To be fair, you're taking on a lot of new things at once. You can spin up docker containers on windows too, all while using a UI. I think it's great your exposing yourself to self hosting, linux, command line interface, and containerization all at once, but don't beat yourself up for it taking longer than expected. A lot of it takes time. I encourage you to keep trying and playing. Good luck!

[-] rutrum@lm.paradisus.day 34 points 11 months ago

Theres so many. Check out the awesome list: https://github.com/awesome-selfhosted/awesome-selfhosted

I think your stategy should be one service at a time. Do everything in docker, and start by tackling a simpler service. For example, you should try paperless-ngx. Absolute game changer. I didnt realize how much managing ny own directory structure sucked until I used this. Then, grow your service list more and more!

[-] rutrum@lm.paradisus.day 77 points 1 year ago

You'd have to explain how gimp doesnt suit your needs, because in the open source world its best in class for photo editing.

[-] rutrum@lm.paradisus.day 42 points 1 year ago

Examining my disk partitions with df is ruined now. Every snap gets its own virtual disk.

view more: next ›

rutrum

joined 2 years ago