[-] fatboy93@lemm.ee 22 points 1 month ago

Why do people sleep on KDE connect? It does a lot of things really well and is OS agnostic.

[-] fatboy93@lemm.ee 14 points 1 month ago

Dang, you're Moneyball'ing your kid?

Sounds awesome!

[-] fatboy93@lemm.ee 22 points 6 months ago

This has the same energy as my spouse yelling at me because jellyfin went down

[-] fatboy93@lemm.ee 15 points 7 months ago

If you're on windows, Everything by Void tools is the best at indexing and searching.

[-] fatboy93@lemm.ee 12 points 7 months ago

For $10M, I might just retire, buy a house and raise my kid.

Sure beats having to struggle at work.

[-] fatboy93@lemm.ee 36 points 8 months ago

Gods, I relate to this.

My wife watches Big Boss which is like Bog Brother everyday.

My toddler has some song or the other playing all day long.

I can't put my headphones because either of the two keep talking all the time.

When my inlaws visit:

My dad in law is a media person, so he's on the phone the whole darn day.

My mom in law sing-songs her words.

My sis in law doesn't remember any song past the first two lines, so she sings the two lines whole fucking day.

I live in a hellscape of my own device.

[-] fatboy93@lemm.ee 28 points 9 months ago

I don't even care about other features. The tiling home screen of the OS was really nice to use and when used properly by the apps could result in a "live" OS unlike the iconographic interfaces of iOS and Android. The homescreen was also old-age friendly and really a pleasure to use.

The OS ram like really smooth on 512mb RAM, unlike their counterpart android phones which were struggling back then with 2-3GB RAM.

The lumias themselves had a ton of useful features like tap to wake etc, which didn't consume much battery and in general the Nokia cameras were top notch for the time.

Basically, the OS got killed because of a chicken and egg problem with the apps, and the OS being from Microsoft, got a death knell because of the reputation. Also for some fucking reason, Microsoft decided that the already low userbase WP7s were to be depreciated rather than provide an upgrade path fo WP8 and WP10.

[-] fatboy93@lemm.ee 12 points 10 months ago

Unable to read the paywalled article, but ROCm is like Nvidia's CUDA for their GOUs. These used to be supported on and off for consumer GPUs, but get aupported mostly only for datacenter ones. Linux used to be supported, but they are bringing ROCm to windows as well.

Probably AMD is planning to improve support for consumer GPUs, since their GPUs are competitive to Nvidia's at a lower price point, given that local LLMs, Image generation and other AIs are in a kind of a booming trend.

Currently, Nvidia has CUDA which is more or less the industry standard, so AMD with their ROCm and Intel with their OpenVino etc are trying to chip away at the monopoly.

[-] fatboy93@lemm.ee 14 points 11 months ago

British museum is just the "accidental" offsite backup plan for a lot of history and cultural artefacts.

[-] fatboy93@lemm.ee 33 points 1 year ago

Not opensource, but i love juicessh. I keep coming back to it all the time.

[-] fatboy93@lemm.ee 13 points 1 year ago

While I really get what you're saying, the unfortunate situation is that they are always a secondary fiddle to money

16
submitted 1 year ago* (last edited 1 year ago) by fatboy93@lemm.ee to c/localllama@sh.itjust.works

Hi!

I have an ASUS AMD Advantage Edition laptop (https://rog.asus.com/laptops/rog-strix/2021-rog-strix-g15-advantage-edition-series/) that runs windows. I haven't gotten time to install linux and set it up the way I like yet, still after more than a year.

I'm just dropping a small write-up for the set-up that I'm using with llama.cpp to run on the discrete GPUs using clbast.

You can use Kobold but it meant for more role-playing stuff and I wasn't really interested in that. Funny thing is Kobold can be set up to use the discrete GPU if needed.

  1. For starters you'd need llama.cpp itself from here: https://github.com/ggerganov/llama.cpp/tags.

    Pick the clblast version, which will help offload some computation over to the GPU. Unzip the download to a directory. I unzipped it to a folder called this: "D:\Apps\llama"

  2. You'd need a llm now and that can be obtained from HuggingFace or where-ever you'd like it from. Just note that it should be in ggml format. If you have a doubt, just note that the models from HuggingFace would have "ggml" written somewhere in the filename. The ones I downloaded were "nous-hermes-llama2-13b.ggmlv3.q4_1.bin" and "Wizard-Vicuna-7B-Uncensored.ggmlv3.q4_0.bin"

  3. Move the models to the llama directory you made above. That makes life much easier.

  4. You don't really need to navigate to the directory using Explorer. Just open Powershell where-ever and you can also do cd D:\Apps\llama\

  5. Here comes the fiddly part. You need to get the device ids for the GPU. An easy way to check this is to use "GPU caps viewer", go to the tab titled OpenCl and check the dropdown next to "No. of CL devices".

    The discrete GPU is normally loaded as the second or after the integrated GPU. In my case the integrated GPU was gfx90c and discrete was gfx1031c.

  6. In the powershell window, you need to set the relevant variables that tell llama.cpp what opencl platform and devices to use. If you're using AMD driver package, opencl is already installed, so you needn't uninstall or reinstall drivers and stuff.

    $env:GGML_OPENCL_PLATFORM = "AMD"

    $env:GGML_OPENCL_DEVICE = "1"

  7. Check if the variables are exported properly

    Get-ChildItem env:GGML_OPENCL_PLATFORM
    Get-ChildItem env:GGML_OPENCL_DEVICE

    This should return the following:

    Name Value


    GGML_OPENCL_PLATFORM AMD

    GGML_OPENCL_DEVICE 1

    If GGML_OPENCL_PLATFORM doesn't show AMD, try exporting this: $env:GGML_OPENCL_PLATFORM = "AMD"

  8. Once these are set properly, run llama.cpp using the following:

    D:\Apps\llama\main.exe -m D:\Apps\llama\Wizard-Vicuna-7B-Uncensored.ggmlv3.q4_0.bin -ngl 33 -i --threads 8 --interactive-first -r "### Human:"

    OR

    replace Wizard with nous-hermes-llama2-13b.ggmlv3.q4_1.bin or whatever llm you'd like. I like to play with 7B, 13B with 4_0 or 5_0 quantized llms. You might need to trawl through the fora here to find parameters for temperature, etc that work for you.

  9. Checking if these work, I've posted the content at pastebin since formatting these was a paaaain: https://pastebin.com/peSFyF6H

    salient features @ gfx1031c (6800M discrete graphics):
    llama_print_timings: load time = 60188.90 ms
    llama_print_timings: sample time = 3.58 ms / 103 runs ( 0.03 ms per token, 28770.95 tokens per second)
    llama_print_timings: prompt eval time = 7133.18 ms / 43 tokens ( 165.89 ms per token, 6.03 tokens per second)
    llama_print_timings: eval time = 13003.63 ms / 102 runs ( 127.49 ms per token, 7.84 tokens per second)
    llama_print_timings: total time = 622870.10 ms

    salient features @ gfx90c (cezanne architecture integrated graphics):
    llama_print_timings: load time = 26205.90 ms
    llama_print_timings: sample time = 6.34 ms / 103 runs ( 0.06 ms per token, 16235.81 tokens per second)
    llama_print_timings: prompt eval time = 29234.08 ms / 43 tokens ( 679.86 ms per token, 1.47 tokens per second)
    llama_print_timings: eval time = 118847.32 ms / 102 runs ( 1165.17 ms per token, 0.86 tokens per second)
    llama_print_timings: total time = 159929.10 ms

Edit: added pastebin since I actually forgot to link it. https://pastebin.com/peSFyF6H

[-] fatboy93@lemm.ee 14 points 1 year ago

Not necessarily, generally the defaults for most of the tools tend to be sane, but when you have a swiss army knife with dozens of attachments, you'd still need a manual to figure what is what.

Note that many tools use ffmpeg under the hood so users are generally never exposed to the various options. But sometime if they need to, cheatsheets like these are really useful.

1
submitted 1 year ago by fatboy93@lemm.ee to c/meta@lemm.ee

Hi!

I subscribed to a few magazines from kbin.social. Is there something I need to check/do so that the subscriptions gets synced across the two instances?

If not, does this mean we are not federated with them yet?

Regards, fatboy93

view more: next ›

fatboy93

joined 1 year ago