Sims

joined 2 years ago
[–] Sims@lemmy.ml 2 points 9 hours ago

There are only a few areas where China doesn't lead over the US/West in technology - well, and in everything else actually. Amazing how quickly they beat the free market religion - even though they were obstructed all the way..

Go go China !!

[–] Sims@lemmy.ml 16 points 2 days ago (3 children)

Didn't know what uBlue was, so here: https://universal-blue.org/

"The Universal Blue project builds a diverse set of continuously delivered operating system images using bootc. That's nerdspeak for the ultimate Linux client: the reliability of a Chromebook, but with the flexibility and power of a traditional Linux desktop.

These images represent what's possible when a community focuses on sharing best practices via automation and collaboration. One common language between dev and ops, and it's finally come to the desktop.

We also provide tools for users to build their own image using our templates and processes, which can be used to ship custom configurations to all of your machines, or finally make the Linux distribution you've long wished for, but never had the tools to create.

At long last, we've ascended."

[–] Sims@lemmy.ml 1 points 2 days ago* (last edited 2 days ago)

Been enjoying Linux for ~25 years, but have never been happy with how it handled low memory situations. Swapping have always killed the system, tho it have improved a little. It's been a while since I've messed with it. I've buckled up and are using more ram now, but afair, you can play with:

(0. reduce running software, and optimize them for less memory, yada yada)

  1. use a better OOM (out of memory) manager that activates sooner and more gracefully. Search in your OS' repository for it.
  2. use zram as a more intelligent buffer and to remove same (zero) pages. It can lightly compress lesser used memory pages and use a partition backend for storing uncompressible pages. You spend a little cpu, to minimize swap, and when needed, only swap out what can't be compressed.
  3. play with all the sysctl vm settings like swappiness and such, but be aware that there's SO much misinformation out there, so seek the official kernel docs. For instance, you can adapt the system to swap more often, but in much smaller chunks, so you avoid spending 5 minutes to hours regaining control - the system may get 'sluggish', but you have control.
  4. use cgroups to divide you resources, so firefox/chrome (or compilers/memory-hogs) can only use X amount before their memory have to swap out (if they don't adapt to lower mem conditions automatically). That leaves a system for you that can still react to your input (while ff/chrome would freeze). Not perfect, tho.
  5. when gaming, activate a low-system mode, where unnecessary services etc are disabled. I think there's a library/command that helps with that (and raise priority etc), but forgot its name.

EDIT: 6. when NOT gaming, add some of your vram as swap space. Its much faster than your ssd. Search github or your repository for 'vram cache' or something like that. It works via opencl, so everyone with dedicated vram can use it as super fast cache. Perhaps others can remember the name/link ?

Something like that anyway, others will know more about each point.

Also, perhaps ask an AI to create a small interface for you to fiddle with vm settings and cgroups in an automated/permanent way ? just a quick thought. Good luck.

[–] Sims@lemmy.ml 5 points 2 days ago

For those that follow internation geo-politics and how peoples opinions are formed, the only amazing thing is that around 50% of the Western population STILL thinks they are part of a 'Democracy'.

The Western propaganda apparatus really is the wildest show on earth..

[–] Sims@lemmy.ml 3 points 1 week ago (1 children)

Agree. I also shift between them. As the bare minimum, I use a thinking model to 'open up' the conversation, and then often continue with a normal model, but it certainly depends on the topic.

Long ago we got 'routellm' I think, that routed a request depended on its content, but the concept never got traction for some reason. Now it seems that closedai and other big names are putting some attention to it. Great to see DeepHermes and other open players be in front of the pack.

I don't think it will take long before we have the agentic framework do the activation of different 'modes' of thinking dependent on content/context, goals etc. It would be great if a model can be triggered into several modes in a standard way.

[–] Sims@lemmy.ml 2 points 1 week ago (2 children)

You can argue that a 4090 is more of a 'flagship' model on the consumer market, but it could be just a typing error, and then you miss the point and the knowledge you could have learned:

"Their system, FlightVGM, recorded a 30 per cent performance boost and had an energy efficiency that was 4½ times greater than Nvidia’s flagship RTX 3090 GPU – all while running on the widely available V80 FPGA chip from Advanced Micro Devices (AMD), another leading US semiconductor firm."

So they have found a way to use a 'off-the-shelf' FPGA and are using it for video inference, and to me it looks like it could match a 4090(?), but who cares. With this upgrade, these standard Fpga's are cheaper(running 24/7)/better than any consumer Nvidia GPU up to at least 3090/4090.

And here from the paper:

"[problem] ..sparse VGMs [video generating models] cannot fully exploit the effective throughput (i.e., TOPS) of GPUs. FPGAs are good candidates for accelerating sparse deep learning models. However, existing FPGA accelerators still face low throughput ( < 2TOPS) on VGMs due to the significant gap in peak computing performance (PCP) with GPUs ( > 21× ).

[solution] ..we propose FlightVGM, the first FPGA accelerator for efficient VGM inference with activation sparsification and hybrid precision. [..] Implemented on the AMD V80 FPGA, FlightVGM surpasses NVIDIA 3090 GPU by 1.30× in performance and 4.49× in energy efficiency on various sparse VGM workloads."

You'll have to look up what that means yourself, but expect a throng of bitcrap miner cards to be converted to VLM accelerators, and maybe give new life for older/smaller/cheaper fpga's ?

[–] Sims@lemmy.ml 3 points 1 week ago (1 children)

Pretty cool with China's focus on efficiency in the AI stack. DeepSeek was the first eye-opener for how to re-think efficiency, but it appears to happen on all levels of the stack.

Fyi: article is paywalled, so block javascript on page with ublock..

[–] Sims@lemmy.ml 2 points 8 months ago

There's a cheap zbtlink openwrt wifi6 3000Mbps 'z8101ax-d' on AliE for around 50 $. (https://www.aliexpress.com/w/wholesale-zbtlink-openwrt.html?spm=a2g0o.productlist.search.0)

I don't know how long, and haven't tried the product, but maybe some here have tried it ?

[–] Sims@lemmy.ml 1 points 8 months ago

Time to fill the drive with anti-establishment political literature + de-google material + google-service hacking news, fake stories about their leaders etc. Anything that goes against what Google propaganda models wants us to believe or what benefits them most..

[–] Sims@lemmy.ml 3 points 8 months ago (1 children)

https://youtu.be/4bBxMs5mhvs 'Consciousness was invented in the 1600s'

Also, the easiest way to generate 'Consciousness' in any system is to add a feedback mechanism. A simple echo in a system or input stream, is a potential emerging 'consciousness'. It becomes aware of something 'it' did before - and then it discover the illusion of 'I', free will and personal responsibility (vs long-chained cause&effect from society/environment).

[–] Sims@lemmy.ml 2 points 9 months ago

Sounds like standard Capitalist practices if you ask me.

[–] Sims@lemmy.ml 2 points 9 months ago

Cute! My first computer was a C= 64 'beast'. I wonder how many of those this little fella can emulate :)

 

I thought of this rare little sub when watching this. It's difficult to evaluate how much Yank (and western) culture have contaminated our expectation and knowledge of other cultures. This video explores some of that..

And, ..going slightly OT on my own post, when trying to search for African music, the result contains mostly African 'beyonce-like' music. I have to add 'root' or similar to find something that actually sounds 'original', local and non-yank. A bit sad.

 

I am planning my first ai-lab setup, and was wondering how many tokens different AI-workflows/agent network eat up on an average day. For instance talking to an AI all day, have devlin running 24/7 or whatever local agent workflow is running.

Oc model inference speed and type of workflow influences most of these networks, so perhaps it's easier to define number of token pr project/result ?

So I were curious about what typical AI-workflow lemmies here run, and how many tokens that roughly implies on average, or on a project level scale ? Atmo I don't even dare to guess.

Thanks..

 

I could not tell that they were generated..

 

Hi all.

For a long long time I've been very happy with Signal, but have lately become rather annoyed that:

  1. it too often bugs me about about an update and forces me to do an update before I can write to my single/only recipient, and
  2. it too often bugs me about my pin code, even tho I never asked for such an annoying level of security.

These security measures are completely overkill for my/normal use, unnecessary, annoying and very aggressive. I'm an adult, and unless there's a super dangerous zeroday attack/vulnerability, I don't need constant forced updates, I don't want to retype a pin code for any reason or interval, and I certainly don't need to be told how I should run my system, when to upgrade or have software on my system that 'randomly' gets locked down for whatever reason.

Does anyone know how I can turn it off (Linux, Android) ? Is there another client fork that don't force me to follow their idea of what security level is necessary ?

Thanks, and apols for negativity..

view more: next ›