56
46

Terminal Trove showcases the best of the terminal, Discover a collection of CLI, TUI, and more developer tools at Terminal Trove.

6
submitted 9 months ago by daredevil@kbin.social to c/vgmusic@lemmy.world

イニシエノウタ/デボル · SQUARE ENIX MUSIC · 岡部 啓一 · MONACA

NieR Gestalt & NieR Replicant Original Soundtrack

Released on: 2010-04-21

120

On Monday, Mistral AI announced a new AI language model called Mixtral 8x7B, a "mixture of experts" (MoE) model with open weights that reportedly truly matches OpenAI's GPT-3.5 in performance—an achievement that has been claimed by others in the past but is being taken seriously by AI heavyweights such as OpenAI's Andrej Karpathy and Jim Fan. That means we're closer to having a ChatGPT-3.5-level AI assistant that can run freely and locally on our devices, given the right implementation.

Mistral, based in Paris and founded by Arthur Mensch, Guillaume Lample, and Timothée Lacroix, has seen a rapid rise in the AI space recently. It has been quickly raising venture capital to become a sort of French anti-OpenAI, championing smaller models with eye-catching performance. Most notably, Mistral's models run locally with open weights that can be downloaded and used with fewer restrictions than closed AI models from OpenAI, Anthropic, or Google. (In this context "weights" are the computer files that represent a trained neural network.)

Mixtral 8x7B can process a 32K token context window and works in French, German, Spanish, Italian, and English. It works much like ChatGPT in that it can assist with compositional tasks, analyze data, troubleshoot software, and write programs. Mistral claims that it outperforms Meta's much larger LLaMA 2 70B (70 billion parameter) large language model and that it matches or exceeds OpenAI's GPT-3.5 on certain benchmarks, as seen in the chart below.
A chart of Mixtral 8x7B performance vs. LLaMA 2 70B and GPT-3.5, provided by Mistral.

The speed at which open-weights AI models have caught up with OpenAI's top offering a year ago has taken many by surprise. Pietro Schirano, the founder of EverArt, wrote on X, "Just incredible. I am running Mistral 8x7B instruct at 27 tokens per second, completely locally thanks to @LMStudioAI. A model that scores better than GPT-3.5, locally. Imagine where we will be 1 year from now."

LexicaArt founder Sharif Shameem tweeted, "The Mixtral MoE model genuinely feels like an inflection point — a true GPT-3.5 level model that can run at 30 tokens/sec on an M1. Imagine all the products now possible when inference is 100% free and your data stays on your device." To which Andrej Karpathy replied, "Agree. It feels like the capability / reasoning power has made major strides, lagging behind is more the UI/UX of the whole thing, maybe some tool use finetuning, maybe some RAG databases, etc."

Mixture of experts

So what does mixture of experts mean? As this excellent Hugging Face guide explains, it refers to a machine-learning model architecture where a gate network routes input data to different specialized neural network components, known as "experts," for processing. The advantage of this is that it enables more efficient and scalable model training and inference, as only a subset of experts are activated for each input, reducing the computational load compared to monolithic models with equivalent parameter counts.

In layperson's terms, a MoE is like having a team of specialized workers (the "experts") in a factory, where a smart system (the "gate network") decides which worker is best suited to handle each specific task. This setup makes the whole process more efficient and faster, as each task is done by an expert in that area, and not every worker needs to be involved in every task, unlike in a traditional factory where every worker might have to do a bit of everything.

OpenAI has been rumored to use a MoE system with GPT-4, accounting for some of its performance. In the case of Mixtral 8x7B, the name implies that the model is a mixture of eight 7 billion-parameter neural networks, but as Karpathy pointed out in a tweet, the name is slightly misleading because, "it is not all 7B params that are being 8x'd, only the FeedForward blocks in the Transformer are 8x'd, everything else stays the same. Hence also why total number of params is not 56B but only 46.7B."

Mixtral is not the first "open" mixture of experts model, but it is notable for its relatively small size in parameter count and performance. It's out now, available on Hugging Face and BitTorrent under the Apache 2.0 license. People have been running it locally using an app called LM Studio. Also, Mistral began offering beta access to an API for three levels of Mistral models on Monday.

3
submitted 9 months ago* (last edited 9 months ago) by daredevil@kbin.social to c/tech@kbin.social

On Monday, Mistral AI announced a new AI language model called Mixtral 8x7B, a "mixture of experts" (MoE) model with open weights that reportedly truly matches OpenAI's GPT-3.5 in performance—an achievement that has been claimed by others in the past but is being taken seriously by AI heavyweights such as OpenAI's Andrej Karpathy and Jim Fan. That means we're closer to having a ChatGPT-3.5-level AI assistant that can run freely and locally on our devices, given the right implementation.

Mistral, based in Paris and founded by Arthur Mensch, Guillaume Lample, and Timothée Lacroix, has seen a rapid rise in the AI space recently. It has been quickly raising venture capital to become a sort of French anti-OpenAI, championing smaller models with eye-catching performance. Most notably, Mistral's models run locally with open weights that can be downloaded and used with fewer restrictions than closed AI models from OpenAI, Anthropic, or Google. (In this context "weights" are the computer files that represent a trained neural network.)

Mixtral 8x7B can process a 32K token context window and works in French, German, Spanish, Italian, and English. It works much like ChatGPT in that it can assist with compositional tasks, analyze data, troubleshoot software, and write programs. Mistral claims that it outperforms Meta's much larger LLaMA 2 70B (70 billion parameter) large language model and that it matches or exceeds OpenAI's GPT-3.5 on certain benchmarks, as seen in the chart below.

The speed at which open-weights AI models have caught up with OpenAI's top offering a year ago has taken many by surprise. Pietro Schirano, the founder of EverArt, wrote on X, "Just incredible. I am running Mistral 8x7B instruct at 27 tokens per second, completely locally thanks to @LMStudioAI. A model that scores better than GPT-3.5, locally. Imagine where we will be 1 year from now."

LexicaArt founder Sharif Shameem tweeted, "The Mixtral MoE model genuinely feels like an inflection point — a true GPT-3.5 level model that can run at 30 tokens/sec on an M1. Imagine all the products now possible when inference is 100% free and your data stays on your device." To which Andrej Karpathy replied, "Agree. It feels like the capability / reasoning power has made major strides, lagging behind is more the UI/UX of the whole thing, maybe some tool use finetuning, maybe some RAG databases, etc."

Mixture of experts

So what does mixture of experts mean? As this excellent Hugging Face guide explains, it refers to a machine-learning model architecture where a gate network routes input data to different specialized neural network components, known as "experts," for processing. The advantage of this is that it enables more efficient and scalable model training and inference, as only a subset of experts are activated for each input, reducing the computational load compared to monolithic models with equivalent parameter counts.

In layperson's terms, a MoE is like having a team of specialized workers (the "experts") in a factory, where a smart system (the "gate network") decides which worker is best suited to handle each specific task. This setup makes the whole process more efficient and faster, as each task is done by an expert in that area, and not every worker needs to be involved in every task, unlike in a traditional factory where every worker might have to do a bit of everything.

OpenAI has been rumored to use a MoE system with GPT-4, accounting for some of its performance. In the case of Mixtral 8x7B, the name implies that the model is a mixture of eight 7 billion-parameter neural networks, but as Karpathy pointed out in a tweet, the name is slightly misleading because, "it is not all 7B params that are being 8x'd, only the FeedForward blocks in the Transformer are 8x'd, everything else stays the same. Hence also why total number of params is not 56B but only 46.7B."

Mixtral is not the first "open" mixture of experts model, but it is notable for its relatively small size in parameter count and performance. It's out now, available on Hugging Face and BitTorrent under the Apache 2.0 license. People have been running it locally using an app called LM Studio. Also, Mistral began offering beta access to an API for three levels of Mistral models on Monday.

[-] daredevil@kbin.social 15 points 9 months ago

Happy to hear you're getting some time in for yourself. The constant updates have been a joy, but downtime is important too.

[-] daredevil@kbin.social 17 points 10 months ago

I'd imagine this will also be very problematic for non-celebrities from all sorts of backgrounds as well. The harassment potential is very concerning.

[-] daredevil@kbin.social 14 points 10 months ago* (last edited 10 months ago)

kbin has this -- the feature is called collections. https://kbin.social/magazines/collections

you can make public ones that others can follow, or private ones to make curated feeds for yourself.

[-] daredevil@kbin.social 32 points 10 months ago

Pandora's Box is already opened, unfortunately. The Streisand Effect is only going to make this worse.

[-] daredevil@kbin.social 51 points 10 months ago

What scares me more than the fact that this guy exists are how many people chose to and continue to support him

[-] daredevil@kbin.social 12 points 10 months ago* (last edited 10 months ago)

Do you have any favorites that have significantly improved your instance experience?

  • @kbinDevlog -- transparency and continuous updates were probably the biggest thing for me. Thank you @Ernest.
  • Transfer of abandoned magazines
  • Request for magazine moderator
  • #Collections
  • Improved account and magazine deletion
  • Anti-spam protections
  • #Crosspost function

Regarding Collections: I'm wondering if it might be useful to allow transfer ownership of collections as well? Probably low priority, but this could be useful if:

  • The owner becomes inactive
  1. A #kbin magazine or #lemmy community becomes problematic + has inactive moderators
  2. Other magazines/communities become more/less relevant over time, thus the collection needing adjustments

Re 1:

  • Could be mitigated on kbin magazines due to magazine ownership transfers, however, may pose an issue on federated magazines from lemmy
  • Also could be mitigated by the creation of a new collection, hence probably low priority

Re 2:

  • Could be low priority in regard to certain magazines/communities becoming inactive over time, however, the chance to miss out on discussions and threads should possibly be considered
  • However, this is also mitigated by also creating new collections -- I just think people generally like to avoid migrating when followings settle in.

Transferring collection ownership could also be excessive/unnecessary? Thoughts? Does the status of an official collection change automatically based on the user count?

Also regarding collections: After following a collection -- going to a collection's page and attempting to unfollow is described as "delete". This may give off the wrong impression to unfamiliar users. Perhaps "unfollow", or "unfavorite" may be better suited? I also wonder if it may be more appropriate to have favorited collections appear at the top of the collections view so users can quickly find the collections they follow in order to avoid being buried by the popular ones. This could also potentially be solved by giving subscribed collections its own tab in the navbar next to magazines? Is that perhaps too many tabs in the navbar?

Anyways, thank you so much Ernest. I know some people were unhappy about /kbin's development progress for awhile, but I'm really glad I stuck with it. Despite the challenges that came your way, watching you get back into the swing of things and /kbin's growth has been a joy to be a part of. The consistent communication and having our feedback be heard is also a big plus. As always, looking forward to what else is on the way.

Edit: Also, really glad to hear you're prioritizing sustainability, balance, and a strong foundation for the future. I hope you're able to get the support you're looking for, so you can give this project the time and effort you think it deserves. It's clear you care a lot about this project.

16
submitted 10 months ago* (last edited 10 months ago) by daredevil@kbin.social to c/vgmusic@lemmy.world
4
Sleepy teefies (media.kbin.social)
submitted 10 months ago by daredevil@kbin.social to c/Cats@kbin.social
27
Blep (media.kbin.social)
submitted 10 months ago by daredevil@kbin.social to c/Cats@kbin.social
1
submitted 10 months ago by daredevil@kbin.social to c/gaming@kbin.social

Resident Evil 4 Remake has been crowned PlayStation Game of the Year at The Golden Joysticks 2023 powered by Intel.

Capcom's third Resident Evil remake was released in March of this year and took players back to rural Spain to confront the mysterious, and deadly, Los Illuminados cult - 18 years after we originally did on the PlayStation 2. Fans clearly loved revisiting the classic survival horror game as it managed to beat out other games in the category including Final Fantasy 16, Street Fighter 6, and Star Wars Jedi: Survivor.

The other Golden Joystick Awards 2023 nominees in this category can be found below:

  • Final Fantasy 16
  • Resident Evil 4 Remake (winner)
  • Street Fighter 6
  • Humanity
  • Armored Core 6: Fires of Rubicon
  • Star Wars Jedi: Survivor
5
submitted 10 months ago by daredevil@kbin.social to c/AskKbin@kbin.social

First one that comes to my mind is having to travel with an NPC and our walk/run speeds don't match.

15
submitted 10 months ago by daredevil@kbin.social to c/kbinMeta@kbin.social

@Ernest has pushed an update which allows users to request ownership/moderation of abandoned magazines. Ghost/abandoned magazines were fairly prevalent after the initial wave of hype due to users either squatting magazine names or becoming inactive for other reasons. Now is your chance to get involved, if you were waiting to do so.

To request ownership/moderator privileges, scroll down to where it says "MODERATORS" in the sidebar. There will be an icon of a hand pointing upwards that you can click on, then make the request. Cheers, and thank you for your hard work Ernest, as well as future mods.

11

Title: Let the Battles Begin!
Name: Final Fantasy VII
Year Released: 1997
Composer: Nobuo Uematsu
Developer: Square Enix
Platform: PlayStation

[-] daredevil@kbin.social 17 points 1 year ago

This does not spark joy.

[-] daredevil@kbin.social 15 points 1 year ago

I hope you're not refusing to clean the cat's litter box because neglect is a form of abuse. I feel sorry for the cat.

[-] daredevil@kbin.social 14 points 1 year ago* (last edited 1 year ago)

A lot of social engagement through social media is driven by impressions such as up votes, favorites, likes, etc. Unfortunately, an easy way to promote engagement and such lies in rage bait. This is likely due to the visceral emotional response generated by rage baiting. I would also extend this issue to how ubiquitous instant gratification is to the internet and social media. People tend to acquire clout through reacting to something quickly, which isn't always well-thought out. Add in the notion of mob mentality, and you have a recipe for the rapid exponential propagation of negative words, thoughts, and emotions. People also tend to not have productive ways of channeling their frustrations and issues, so they often see other entities on the Internet as just a name, sometimes less than that.

There's also a heavy amount of tribalism across a variety of domains which allows one to take refuge from this rage baiting by finding other like-minded individuals to identify with. In some cases, the stress of everyday life or what have you removes a sense of agency or power in one's life and sometimes people cope with this by developing a sense of superiority through whichever group or ideal that they identify with. This cycle repeats itself until there is a constant battle between any given groups where people attempt to elevate their self-worth by putting those that they dont agree with down, while emphasizing the superiority of their own ideal, IMO. I could be totally wrong ofc. I'm hardly perfect.

It's been a pretty exhausting experience. I'm tired of it as well; my fondness for engaging with people has diminished greatly.

[-] daredevil@kbin.social 26 points 1 year ago

Even though this is a nice development, I'm pretty disappointed in the resources directed at Linux support. I'm considering dropping Proton soon.

[-] daredevil@kbin.social 23 points 1 year ago

Are you shore?

[-] daredevil@kbin.social 17 points 1 year ago

I'm okay just finding a niche community that aligns more with my perspective. Even if it didn't dent Reddit's numbers like many has hoped, more people have joined here much more quickly than I would've thought. While it also may be a hindrance to some, I think it's quite fascinating being a part of a rapidly expanding community going through growing pains. I have no idea how this will pan out, but I'm glad I'm here for the ride.

view more: next ›

daredevil

joined 1 year ago