14
submitted 2 weeks ago by Akisamb@programming.dev to c/france@jlai.lu
[-] Akisamb@programming.dev 23 points 3 weeks ago* (last edited 3 weeks ago)

https://notesfrompoland.com/2024/06/17/poland-demands-explanation-after-german-police-transport-migrant-family-back-over-border/

“They had Polish refugee status certificates for the adults and Polish ID cards for the children. These people were taken to the police station for further investigation,” said von Vegesack. He added that, after their legal situation was confirmed, the family had to be sent back to Poland.

According to the spokesman, German officials informed the Polish border guard about the situation through the Polish-German Centre for Cooperation of Border, Police and Customs Services in Świecko.

“Since there was no reaction from the Polish side for several hours, the officers decided to take the family to the Polish-German border…in order to release them to Poland,” said von Vegesack.

The article above should be more "polish refugees left alone in Poland by German police" rather than the current one which suggests that Germany is getting rid of its migrants illegally.

[-] Akisamb@programming.dev 69 points 4 weeks ago

Separately, Jong has also alleged that Apple subjected her to a hostile work environment after a senior member of her team, Blaine Weilert, sexually harassed her. After she complained, Apple investigated and Weilert reportedly admitted to touching her "in a sexually suggestive manner without her consent," the complaint said. Apple then disciplined Weilert but ultimately would not allow Jong to escape the hostile work environment, requiring that she work with Weilert on different projects. Apple later promoted Weilert.

As a result of Weilert's promotion, the complaint said that Apple placed Weilert in a desk "sitting adjacent" to Jong's in Apple’s offices. Following a request to move her desk, a manager allegedly "questioned" Jong's "willingness to perform her job and collaborate" with Weilert, advising that she be “professional, respectful, and collaborative,” rather than honoring her request for a non-hostile workplace.

...

2
6

cross-posted from: https://lemmy.one/post/13942290

Abstract: We present Scallop, a language which combines the benefits of deep learning and logical reasoning. Scallop enables users to write a wide range of neurosymbolic applications and train them in a data- and compute-efficient manner. It achieves these goals through three key features: 1) a flexible symbolic representation that is based on the relational data model; 2) a declarative logic programming language that is based on Datalog and supports recursion, aggregation, and negation; and 3) a framework for automatic and efficient differentiable reasoning that is based on the theory of provenance semirings. We evaluate Scallop on a suite of eight neurosymbolic applications from the literature. Our evaluation demonstrates that Scallop is capable of expressing algorithmic reasoning in diverse and challenging AI tasks, provides a succinct interface for machine learning programmers to integrate logical domain knowledge, and yields solutions that are comparable or superior to state-of-the-art models in terms of accuracy. Furthermore, Scallop's solutions outperform these models in aspects such as runtime and data efficiency, interpretability, and generalizability.

[-] Akisamb@programming.dev 15 points 3 months ago

I'm afraid that would not be sufficient.

These instructions are a small part of what makes a model answer like it does. Much more important is the training data. If you want to make a racist model, training it on racist text is sufficient.

Great care is put in the training data of these models by AI companies, to ensure that their biases are socially acceptable. If you train an LLM on the internet without care, a user will easily be able to prompt them into saying racist text.

Gab is forced to use this prompt because they're unable to train a model, but as other comments show it's pretty weak way to force a bias.

The ideal solution for transparency would be public sharing of the training data.

20
[-] Akisamb@programming.dev 26 points 3 months ago

And cow feed is also grown with tons of pesticides and you need much more of it for less tissue at the end.

I have hard time seeing clothing with a bigger environmental than leather.

14

cross-posted from: https://lemmy.ml/post/13088944

5
The Annotated S4 (srush.github.io)
[-] Akisamb@programming.dev 14 points 5 months ago

They gave them a birth control shot without properly informing them of what it was. Still scandalous, but not what you are saying.

[-] Akisamb@programming.dev 12 points 6 months ago

Didier Raoult for a large part. He was the one who published the paper that really started this whole mess. His shoddy research practices and non-respect for patients did plenty of harm.

Good thing that they've forced his retirement.

17
1

abstract :

How do sequence models represent their decision-making process? Prior work suggests that Othello-playing neural network learned nonlinear models of the board state (Li et al., 2023). In this work, we provide evidence of a closely related linear representation of the board. In particular, we show that probing for "my colour" vs. "opponent's colour" may be a simple yet powerful way to interpret the model's internal state. This precise understanding of the internal representations allows us to control the model's behaviour with simple vector arithmetic. Linear representations enable significant interpretability progress, which we demonstrate with further exploration of how the world model is computed.

[-] Akisamb@programming.dev 13 points 7 months ago

Some members have stated as such but have been corrected by the leadership. Hamas, at least publicly, only said that they wanted to forcefully displace the Jews and that they would not hesitate to kill civilians to attain that objective.

Example: https://www.washingtonexaminer.com/news/senior-hamas-official-urges-palestinians-worldwide-to-kill-every-jew-on-the-globe

8
submitted 7 months ago* (last edited 7 months ago) by Akisamb@programming.dev to c/machine_learning@programming.dev

Paper here : https://arxiv.org/pdf/2312.00752.pdf

Abstract :

Foundation models, now powering most of the exciting applications in deep learning, are almost universally based on the Transformer architecture and its core attention module. Many subquadratic-time architectures such as linear attention, gated convolution and recurrent models, and structured state space models (SSMs) have been developed to address Transformers’ computational inefficiency on long sequences, but they have not performed as well as attention on important modalities such as language. We identify that a key weakness of such models is their inability to perform content-based reasoning, and make several improvements. First, simply letting the SSM parameters be functions of the input addresses their weakness with discrete modalities, allowing the model to selectively propagate or forget information along the sequence length dimension depending on the current token. Second, even though this change prevents the use of efficient convolutions, we design a hardware-aware parallel algorithm in recurrent mode. We integrate these selective SSMs into a simplified end-to-end neural network architecture without attention or even MLP blocks (Mamba). Mamba enjoys fast inference (5× higher throughput than Transformers) and linear scaling in sequence length, and its performance improves on real data up to million-length sequences. As a general sequence model backbone, Mamba achieves state-of-the-art performance across several modalities such as language, audio, and genomics. On language modeling, our Mamba-3B model outperforms Transformers of the same size and matches Transformers twice its size, both in pretraining and downstream evaluation.

[-] Akisamb@programming.dev 12 points 7 months ago

These companies pollute to satisfy a demand. If people stopped driving cars, TotalEnergy and Shell would sell less oil.

Where these companies are evil is when they try to influence people and politicians. For example jay walking is a crime because of them.

That said taxing the hell out of these polluting industries is a solution, as it will raise the oil price and force people to consume less.

[-] Akisamb@programming.dev 21 points 7 months ago

Most surprisingly, the inspectors observed barefoot employees working in a sterile area of the facility, where they should have been wearing shoes—plus gowns, gloves, and shoe booties. (The barefoot workers were also not wearing gowns or gloves.) A production manager puzzlingly told FDA inspectors that shoeless work is "standard practice."

They were supposed to cover everything including the feet.

[-] Akisamb@programming.dev 12 points 9 months ago

Even if 99% of it would evaporate that would still be a ridiculous amount of power.

But Bill Gates proved that diversifying a stock of mainly one company while having that company keep all its value is possible. Elon Musk is horrifyingly rich like it or not. His power and the damage he can do is huge.

[-] Akisamb@programming.dev 37 points 11 months ago

A particular flavour of right wing though. One that goes out of it's way to say the most aboherent shit on people that are not like them.

In France we've got laws to regulate this type of speech. Say that AIDS is god's punishment to the homosexuals -> get a fine (or if it's the twentieth time you do it, go to jail).

I don't find it shocking that people are banning content that some countries judge so problematic that they are ready to put people in jail for it.

1

Was looking at EAP6 release notes and was pleasantly surprised to see this there.

I'm quite happy that intellij provides on premise solutions, it gives a small chance of this coming to my job one day. I believe this will be quite useful for repetitive code and certain types of tests.

1

cross-posted from: https://kbin.social/m/machinelearning/t/98088

Abstract:

Work on scaling laws has found that large language models (LMs) show predictable improvements to overall loss with increased scale (model size, training data, and compute). Here, we present evidence for the claim that LMs may show inverse scaling, or worse task performance with increased scale, e.g., due to flaws in the training objective and data. We present empirical evidence of inverse scaling on 11 datasets collected by running a public contest, the Inverse Scaling Prize, with a substantial prize pool. Through analysis of the datasets, along with other examples found in the literature, we identify four potential causes of inverse scaling: (i) preference to repeat memorized sequences over following in-context instructions, (ii) imitation of undesirable patterns in the training data, (iii) tasks containing an easy distractor task which LMs could focus on, rather than the harder real task, and (iv) correct but misleading few-shot demonstrations of the task. We release the winning datasets at https://inversescaling.com/data to allow for further investigation of inverse scaling. Our tasks have helped drive the discovery of U-shaped and inverted-U scaling trends, where an initial trend reverses, suggesting that scaling trends are less reliable at predicting the behavior of larger-scale models than previously understood. Overall, our results suggest that there are tasks for which increased model scale alone may not lead to progress, and that more careful thought needs to go into the data and objectives for training language models.

[-] Akisamb@programming.dev 11 points 1 year ago

R2 by cloudfare does not charge egress costs. It's 0.015$/GB/month for storage. Read operations are 0.36$ per million.

I do have a hard time believing that they will remain this cheap though, but who knows.

1
Inside the AI Factory (www.theverge.com)
view more: next ›

Akisamb

joined 1 year ago
MODERATOR OF