writing.exchange is one of the many independent Mastodon servers you can use to participate in the fediverse.
A small, intentional community for poets, authors, and every kind of writer.

Administered by:

Server stats:

331
active users

#anthropic

5 posts5 participants0 posts today

No, #AI frontier models don't "just guess words", it's far more complicated than that.

#Anthropic built an #LLM "brain scanner" (so far AIs have been black boxes).

According to Anthropic, "it currently takes a few hours of human effort to understand the circuits we see, even on prompts with only tens of words." And the research doesn't explain how the structures inside LLMs are formed in the first place.

#OpenAi, #Anthropic, and other #LLM model vendors are starting to look a lot like #Docker - a ubiquitous technology with no real moat and no way to avoid becoming a commodity with razor thin profit margins.
These companies will have a hard time competing with small end-user focused competitors that provide nicely packed #AI based apps for specific users and use-cases.

Who is *really* using the newest AI models?

Anthropic recent launched the Anthropic Economic Index, "aimed at understanding AI's effects on labor markets and the economy over time." In their latest post, they explore who is making the most use of their new extended thinking models.

The biggest users of the new models? Computer research scientists, software developers, multimedia artists, and video game designers!

Find out more here:

anthropic.com/news/anthropic-e
#AI #GenAI #Anthropic #GenerativeAI

#Anthropic vient de publier des études révélant comment son modèle #Claude "réfléchit" réellement.

Les chercheurs ont découvert que l' #IA #AI planifie ses réponses à l'avance, pense dans un #langage conceptuel #universel et peut même parfois fournir des explications qui ne reflètent pas son véritable processus interne.

lesnumeriques.com/intelligence

Les Numériques · Comment fonctionne vraiment une IA ? Les chercheurs d'Anthropic ont enfin un début de réponseBy Sofian Nouira
Continued thread

"Why do language models sometimes hallucinate—that is, make up information? At a basic level, language model training incentivizes hallucination: models are always supposed to give a guess for the next word. Viewed this way, the major challenge is how to get models to not hallucinate. Models like Claude have relatively successful (though imperfect) anti-hallucination training; they will often refuse to answer a question if they don’t know the answer, rather than speculate. We wanted to understand how this works.

It turns out that, in Claude, refusal to answer is the default behavior: we find a circuit that is "on" by default and that causes the model to state that it has insufficient information to answer any given question. However, when the model is asked about something it knows well—say, the basketball player Michael Jordan—a competing feature representing "known entities" activates and inhibits this default circuit (see also this recent paper for related findings). This allows Claude to answer the question when it knows the answer. In contrast, when asked about an unknown entity ("Michael Batkin"), it declines to answer.

Sometimes, this sort of “misfire” of the “known answer” circuit happens naturally, without us intervening, resulting in a hallucination. In our paper, we show that such misfires can occur when Claude recognizes a name but doesn't know anything else about that person. In cases like this, the “known entity” feature might still activate, and then suppress the default "don't know" feature—in this case incorrectly. Once the model has decided that it needs to answer the question, it proceeds to confabulate: to generate a plausible—but unfortunately untrue—response."

anthropic.com/research/tracing