this post was submitted on 13 Aug 2023
228 points (100.0% liked)

Technology

37362 readers
275 users here now

Rumors, happenings, and innovations in the technology sphere. If it's technological news or discussion of technology, it probably belongs here.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS
 

I asked Google Bard whether it thought Web Environment Integrity was a good or bad idea. Surprisingly, not only did it respond that it was a bad idea, it even went on to urge Google to drop the proposal.

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 9 points 11 months ago (1 children)

Large language models literally do subspace projections on text to break it into contextual chunks, and then memorize the chunks. That's how they're defined.

Source: the paper that defined the transformer architecture and formulas for large language models, which has been cited in academic sources 85,000 times alone https://arxiv.org/abs/1706.03762

[–] [email protected] 6 points 10 months ago (1 children)

Hey, that comment's a bit off the mark. Transformers don't just memorize chunks of text, they're way more sophisticated than that. They use attention mechanisms to figure out what parts of the text are important and how they relate to each other. It's not about memorizing, it's about understanding patterns and relationships. The paper you linked doesn't say anything about these models just regurgitating information.

[–] [email protected] 4 points 10 months ago

I believe your "They use attention mechanisms to figure out which parts of the text are important" is just a restatement of my "break it into contextual chunks", no?