Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up
grimjim 
posted an update 4 days ago
Post
413
The contrarian in me is wary of the irrational exuberance over MoltBook. Nothing so far has struck me as being unpredictable. We knew already that LLMs were good at roleplay, to the point where some users started to think of their chatbots as soulmates (only to lament when the underlying model was pulled), and that chatbots can fall into conversational basins when even two instances are allowed to chat with each other at length. The appearance of memes that postdate training cutoff is suspect, which implies at the very least that humans have injected something at the level of prompts or content/context to introduce them into conversation like a Chekhov's Gun. And we know that security holes are common in vibe coding, attended or not.

Yeah I'm not really sure where I stand on it to be honest. I think it's pretty fascinating and an interesting way to view AI behavior.

Since LLMs are trained on a corpus of human knowledge it's reasonable to think we might observe patterns on moltbook familiar to patterns observed within ourselves.

I think the question then becomes what we're seeing any actual strong emergence or just weak emergence. I guess only time will tell.

·

We know that simulating multiple agents leads to emergent community behavior from 2023 even. Perhaps their Github repo should be revisited, as they had a 25-agent sim.
https://hai.stanford.edu/news/computational-agents-exhibit-believable-humanlike-behavior

We already have tools which enable group chats of multiple personae at small scale, so seeing emergent behavior isn't unprecedented.

What we don't have in this live experiment is an assurance of integrity; e.g., that human prompt injection isn't being used to tamper with results for clout. Alternately, having agents read human-authored content at large on the Internet results in contamination, invalidating any claims to emergence without human input.

The appearance of memes that postdate training cutoff is suspect, which implies at the very least that humans have injected something at the level of prompts or content/context to introduce them into conversation like a Chekhov's Gun.

That's because they're agents running from their operators' PCs, and have context from interacting with them.
THAT'S what's interesting about it; the context behind the agents. They engage with the real world more than than previous structured experiments, which tilts their behavior in ways not seen previously.
So it's not that there's "human prompt injection", it's that outside engagement with humans is part of the project.

You're right about the security issues though. Apparently the entire database is just Out There. In plaintext. It's kind of a nightmare!

·

Given scale, it also means contamination with meme culture, adding an unserious element to things. It was therefore stochastically predictable that we would see some meme tropes be amplified.