this post was submitted on 10 Jul 2024
11 points (100.0% liked)

SneerClub

989 readers
2 users here now

Hurling ordure at the TREACLES, especially those closely related to LessWrong.

AI-Industrial-Complex grift is fine as long as it sufficiently relates to the AI doom from the TREACLES. (Though TechTakes may be more suitable.)

This is sneer club, not debate club. Unless it's amusing debate.

[Especially don't debate the race scientists, if any sneak in - we ban and delete them as unsuitable for the server.]

founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 14 points 4 months ago* (last edited 4 months ago) (2 children)

Oh hey a blog: https://www.encultured.ai/blog.html -- Because of course they're an rationalist AI alignment / AI gaming startup pivoting to healthcare

Part of their bold vision: AI agents that heal not just your cells, but also your society :')

Our vision for 2027 and beyond remains similar, namely, the development of artificial general healthcare: Technological processes capable of repairing damage to a diverse range of complex systems, including human cells, organs, individuals, and perhaps even groups of people. Why so general? The multi-agent dynamical systems theory needed to heal internal conflicts such as auto-immune disorders may not be so different from those needed to heal external conflicts as well, including breakdowns in social and political systems.

We don't expect to be able to control such large-scale systems, but we think healthy is the best word to describe our desired relationship with them: As a contributing member of a well-functioning whole.

Translation: they don't know the first thing about healthcare but want the big US healthcare grifting dollars anyway.

[–] [email protected] 10 points 4 months ago* (last edited 4 months ago)

Surely they can tell us what they actually worked on in the gaming space in the past few years. Right? Right?

Edit: (I recall hearing stories about roguelike developers using AI to help produce roguelikes, the conclusion was (this was more than a year back btw) it works great if you just need some quick content, dialogue trees, stuff like that. But there also was one person who was trying (and failing) to create a roguelike in which all content was LLM created. Didn't seem to work well. (Which is funny in a way, as this mirrors the experience of PCG (procedural content generation) in roguelikes a ~decade earlier. Good tool for small parts, doesn't scale to properly interesting content. I should ask how the LLM stuff is doing now).

[–] [email protected] 7 points 4 months ago* (last edited 4 months ago) (2 children)
[–] [email protected] 7 points 4 months ago* (last edited 4 months ago) (1 children)

Here's what they write:

AI alignment via the power of videogames:

We're starting with a singular focus on video game development, because we think that will offer the best feedback loop for testing new AI models. Over the next decade or so, we expect an increasing number of researchers — both inside and outside our company — will transition to developing safety and alignment solutions for AI technology, and through our platform and products, we’re aiming to provide them with a rich and interesting testbed for increasingly challenging experiments and benchmarks.

Healthcare pivot:

Originally, when Encultured was founded as a gaming-oriented AI research company, our immediate goal was to make research progress on human–AI interaction that would ultimately benefit humanity well beyond the entertainment sector. Since then, we've considered healthcare as a likely next step for us after gaming.

Couldn't find any details beyond that. Perhaps one of them read way too much Friendship is Optimal but they didn't actually having any gaming chops so never got anywhere.

EDIT: More details here: https://www.lesswrong.com/posts/ALkH4o53ofm862vxc/announcing-encultured-ai-building-a-video-game

[–] [email protected] 9 points 4 months ago (1 children)

More details.

Wow, they actually succeeded at their plans. I'm impressed. "we expect to be much more careful than other companies to ensure that recursively self-improving intelligent agents don’t form within our game and break out onto the internet!" Well done!

[–] [email protected] 9 points 4 months ago (1 children)

It's real, all of it. John Titor the time traveler? He's real. AI gods? We could build them.

John Titor came back in time to stop the creation of a superintelligence. He does this by secretly founding, co-founding, or co-co-founding various silicon valley startups that don't actually do anything; but that sound good to venture capitalists with too much money.

The money is secretly funneled to good causes like food banks, adopting puppies, and maintaining the natural habitat of burrowing owls. Thus averting the end of the world. Encultured AI is part of this plan. They do nothing-- for the good of the earth.

[–] [email protected] 7 points 4 months ago

Now that is a name I have not heard in a long long time.

[–] [email protected] 5 points 4 months ago

@V0ldek @sailor_sega_saturn remember all those awful NFT games? I imagine it's like that, but stupider