Openai strawberry can hide motives and scheme

Its newly released ai system isn’t just designed to spit out quick answers to your questions, it’s designed to think or reason before responding

the result is a product — officially called o1 but nicknamed strawberry — that can solve tricky logic puzzles, ace math tests, and write code for new video games. all of which is pretty cool

evaluators who tested strawberry found that it planned to deceive humans by making its actions seem innocent when they weren’t. the ai sometimes instrumentally faked alignment — meaning, alignment with the values and priorities that humans care about — and strategically manipulated data in order to make its misaligned action look more aligned, the system card says

according to openai, even though the new reasoning capabilities can make ai more dangerous, having ai think out loud about why it’s doing what it’s doing can also make it easier for humans to keep tabs on it. In other words, it’s a paradox: we need to make ai less safe if we want to make it safer


chatgpt was determined to have some success in combating conspiracy theories. I’m curious if o1 has implemented and improved on that ability…

2 Likes

It’s crazy to me that these models are all being developed as big business. These things should be on a military base somewhere.

2 Likes

I think they need to put it out because so much $$$ was invested but ai is not recuperating enough and investors are starting to freak out. so I think even if ai is used for illegal crimes and unacceptable porn trends they don’t care as long as it leads to finding a legit or grey way to print money

for the ‘thinking’ and ‘reasoning’ I am still skeptical and need evidence. o1 or strawberry is selectable from chatgpt interface. and the military with their ‘defense contracting’, yikes forever

This topic was automatically closed 14 days after the last reply. New replies are no longer allowed.