top | item 47051604

(no title)

the_duke | 12 days ago

An Anthropic safety researcher just recently quit with very cryptic messages , saying "the world is in peril"... [1] (which may mean something, or nothing at all)

Codex quite often refuses to do "unsafe/unethical" things that Anthropic models will happily do without question.

Anthropic just raised 30 bn... OpenAI wants to raise 100bn+.

Thinking any of them will actually be restrained by ethics is foolish.

[1] https://news.ycombinator.com/item?id=46972496

discuss

order

mobattah|12 days ago

“Cryptic” exit posts are basically noise. If we are going to evaluate vendors, it should be on observable behavior and track record: model capability on your workloads, reliability, security posture, pricing, and support. Any major lab will have employees with strong opinions on the way out. That is not evidence by itself.

Aromasin|12 days ago

We recently had an employee leave our team, posting an extensive essay on LinkedIn, "exposing" the company and claiming a whole host of wrong-doing that went somewhat viral. The reality is, she just wasn't very good at her job and was fired after failing to improve following a performance plan by management. We all knew she was slacking and despite liking her on a personal level, knew that she wasn't right for what is a relatively high-functioning team. It was shocking to see some of the outright lies in that post, that effectively stemmed from bitterness at being let go.

The 'boy (or girl) who cried wolf' isn't just a story. It's a lesson for both the person, and the village who hears them.

spondyl|12 days ago

If you read the resignation letter, they would appear to be so cryptic as to not be real warnings at all and perhaps instead the writings of someone exercising their options to go and make poems

axus|12 days ago

I think the perils are well known to everyone without an interest in not knowing them:

Global Warming, Invasion, Impunity, and yes Inequality

imiric|12 days ago

[deleted]

skybrian|12 days ago

The letter is here:

https://x.com/MrinankSharma/status/2020881722003583421

A slightly longer quote:

> The world is in peril. And not just from AI, or from bioweapons, gut from a whole series of interconnected crises unfolding at this very moment.

In a footnote he refers to the "poly-crisis."

There are all sorts of things one might decide to do in response, including getting more involved in US politics, working more on climate change, or working on other existential risks.

user2722|12 days ago

Similar to Peripheral TV series' Jackpot?

stronglikedan|12 days ago

Not to diminish what he said, but it sounds like it didn't have much to do with Anthropic (although it did a little bit) and more to do with burning out and dealing with doomscoll-induced anxiety.

zamalek|12 days ago

I think we're fine: https://youtube.com/shorts/3fYiLXVfPa4?si=0y3cgdMHO2L5FgXW

Claude invented something completely nonsensical:

> This is a classic upside-down cup trick! The cup is designed to be flipped — you drink from it by turning it upside down, which makes the sealed end the bottom and the open end the top. Once flipped, it functions just like a normal cup. *The sealed "top" prevents it from spilling while it's in its resting position, but the moment you flip it, you can drink normally from the open end.*

Emphasis mine.

lanyard-textile|12 days ago

He tried this with ChatGPT too. It called the item a "novelty cup" you couldn't drink out of :)

vunderba|12 days ago

> Codex quite often refuses to do "unsafe/unethical" things that Anthropic models will happily do without question.

I can't really take this very seriously without seeing the list of these ostensible "unethical" things that Anthropic models will allow over other providers.

ljm|12 days ago

I'm building a new hardware drum machine that is powered by voltage based on fluctuations in the stock market, and I'm getting a clean triangle wave from the predictive markets.

Bring on the cryptocore.

ReptileMan|12 days ago

>Codex quite often refuses to do "unsafe/unethical" things that Anthropic models will happily do without question.

Thanks for the successful pitch. I am seriously considering them now.

tsss|12 days ago

Good. One thing we definitely don't need any more of is governments and corporations deciding for us what is moral to do and what isn't.

bflesch|12 days ago

Wasn't that most likely related to the US government using claude for large-scale screening of citizens and their communications?

astrange|12 days ago

I assumed it's because everyone who works at Anthropic is rich and incredibly neurotic.

WesolyKubeczek|12 days ago

> Codex quite often refuses to do "unsafe/unethical" things that Anthropic models will happily do without question.

That's why I have a functioning brain, to discern between ethical and unethical, among other things.

catoc|12 days ago

Yes, and most of us won’t break into other people’s houses, yet we really need locks.

toddmorey|12 days ago

You are not the one folks are worried about. US Department of War wants unfettered access to AI models, without any restraints / safety mitigations. Do you provide that for all governments? Just one? Where does the line go?

idiotsecant|12 days ago

That guys blog makes him seem insufferable. All signs point to drama and nothing of particular significance.

manmal|12 days ago

Codex warns me to renew API tokens if it ingests them (accidentally?). Opus starts the decompiler as soon as I ask it how this and that works in a closed binary.

kaashif|12 days ago

Does this comment imply that you view "running a decompiler" at the same level of shadiness as stealing your API keys without warning?

I don't think that's what you're trying to convey.

ACCount37|12 days ago

Opus <3. My go-to for reverse engineering tasks.