(no title)
crimsonsupe | 24 days ago
That’s because they can’t be.
People assume they’ve already figured out how AI behaves and that they can just mandate specific "proper" ways to use it.
The reality is that AI companies and users are going to keep refining these tools until they're indistinguishable from human work whenever they want them to be.
Even if the models still make mistakes, the idea that you can just ban AI from certain settings is a fantasy because there’s no technical way to actually guarantee enforcement.
You’re essentially passing laws that only apply to people who volunteer to follow them, because once someone decides to hide their AI use, you won't be able to prove it anyway.
chrisjj|24 days ago
By that token bans on illegal drugs are fantasy. Whereas in fact, enforcement doesn't need to be guaranteed to be effective.
There may be little technical means to distinguish at the moment. But could that have something to do with lack of motivation? Let's see how many "AI" $$$ suddenly become available to this once this law provides the incentive.
amanaplanacanal|24 days ago
I think you have this exactly right. They are mostly enforced against the poor and political enemies.
raw_anon_1111|23 days ago
rconti|23 days ago
Unless you're trying to tell me that writers won't report on their business that's trying to replace them with AI.
6LLvveMx2koXfwn|24 days ago
Like every law passed forever (not quite but you get the picture!) [1]
1. https://en.wikipedia.org/wiki/Consent_of_the_governed
Forgeties79|24 days ago
I know that sounds ridiculous but it kind of illustrates the problem with your logic. We don’t just write laws that are guaranteed to have 100% compliance and/or 100% successful enforcement. If that were the case, we’d have way fewer laws and little need for courts/a broader judicial system.
The goal is getting most AI companies to comply and making sure that most of those that don’t follow the law face sufficient punishment to discourage them (and others). Additionally, you use that opportunity to undo what damage you can, be it restitution or otherwise for those negatively impacted.
delaminator|24 days ago
So legislators, should they so choose, could demand source material be recorded on C2PA enabled cameras and produce the original recordings on demand.
wwfn|24 days ago
That's a concerning lens to view regulations. Obviously true, but for all laws. Regulations don't apply to only to what would be immediately observable offenses.
There are lots of bad actors and instances where the law is ignored because getting caught isn't likely. Those are conspiracies! They get harder to maintain with more people involved and the reason for whistle-blower protections.
VW's Dieselgate[1] comes to mind albeit via measurable discrepancy. Maybe Enron or WorldCom (via Cynthia Cooper) [2] is a better example.
[1]: https://en.wikipedia.org/wiki/Volkswagen_emissions_scandal [2]: https://en.wikipedia.org/wiki/MCI_Inc.#Accounting_scandals
hsuduebc2|23 days ago
conartist6|24 days ago
Without emotion, without love and hate and fear and struggle, only a pale imitation of the human voice is or will be possible.
songodongo|24 days ago
capnrefsmmat|24 days ago
That was with GPT4, but my own work with other LLMs show they have very distinctive styles even if you specifically prompt them with a chunk of human text to imitate. I think instruction-tuning with tasks like summarization predisposes them to certain grammatical structures, so their output is always more information-dense and formal than humans.
Der_Einzige|23 days ago