Shall we trust LM defining legal definitions, deepfake in this case? It seems the state rep. is unable to proof read the model output as he is “really struggling with the technical aspects of how to define what a deepfake was.”
I understand the irony. But can we not pretend they blindly used an output or even generated a full page. It was a specific section to provide a technical definition of “what is a deepfake”.
“I was really struggling with the technical aspects of how to define what a deepfake was. So I thought to myself, ‘Well, why not ask the subject matter expert (i do not agree with that wording, lol) , ChatGPT?’” Kolodin said.
The legislator from Maricopa County said he “uploaded the draft of the bill that I was working on and said, you know, please, please put a subparagraph in with that definition, and it spit out a subparagraph of that definition.”
“There’s also a robust process in the Legislature,” Kolodin continued. “If ChatGPT had effed up some of the language or did something that would have been harmful, I would have spotted it, one of the 10 stakeholder groups that worked on or looked at this bill, the ACLU would have spotted, the broadcasters association would have spotted it, it would have got brought out in committee testimony.”
But Kolodin said that portion of the bill fared better than other parts that were written by humans. “In fact, the portion of the bill that ChatGPT wrote was probably one of the least amended portions,” he said.
I do not agree on his statement that any mistakes made by ai could also be made by humans. The reasoning and errors in reasoning is quite different in my experience but the way chatgpt was used is absolutely fair.
No kidding. When I read that, my first thought was, “He’s clearly at least above the median intelligence of his fellow Arizona GOP reps, if not in the top 10% of their entire conference”
Anyone who read the article AND has experience with the Arizona GOP, probably thought the same thing.
The Arizona GOP collects some of the dumbest people alive.
I get this feeling this will generally be the peak of generative AI. Used for assistance when needed and with lots of oversight. The problem is that not all people bother to check the AI’s work.
These types of things are exactly what Generative AI models are good for, as much as Internet people don’t want to hear it.
Things that are massively repeatable based off previous versions (like legislation, contracts, etc) are pretty much perfect for it. These are just tools for already competent people. So in theory you have GenAI crank out the boring stuff and have an expert “fill in the blanks” so to speak
Ideally it would be a generative AI trained specifically on legal textbooks.
I don’t know why there seem to be no LLMs trained specifically on expert subject matter.
There are, just not available publicly. Tons of enterprises (law firms included) are paying to have models trained on their data
The stupid. It hurts.
And yet again it cynically amuses me that AI has become “artificial” intelligence in the sense of “fake.”
It’s a shabby substitute for real intelligence, used by people who don’t possess any of their own to impress other people who don’t possess any of their own.
This is actually true.
Most notably to me, the ability to sift through and collate enormous amounts of data has led to surprising things like diagnosing diabetes through retinal scans.
But those sorts of things, beneficial and impressive though they might be, remain at the fringe of AI research for the simple reason that those sorts of uses are too niche to provide the revenue stream that all of the bubble-building corporate parasites demand. Their focus is on the AI-as-a-substitute-for-real-intelligence aspect (and increasingly “AI” as just a meaningless marketing buzzword), since that’s where the money is. And unfortunately but not coincidentally, that’s where most of the public attention is too.