The decision Sunday is a major blow to efforts in the United States attempting to rein in the homegrown industry that is rapidly evolving with little oversight
The measure, aimed at reducing potential risks created by AI, would have required companies to test their models and publicly disclose their safety protocols to prevent the models from being manipulated to, for example, wipe out the state’s electric grid or help build chemical weapons.
How exactly do LLMs do that? If you’ve given an LLM’s pseudorandom output control over your electrical grid, no regulation will mitigate your stupidity.
If you hook an LLM up as an interface replacement for a manual/analog Power Plant interface and start asking the translator to intuit decisions based on fuzzy inputs, you can create a cascade of errors that result in grid failure.
If you’ve given an LLM’s pseudorandom output control over your electrical grid, no regulation will mitigate your stupidity.
This rule would prevent a business or public regulator from doing such a thing without proving out safeguards.
How exactly do LLMs do that? If you’ve given an LLM’s pseudorandom output control over your electrical grid, no regulation will mitigate your stupidity.
Removed by mod
Could he understand the halting problem? I doubt he does, but the legislators evidently don’t either
If you hook an LLM up as an interface replacement for a manual/analog Power Plant interface and start asking the translator to intuit decisions based on fuzzy inputs, you can create a cascade of errors that result in grid failure.
This rule would prevent a business or public regulator from doing such a thing without proving out safeguards.
And the governor vetoed it.