As they improve, we’ll likely trust AI models with more and more responsibility. But if their autonomous decisions end up causing harm, our current legal frameworks may not be up to scratch.
Thankfully there’s not: you’d expect someone at a pharmacy to provide reasonable medical advice, or your mechanic to tell you the right thing to do with your car. Once you walk outside the field where a reasonable person would reasonably expect what they’re being told to be uh, reasonable, then there’s usually no real case for liabilities.
Buuuuuut, in the US at least, this is entirely civil law, and that means the law is mostly whatever you can convince a jury of, so you can end up with some wacky shit happening.
Thankfully there’s not: you’d expect someone at a pharmacy to provide reasonable medical advice, or your mechanic to tell you the right thing to do with your car. Once you walk outside the field where a reasonable person would reasonably expect what they’re being told to be uh, reasonable, then there’s usually no real case for liabilities.
Buuuuuut, in the US at least, this is entirely civil law, and that means the law is mostly whatever you can convince a jury of, so you can end up with some wacky shit happening.