• veee@lemmy.ca
    link
    fedilink
    English
    arrow-up
    0
    ·
    10 months ago

    According to Air Canada, Moffatt never should have trusted the chatbot and the airline should not be liable for the chatbot’s misleading information because Air Canada essentially argued that “the chatbot is a separate legal entity that is responsible for its own actions,” a court order said.

    “Air Canada argues it cannot be held liable for information provided by one of its agents, servants, or representatives—including a chatbot,” Rivers wrote.

    • RegalPotoo@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      10 months ago

      The thing is, none of that is even slightly true; even if the chatbot were it’s own legal entity, it would still be an employee and air Canada are liable for bad advice given by their representatives

  • RegalPotoo@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    10 months ago

    I wonder if this will turn into a new attack vector against companies; talk their LLM chat bots into promising a big discount, take the company to a small claims court to cash out