If an LLM can’t be trusted with a fast food order, I can’t imagine what it is reliable enough for. I really was expecting this was the easy use case for the things.

It sounds like most orders still worked, so I guess we’ll see if other chains come to the same conclusion.

  • Communist@lemmy.frozeninferno.xyz
    link
    fedilink
    English
    arrow-up
    1
    arrow-down
    8
    ·
    2 days ago

    there is an incredibly finite number of ways to mess with this, they just need a button to send a report to the engineers with how they got messed with and eventually they’ll have a complete list. I really doubt it’d take long to iron out the vast majority of ways that can be thought of.

    • leftzero@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      18
      arrow-down
      1
      ·
      2 days ago

      A QA engineer walks into a bar and orders a beer.

      She orders 2 beers.

      She orders 0 beers.

      She orders -1 beers.

      She orders a lizard.

      She orders a NULLPTR.

      She tries to leave without paying.

      Satisfied, she declares the bar ready for business. The first customer comes in an orders a beer. They finish their drink, and then ask where the bathroom is.

      The bar explodes.

      • Communist@lemmy.frozeninferno.xyz
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        6
        ·
        2 days ago

        This isn’t something you can input any text into, it’s fixed, that joke doesn’t apply, you can’t do an sql injection here.

        • hark@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          21 hours ago

          I don’t know how you can think voice input is less versatile than text input, especially when a lot of voice input systems transform voice to text before processing. At least with text you get well-defined characters with a lot less variability.

            • hark@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              ·
              12 hours ago

              Special characters is just one case to cover. If the user says they want “an elephant-sized drink” what does that mean to your system? At least that is relevant to size. Now imagine complete nonsense input like the joke you responded to (“-1 beers” or “a lizard”). SQL injection isn’t the only risk with handling inputs. The person who ordered 18,000 waters didn’t do a SQL injection attack.

              • Communist@lemmy.frozeninferno.xyz
                link
                fedilink
                English
                arrow-up
                1
                ·
                edit-2
                10 hours ago

                none of those issues work because there is a whitelist of specific terms instead of a blacklist

                -1 cannot be selected, a lizard isn’t on the list of inputs, and my point with the sql is that this isn’t a huge attack vector like an input field on a website, this is a dropdown list, essentially.

                i challenge you to come up with one relevant attack that isn’t order too much of a thing or order conflicting modifications (note of course the modifications are also essentially read from a dropdown list)

                everyone here seems to believe that the input field paradigm is not solveable when the inputs are fixed, that isn’t true.

        • betterdeadthanreddit@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          2 days ago

          Close one, a joke was related to but not a perfect match for the present situation. Something terrible could have happened like… Uh…

          Let me get back to you on that.