• Overspark@feddit.nl
    link
    fedilink
    English
    arrow-up
    5
    ·
    2 days ago

    You can. There are simple options, that only recognise predefined sentences, that even work on a Raspberry Pi, and at the other end of the spectrum you can host an LLM locally and chat with that if you have the right hardware (Coral isn’t powerful enough for that, you want a GPU with lots of VRAM). Obviously setting this up is more complicated, but there are a lot of options to do it your way.