In A Fire Upon the Deep, there's this Oliphaunt toy that some of the characters come to consult as a dataset or knowledge oracle, right?
I want to make something like that, a plushie with an embedded LLM engine. Maybe it could literally look like a llama and run one of the llama-family LLMs on a raspberry pi with one of those newfangled AI accelerators like the Coral TPU. Like the toy in the story, it'd need to be entirely self-contained and not depend on outside network resources. Presumably it has a battery you can charge. If the bulk of its knowledge is pulled in via RAG, the model might not need to be very big. Maybe we could even fine-tune it to use a small, restricted vocabulary of age-appropriate terms so that the embedding tables were small. Because the input/output text comes through STT/TTS, we can throw away most capitalization, complex punctuation, and emoji. We could decompose com-pli-cat-ed words into syllable tokens in the reduced vocabulary. It makes sense that the plushie has a hard ti