Question: where would a flash-resident tiny language runtime fit relative to on-device GenAI runtimes? #2033
Alpha-Guardian
started this conversation in
Show and tell
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi ONNX Runtime GenAI folks,
I wanted to share a small on-device language-runtime experiment and ask how systems like this should be viewed relative to more familiar on-device GenAI runtimes.
We built a public demo line called Engram and deployed it on a commodity ESP32-C3.
Current public numbers:
Host-side benchmark capability
LogiQA = 0.392523IFEval = 0.780037Published board proof
LogiQA 642 = 249 / 642 = 0.3878504672897196host_full_match = 642 / 6421,380,771 bytesImportant scope note:
This is not presented as unrestricted open-input native LLM generation on MCU.
The board-side path is closer to a flash-resident, table-driven runtime with:
So this is not a standard portable GenAI runtime story. It is closer to a task-specialized language runtime whose behavior has been compiled into a very constrained execution form.
Repo:
https://github.com/Alpha-Guardian/Engram
What I’m curious about is whether systems like this should be viewed as:
Would love to hear any thoughts.
Beta Was this translation helpful? Give feedback.
All reactions