< Notes...

there really are no uninteresting things...
24 Mar 2026 @ 14:00:21

I feel morally obligated to say I did not write the code in this repository myself. This project is an exploration of using LLMs to carry out tasks based on my direction. The majority of prompts I used to get here were derived using the socratic method, genuine curiosity, and a hunch that NVMe supporting inference is underutilized despite being a (slow but) perfectly valid form of memory.

Translation: I slopcoded this sh*t. Not that it is super bad, I just found funny the way it was written.

➝ Via Hacker News.

llm tech via
< Older Newer >