Building an LLM Robot with My Son — EP 9. 4-Month Retrospective, and What Comes Next

Building an LLM Robot with My Son — EP 9. 4-Month Retrospective, and What Comes Next

When I wrote EP 0, there was a half-assembled acrylic chassis on the desk with one wheel turning in reverse.

Now that robot walks to the kitchen on its own and finds a water glass. Can't pick it up — but it finds it.

Four months.


What Worked

The agent harness approach held up

Injecting domain knowledge through CLAUDE.md worked better than expected. The repeated context fatigue disappeared — AI wrote code that respected the project's rules without having to be reminded every session. The file started at ten lines and grew to 120. That growth is a record of the project itself.

My son completed a behavior through eight prompts

The scene from EP 5 is the one that stays with me. He sat down alone, iterated eight prompts, and built a working obstacle avoidance behavior. Never touched a line of code. "I made this" was not wrong.

Apple Silicon local LLM actually works

112 tok/s on M4 Pro. Sub-500ms round trips over home LAN, no cloud API latency. Running a local LLM server at home turned out to be more practical than I expected.


What Surprised Me

The Arduino → Pi migration came sooner than planned

The moment I built the laptop bridge in EP 6, it was obvious that was temporary. The series structure had Pi migration planned for EP 7, but honestly I wanted to do it by EP 4–5. The episode-by-episode format made me wait.

Vision LLM worked better than expected

When "go to the kitchen and find the water glass" actually succeeded on the third try in EP 8 — I was genuinely surprised. No hardcoded map, no coordinates, just a camera feed and a goal. A 7B parameter model doing that was not something I'd have predicted two years ago.

He started reading error messages

Early in the project, any error meant coming to find me. Now when the terminal shows red, he reads it first. He doesn't understand everything, but he finds the line number. I didn't teach him that. It just happened through doing.


What Didn't Work

He still doesn't understand code structure

What a function is, why there's a loop — he doesn't know. He can build behaviors through vibe coding but can't explain why the code works. I don't think that's a problem right now, but there will be a point where fundamentals need to come in.

Sensor noise was never fully resolved

HC-SR04 returning 0cm or 400cm with no obstacle present. We filtered it multiple times. It never went completely away. Hardware-level noise that software can't fully suppress — would need a sensor swap or a reworked level-shifter circuit. Didn't get there this season.

Didn't buy the Pi 5

Honest admission: I wrote in EP 0 that I'd buy Pi 5 if blog revenue materialized. Four months later, revenue did materialize. Pi 4 works well enough that I didn't buy it. Didn't need to.


What He Learned in Four Months

The original goal was "teach him to code." More accurately, we did it together.

What he can do now:
- Describe a desired behavior to AI in plain language
- Narrow down what's wrong when something behaves unexpectedly
- Read Serial monitor output and notice when sensor values are off
- Context check prompt (the routine he invented in EP 2)
- Hardware wiring — power rail separation, checking voltage before connecting

What he still can't do:
- Modify code directly
- Interpret compile errors
- Design circuits

The can't list is longer. Of course — four months, twelve years old.

But the instinct for describing what you want, working through conversation with AI, and getting a result is there. That's what this project set out to build. The rest can follow.


Season 2: LEGO

He already knows.

"Next time let's build it in LEGO." That came up in EP 0 and kept resurfacing for four months. Every time he watched the acrylic chassis move: "I think it would work better in LEGO Technic."

Season 2: my son designs the robot body himself in LEGO Technic. Gear ratios, torque, wheel configuration — this is territory he already understands intuitively from years of building. LEGO Mindstorms or Spike Prime integration means programmable motors without needing Pi. The existing LLM server harness carries over.

Pi 5 happens then. LEGO Technic too. When the budget is there.


If you've read from EP 0 to here: I wanted it to feel like watching someone figure it out in real time, not a polished retrospective. Not a finished project showcase. A record of conversations that built something.

My son knows this series has been going up. At first he was a little self-conscious about his story being on the internet. Now he's the one who says "post what we did today."

That's probably the best signal.

댓글

이 블로그의 인기 게시물

개발자는 코드를 쓰는 사람이 아니다 — AI 시대에 남는 자리는 '책임'에 있다

Harness Engineering in Practice — How Anthropic Designs AI Agents

What Is Harness Engineering — Designing the Reins for AI Agents