Weekly Catch-Up: Mar 16-22, 2026¶
Prepared Sunday March 22, 3:09 PM. Tomorrow is Meeting #10.
Sync Status¶
| Source | Result |
|---|---|
| Google Docs | 17 docs synced (added Meeting #9 notes) |
| Telegram | 861 messages across 6 topics |
| GitHub | 0 open PRs, 8 open issues |
Meeting #9 Happened (Mar 16)¶
Yaroslav presented the roadmap. Participants: Yaroslav, Moorissa Tjokro (SPC, robotics/autonomous vehicles), Anastasiia Zhiboedova (ML Engineer, Amazon AGI), Michael Keating (data center cooling, attending NVIDIA GTC), Jonathan Belay, Yad Konrad (async via pre-recorded video), JackJack Ganbold (SPC), Andrew, Preston Schmittou.
Key outcomes:
- Metric shift: ARD to DMC (Data Movement Complexity, Ding et al. arXiv:2312.14441). The new homework is to optimize DMC instead of ARD. DMC uses sqrt(stack_distance) per access, which maps to physical wire-length energy on a 2D memory layout.
- Meta-goal: iterate on the process of going from "metric + problem specification" to a fast sequence of experiments. Not just solving the problem, but making the solving fast.
- Meeting video: YouTube
- AI slides: Sutro_Efficiency_Synthesis.pdf
New Ideas from Telegram This Week¶
Potential high-profile visitors (chat-yaroslav, Mar 21)¶
- Lukas Kaiser left OpenAI, doing open-source research. Plans to stop by the Mar 30 meeting.
- Alec Radford also doing OSS research now. Yaroslav wants to involve both.
RL environment framing (chat-yaroslav, Mar 21)¶
- Yaroslav wants to wrap our algorithmic challenges into RL environments and give them to companies like Anthropic. If it makes Claude better, that accelerates our own auto-research loops.
- PrimeIntellect has a research grants program (compute + stipends) for novel environments.
- Yad noted that our 33 experiments are basically an answer key -- did the agent rediscover GF(2)? Did it figure out local learning rules fail? That's richer signal than most RL envs.
Discrete ML / Wolfram (general, Mar 19-20)¶
- Yaroslav shared Wolfram's work on training neural nets with pure discrete Boolean logic (AND/XOR grids). No backprop, no floats.
- Seth Stafford: "neural networks are just a quantization of random forests. You recover the random forest in a semi-classical limit."
- 8-bit integer multiply is 5x less Joules than 16-bit FP. 8-bit integer adds: ~50x cheaper.
Repo / logistics (chat-yad, Mar 21)¶
- Yaroslav asked if SutroYaro can be designated Public Domain. Yad said yes.
- Yaroslav visiting Manhattan Mar 24-30 (this coming week). Wants to meet up.
- Video quality note: videos came through at 720p, Yaroslav asked for 4k next time.
Michael's Autoresearch fork (challenge #1, Mar 16)¶
- Michael forked Karpathy's new Autoresearch, pointed it at sparse parity, asked Opus to use "unconventional or ancient mathematical theories" to avoid leaning on conventional wisdom.
DMC Infrastructure Inventory¶
Before running experiments, we audited what exists for the DMC metric shift.
| Component | Status | Notes |
|---|---|---|
tracker.py (DMC formula) |
DONE | sum(size * sqrt(distance)) per Ding et al. |
cache_tracker.py (LRU + DMC) |
DONE | Inherits DMC, adds cache simulation |
harness.py (all 5 methods) |
DONE | GF2, KM, SGD, SMT, Fourier all report DMC |
fast.py (quick iteration) |
MISSING | Zero tracker integration (#15) |
| 33 experiment files | MISSING | None instantiate MemTracker for DMC |
scoreboard.tsv |
PARTIAL | DMC column exists, only 5 of 35 rows filled (#16) |
| DMC visualization / plotting | MISSING | Nothing exists (#18) |
| Cross-method DMC comparison | MISSING | CLAUDE.md shows ARD table, no DMC equivalent |
GitHub Issues (16 open, no new PRs)¶
Homework (due tomorrow)¶
| Issue | Title |
|---|---|
| #17 | DMC baseline sweep: measure all methods |
| #22 | DMC optimization experiment: beat baseline on at least one method |
Infrastructure (DMC shift)¶
| Issue | Title |
|---|---|
| #15 | Add tracker integration to fast.py |
| #16 | Backfill scoreboard.tsv with DMC values |
| #18 | DMC visualization and plotting |
| #6 | Compare DMC vs ARD vs real GPU joules |
Strategic¶
| Issue | Title |
|---|---|
| #19 | Prototype sparse parity as RL/eval environment |
| #20 | Add Public Domain license |
| #21 | Prep for Mar 30 meeting: Lukas Kaiser + Alec Radford visiting |
Existing (from before)¶
| Issue | Title |
|---|---|
| #4 | Push SGD under 10ms on sparse parity |
| #5 | Test agent loop on sparse sum and sparse AND |
| #7 | Add more task variations |
| #8 | Agent complexity budget |
| #9 | Modal integration for nanoGPT energy baseline |
| #13 | Agent compatibility layer |
| #14 | Agent notification bridge |
What's Due Tomorrow (Meeting #10)¶
From the Meeting #9 homework:
- Get agents to improve sparse parity using DMC (not ARD) as the energy proxy (#17, #22)
- Iterate on prompts and meta-approaches -- make it fast to go from "metric spec + problem spec" to a sequence of experiments
Our DMC metric is already implemented (task #1 in docs/tasks/INDEX.md is DONE). But we haven't run experiments optimizing DMC yet. The CacheTracker/MemTracker already tracks DMC alongside ARD (baseline: ARD 4,104 / DMC 300,298).
Task Lists¶
Homework Tasks (Due Monday)¶
See 007-homework-meeting10.md for full breakdown.
- Run DMC baseline sweep across top methods (#17)
- Run at least one DMC optimization experiment (#22)
- Prepare results summary for presentation
Infrastructure (This Week)¶
- Add tracker integration to fast.py (#15)
- Backfill scoreboard.tsv with DMC values (#16)
- Create DMC visualization scripts (#18)
Strategic (Before Mar 30)¶
- Prototype sparse parity as RL env (#19)
- Add Public Domain license (#20)
- Prep for Mar 30 meeting (#21)
- Compare DMC vs ARD rankings (#6)