Compute ≠ Power. The language AI writes, not humans.
YONG (永年语言) is a declarative programming language designed for AI to generate, not humans to learn. Same 30 tokens of YONG replace 500 tokens of Python. The compiler materializes the rest.
What AI writes today What AI should write
┌─────────────────────────────┐ ┌─────────────────────────────┐
│ from flask import Flask... │ │ struct Todo { │
│ from sqlalchemy import... │ │ title: string │
│ app = Flask(__name__) │ │ done: bool │
│ Base = declarative_base() │ │ } │
│ class Todo(Base): │ │ @api(POST, "/todos") │
│ __tablename__ = 'todos' │ │ fn create(req) -> Todo { │
│ id = Column(Integer...) │ │ return req |> save; │
│ title = Column(String...) │ │ } │
│ done = Column(Boolean...) │ │ │
│ ... 40 more lines │ │ │
├─────────────────────────────┤ ├─────────────────────────────┤
│ ~500 tokens │ │ ~30 tokens │
│ ~$15 per 1000 files │ │ ~$1.50 per 1000 files │
└─────────────────────────────┘ └─────────────────────────────┘
10× fewer tokens = 10× less GPU compute = 10× less power.
CUDA's moat isn't hardware — it's ecosystem lock-in. But there's a bigger waste hiding in plain sight:
The language AI writes in IS a form of compute cost.
Every import, every ORM config, every route handler that AI generates is a token that burns GPU power. Billions of tokens, every day, across every AI coding assistant on the planet.
YONG eliminates the waste:
| Python (what AI writes today) | YONG (what AI should write) | |
|---|---|---|
| Todo App | ~500 tokens | ~30 tokens |
| User Auth API | ~800 tokens | ~50 tokens |
| SNN Chip Design | ~900 lines Verilog | ~30 lines YONG |
| AI generation cost | 1× | 0.1× |
And here's the double kill: YONG also compiles to SNN neuromorphic hardware at 28 pJ/spike — 100,000× more efficient than GPU inference.
One language. Two kills. Generation cost down 10×. Execution cost down 100,000×.
Humans don't write YONG. Humans don't learn YONG.
Today's pipeline:
User → "Build a todo app" → AI → 500 tokens Python → interpreter → app
YONG pipeline:
User → "Build a todo app" → AI → 30 tokens YONG → compiler → app/silicon
The compiler infers everything from declarations: database schema, HTTP routing, middleware, serialization, error handling — or LIF neurons, STDP learning, WTA competition, BRAM pipelines.
Declare intent. Compiler materializes.
┌─────────────────────────────────────┐
│ Human (natural language) │
└──────────────┬──────────────────────┘
│
▼
┌─────────────────────────────────────┐
│ VGO Brain 2.0 (109M param SNN) │
│ Natural Language → YONG │
└──────────────┬──────────────────────┘
│
▼
┌─────────────────────────────────────┐
│ .yong source file │
│ (30 tokens, declarative intent) │
└──────────────┬──────────────────────┘
│
┌────────────────────┼────────────────────┐
▼ ▼ ▼
┌────────────┐ ┌──────────────┐ ┌──────────────┐
│ Lexer │ │ Parser │ │ IR Gen │
│ (tokens) │ ──▶ │ (AST) │ ──▶ │ (typed IR) │
└────────────┘ └──────────────┘ └──────┬───────┘
│
┌──────────────────┼──────────────────┐
▼ ▼
┌─────────────────────┐ ┌─────────────────────┐
│ Web Backend │ │ RTL Backend │
│ HTML+CSS+JS+API+DB │ │ Verilog → Yosys │
│ (App Dialect) │ │ → FPGA/ASIC │
└─────────────────────┘ │ (Hardware Dialect) │
└─────────────────────┘
5-stage pipeline: Lexer → Parser → AST → Typed IR → Backend-specific code generation.
Same .yong file can target web applications or synthesizable silicon.
@db(table="todos")
struct Todo {
id: uid
text: string
done: bool
}
@api(GET, "/todos")
fn list_todos() -> list[Todo] {
return db.todos.all()
}
@api(POST, "/todos")
fn add_todo(text: string) -> Todo {
return db.todos.create({ text, done: false })
}
@app(route="/")
component TodoApp {
state todos = query(list_todos)
view {
Header("My Todos")
Input(placeholder="Add...", on_enter=add_todo)
List(todos) -> |todo| {
Row { Checkbox(todo.done) Text(todo.text) }
}
}
}
Compiles to: HTML + CSS + JS + REST API + Database schema + ORM
network MNIST {
layer input(784);
layer hidden(400, type=lif, threshold=300, leak=1);
layer output(10, type=lif, threshold=200, leak=2);
connect input -> hidden with stdp(ltp=5, ltd=2);
connect hidden -> output with stdp(ltp=10, ltd=4);
config {
weight_storage: bram
wta_mode: enable
homeostasis: enable
target: zynq7020
}
}
Compiles to: 47KB synthesizable Verilog RTL (verified on Yosys + iCE40 FPGA)
| Feature | Why it matters for AI |
|---|---|
| No imports | Compiler infers dependencies from declarations. Zero wasted tokens. |
| Decorators as directives | @api(POST, "/todos") replaces 20 lines of routing + middleware + serialization. |
| Pipe operators | `data |
| No boilerplate | No ORM config, no app factory, no session management. Declared once, inferred everywhere. |
| Dual compilation | Same syntax targets web apps OR silicon. AI doesn't need to know the target. |
| Principle | Description |
|---|---|
| AI-First | Designed for AI to generate, not humans to type |
| Declarative | Write what, not how |
| Minimum Tokens | Maximum semantics in minimum surface area |
| Unit Safety | Time<ms> + Energy<pJ> → compile error E202 |
| Bit-Accurate | Same .yong → same behavior on every backend |
| Extensible | Frozen core, open everything else |
| Metric | NVIDIA GPU | VGO SNN (YONG-compiled) |
|---|---|---|
| Power per inference | ~2,700,000 pJ | ~28 pJ/spike |
| Efficiency ratio | 1× | ~100,000× |
| Lines of code | 900+ (CUDA/Verilog) | 30 (YONG) |
| FPGA bitstream | N/A | 132 KB |
📊 Full methodology, hardware specs, and reproduction scripts: benchmarks/README.md
We sent YONG code to 5 major AI models — none had seen YONG before.
Round 1 — Raw code, zero context: 5/5 understood instantly.
Round 2 — Full README review:
| AI | Key Quote |
|---|---|
| Claude | "The token economics argument is spot-on. You're competing for AI token budgets — that's a different game entirely." |
| DeepSeek | "A brilliant, necessary, high-risk/high-reward thought experiment made real." |
| Gemini | "YONG seems like a perfect Intermediate Representation (IR) — human describes, AI generates YONG, then it compiles to Rust, Go, or TypeScript." |
| ChatGPT 5.2 | "README isn't just vapor — there are concrete status claims. That's the right kind of credibility signal." |
| Grok | "The bottleneck is no longer transistors or watts — it's tokens and unnecessary syntax." |
These AIs are YONG's target users. Their instant comprehension validates the core design thesis.
| Component | Status |
|---|---|
| Language Specification v4.3 | ✅ Complete (v4.2 core + v4.3 addendum) |
| Compiler Specification v1.0 | ✅ Complete |
| Parser (Lexer → AST → IR) | ✅ Working (822 lines) |
| Native Engine (.yong → GUI) | ✅ Working |
| Verilog Backend (.yong → RTL) | ✅ Working (47KB output, Yosys verified) |
| VGO Brain 2.0 (NL → YONG) | ✅ Working (109M param SNN) |
| FPGA Synthesis (iCE40) | ✅ Verified (132KB bitstream) |
| VSCode Extension | ✅ Syntax highlighting |
| Fine-tuning Dataset | 🔄 16 seed pairs (target: 1000+) |
| Conformance Tests | 🔄 31 test cases defined |
| Document | English | 中文 |
|---|---|---|
| Language Specification | spec/language-spec.md | spec-zh/YONG语言规范.md |
| Spec Addendum v4.3 | spec/language-spec-addendum-v4.3.md | spec-zh/YONG语言规范补充v4.3.md |
| Compiler Specification | spec/compiler-spec.md | spec-zh/YONG编译器规范.md |
| Benchmark Methodology | benchmarks/README.md | — |
See the examples/ directory:
hello.yong— Hello Worldtodo-app.yong— Full-stack Todo App (30 tokens)snn-mnist.yong— MNIST classifier SNN chipauth-api.yong— Authenticated API with RBACerror-handling.yong— Railway-oriented error handlingblog-engine.yong— Full CMS with posts, comments, tags, auth, and searchecommerce-api.yong— E-commerce with state machines, transactions, and webhookssnn-speech.yong— SNN speech recognition with reservoir computing
- VSCode Extension: vscode-yong/ — Syntax highlighting for
.yongfiles - Fine-tuning Dataset: datasets/ — NL→YONG training pairs for AI models
- Conformance Tests: tests/ — 31 golden test cases
YONG is in its early stages. We welcome contributions in:
- 🧪 Compiler implementation — Help build the 5-stage pipeline
- 📝 Language design feedback — File issues on the spec
- 🔌 Backend plugins — Implement new compilation targets
- 📚 Documentation — Improve specs, add examples, translate
Robert Hu — Chongqing, China 🇨🇳
📧 roberthxr@qq.com
Declare intent. Compiler materializes. | 声明意图,编译器物化。
Compute ≠ Power. | 算力不等于电力。
The writer is AI, not human. | 写代码的是 AI,不是人。