micro-kiki
35-domain expert model built on Qwen3.5-35B-A3B (MoE, 256 experts, 3B active/token) with LoRA adapters and a cognitive layer (memory palace + negotiator + anti-bias).
Model Description
micro-kiki is a multi-domain language model designed for technical applications spanning electronics, firmware, CAD, manufacturing, and general-purpose conversation. It uses a router-based architecture that selects up to 4 domain-specific LoRA stacks per request.
| Property | Value |
|---|---|
| Base model | Qwen3.5-35B-A3B |
| Architecture | MoE (256 experts, 3B active/token) |
| Adapter | LoRA rank 16 (q/k/v/o projections) |
| Domains | 35 |
| Max active stacks | 4 |
| Context length | 262,144 tokens |
| Quantization | Q4_K_M (inference), BF16 (training) |
| License | Apache 2.0 |
Architecture
+-------------------+
| Domain Router |
| (classifier, top4)|
+--------+----------+
|
+----------+--------+--------+----------+
| | | |
+----v----+ +---v---+ +----v----+ +---v---+
| Stack 1 | |Stack 2| ... |Stack 34 | |Stack35|
| chat-fr | |python | |ml-train | |securi.|
+---------+ +-------+ +---------+ +-------+
| | | |
+----------+--------+--------+----------+
|
+--------v----------+
| Negotiator |
| CAMP + Catfish |
+--------+----------+
|
+--------v----------+
| Anti-Bias |
| KnowBias + RBD |
+--------+----------+
|
+--------v----------+
| Aeon Memory |
| Atlas + Trace |
+-------------------+
Intended Use
- French/English conversational AI with domain expertise
- Code generation (Python, C/C++, Rust, TypeScript, embedded firmware)
- Electronics design (KiCad DSL, schematic review, component selection, SPICE)
- Manufacturing (process optimization, quality control)
- Multi-domain routing with cognitive arbitration
Limitations
- Not designed for medical, legal, or financial advice
- Optimized for technical domains; general knowledge may be weaker than base model
- Requires Q4_K_M or higher quantization; quality degrades below Q4
- Maximum 4 concurrent LoRA stacks; performance varies with stack combinations
- Memory (Aeon) requires external backends (Qdrant/Neo4j) for production use
Training Data โ V3 (489K examples, 35 domains)
Sources
| Source | Examples | Description |
|---|---|---|
| Claude CLI sessions | 50,116 | Real user-tool interactions extracted from 5 machines (GrosMac, kxkm-ai, Studio, Tower, CILS) |
| Codex/Copilot sessions | 2,529 | OpenAI Codex + GitHub Copilot sessions extracted from 4 machines |
| HuggingFace datasets | 364,045 | 19 open datasets (see below) |
| Opus teacher distillation | โ | chat-fr, reasoning domains |
| Original curated | โ | 32 domain seed datasets |
HuggingFace Datasets
| Dataset | Examples | License |
|---|---|---|
| CodeFeedback-Filtered-Instruction | 157,000 | Apache 2.0 |
| French-Alpaca-Instruct-110K | 110,000 | Apache 2.0 |
| Electronics StackExchange | 95,000 | CC-BY-SA-3.0 |
| CJJones/LLM_EE_Educational_Synthetic_Dialog | 50,000 | CC-BY-NC-SA-4.0 |
| MuratKomurcu/stm32-hal-dataset | 29,700 | MIT |
| redcathode/thingiverse-openscad | 7,400 | โ |
| ThomasTheMaker/OpenSCAD | 4,900 | โ |
| STEM-AI-mtl/Electrical-engineering | 1,100 | โ |
| JITX open-components-database | 151 | โ |
| Vrindarani/netlistgen | 106 | โ |
35 Domains
| Group | Domains |
|---|---|
| Conversation | chat-fr, reasoning |
| Code | python, typescript, cpp, rust, html-css, shell, sql, yaml-json, lua-upy |
| Infrastructure | docker, devops, llm-orch, llm-ops (NEW), ml-training (NEW) |
| Electronics | kicad-dsl, kicad-pcb, spice, electronics, components (NEW), power, emc, dsp |
| Hardware | embedded, stm32, iot, platformio |
| CAD | freecad |
| Web | web-frontend, web-backend |
| Other | music-audio, math, security |
Changes from V2: 3 new domains (components, llm-ops, ml-training). spice-sim merged into spice. stm32 is a sub-category of embedded.
New Domain: components
57K Q&A about electronic component specs, datasheets, sourcing, BOM, and cross-reference. Sources: Electronics StackExchange (filtered by component tags) + JITX open-components-database.
Training โ V3
| Property | Value |
|---|---|
| Base model | Qwen3.5-4B |
| Adapter | MoE-LoRA: 4 experts/projection, rank 16, top-2 routing |
| Null-space projection | ENABLED (prevents catastrophic forgetting between stacks) |
| Curriculum | Sequential, 35 stacks trained in order |
| Platform (MLX) | Mac Studio M3 Ultra 512 GB |
| Platform (CUDA) | kxkm-ai RTX 4090 24 GB |
Evaluation
| Metric | Value |
|---|---|
| Router accuracy (35-class) | [PENDING] |
| Forgetting check (angle) | [PENDING] |
| Perplexity (base) | [PENDING] |
| Perplexity (debiased) | [PENDING] |
| Aeon recall@1 | [PENDING] |
| Aeon recall@5 | [PENDING] |
| Aeon recall@10 | [PENDING] |
| Anti-bias flag rate | [PENDING] |
| Average inference latency | [PENDING] |
Hardware Requirements
| Setup | RAM/VRAM | Use |
|---|---|---|
| Mac Studio M3 Ultra | 512 GB unified | Training (BF16 LoRA) + serving (MLX) |
| RTX 4090 | 24 GB VRAM | Q4 inference (vLLM) |
| Apple Silicon 32 GB+ | 32 GB unified | Q4_K_M inference (MLX/llama.cpp) |
Citation
@misc{micro-kiki-2026,
title={micro-kiki: Multi-Domain Expert Model with Cognitive Layer},
author={L'Electron Rare},
year={2026},
url={https://huggingface.co/electron-rare/micro-kiki}
}
Related Projects & Ecosystem
micro-kiki-v3 is one component of the FineFab platform built by L'รlectron Rare โ a local-first, multi-machine AI-native manufacturing and electronics platform.
| Role | Project | Description |
|---|---|---|
| Training toolkit | L-electron-Rare/KIKI-Mac_tunner | MLX fine-tuning toolkit (Mac Studio) โ Opus reasoning distilled into Mistral Large 123B |
| Fine-tuning pipeline | L-electron-Rare/KIKI-models-tuning | FineFab fine-tuning pipeline โ training, evaluation, registry (Unsloth, LoRA) |
| Methodology | electron-rare/Kill_LIFE | Spec-first agentic methodology for embedded systems โ BMAD agents, gates, evidence packs |
| Orchestration | electron-rare/mascarade | Multi-machine agentic LLM orchestration โ P2P mesh, 8 providers, RAG pipeline |
| AI backend | L-electron-Rare/life-core | FineFab AI backend โ LLM router, RAG, caching, orchestration |
| CAD assistant | electron-rare/KiC-AI | AI-powered PCB design assistant for KiCad |
See the full org at github.com/L-electron-Rare โ 13 public repos covering platform, hardware, firmware, CAD, and ML.
Infrastructure: the 50K+ Claude CLI examples in the training dataset were captured on our 5-node P2P mesh โ GrosMac (Apple M5), Tower (28 threads), CILS (i7), KXKM-AI (RTX 4090), VM bootstrap. Ed25519 auth, DHT discovery.
- Downloads last month
- 136
4-bit