Cluster1 sources· last seen 5h ago· first seen 5h ago
Used over a million tokens in three separate sessions to test Qwen 3.6 35b (new Multi-token Prediction version)
In my opinion, MTP models are 100% game changer for local LLMs. In terms of speed, I was getting around 1.5x the tok/sec of previous tests. The project was a test - building a full iterative step-by-step pygame; a small mystery dungeon-style game. At first I set 100-200k context and raised it to
Lead: r/LocalLLaMABigness: 24usedmilliontokensthreeseparate
📡 Coverage
10
1 news source
🟠 Hacker News
0
🔴 Reddit
55
49 upvotes across 1 sub
📈 Google Trends
0
Full methodology: How scoring works
Receipts (all sources)
Used over a million tokens in three separate sessions to test Qwen 3.6 35b (new Multi-token Prediction version)
REDDIT · r/LocalLLaMA · 5h ago · ⬆ 49 · 💬 32
score 119
In my opinion, MTP models are 100% game changer for local LLMs. In terms of speed, I was getting around 1.5x the tok/sec of previous tests. The project was a test - building a full iterative step-by-step pygame; a small mystery dungeon-style game. At first I set 100-200k context and raised it to