It’s been great seeing more teams run real workloads on Doubleword over the past week.
Last week, we added Qwen3.5-397B-A17B, Qwen3-14B, and GPT-OSS-20B to the platform - and lots of users have already been putting them to work across agent workflows, extraction pipelines, and async evaluation jobs.
We’ve seen very strong early results, particularly:
- Qwen3.5-397B-A17B for multimodal background agents and OpenClaw-style systems
- Qwen3-14B for high-volume classification, routing, and lightweight reasoning workloads
It’s been exciting to see how quickly these models are being tested in production-style pipelines.
This week, we’re expanding that lineup further.
New model support: Qwen3.5-35B-A3B
Pricing: $0.07 / $0.30 (High Priority) and $0.05 / $0.20 (Standard Priority)
Qwen3.5-35B-A3B is a high-intelligence, mid-sized model that hits a very compelling price/performance point for async workloads.
In Qwen’s published benchmarks, this model outperformed GPT-5-mini, GPT-OSS-120B, and Claude Sonnet 4.5.
Even more interesting: it delivers higher quality than our previous largest model, Qwen3-235B - at a fraction of the cost.
In practical terms, this means:
- Stronger reasoning
- More robust outputs
- Better multimodal performance
- High efficiency for async and batch jobs
Qwen3.5 represents a meaningful architectural step forward - combining improved reasoning, multimodal capability, and efficiency in a way that’s particularly well suited to large-scale background agents and evaluation workflows.
🥳 I’m also very proud to say Doubleword is offering the best pricing for these models on the market.
Should you migrate?
If you’re currently using Qwen3-30B or Qwen3-235B, we’d strongly recommend testing:
- Qwen3.5-35B-A3B
- Qwen3.5-397B-A17B
Both offer higher intelligence and improved reasoning quality for the price - particularly for:
- Multi-step agents
- Extraction pipelines
- Large-scale evals
- Background async processing
Qwen3-30B and Qwen3-235B remain supported. But for most users, the new Qwen3.5 series should now be the better default starting point.
Try it out
As always, the best evaluation is against your own workload.
Benchmark Qwen3.5-35B-A3B against your current setup and see how it performs. If you notice meaningful differences - positive or negative - we’d love to hear what you see.
We’ll continue expanding model coverage across performance tiers while keeping pricing aggressively competitive for async and batch inference.
More to come next week.


