Doubleword logo black
Product
Products
Doubleword API
NEW
Inference built for scale
Doubleword Inference Stack
High performance inference stack
Use Cases
Async Agents
Long running background agents
Synthetic Data Generation
Generate high volumes of data for fine- tuning
Data Processing
Apply intelligence to large volumes of data
Resources
Documentation
Technical docs and API reference
Workbooks
Ready-to-run examples
Seen in the Wild
Community content and projects
Resource Centre
All our blogs and guides
Technical Blog
Our blog on building inference systems
Al Dictionary
Key Al terms explained
Savings Calculator
See how much you save with Doubleword
Solutions
By Deployment Option
On-premiseCloudHybrid
By Team
AI, ML & Data SciencePlatform, DevOps & ITCompliance & Cyber
Pricing
Docs
Pricing
Get started - Free
Get started - Free
Resources
/
Blog
/
20/2 Weekly Update: New Qwen Models, GPT-OSS 20B & Webhooks
February 20, 2026

20/2 Weekly Update: New Qwen Models, GPT-OSS 20B & Webhooks

Meryem Arik
Share:
https://doubleword.ai/resources/doubleword-weekly-update---20-2
Copied
To Webinar
•

It’s been really great seeing what people are starting to build with Doubleword - and a lot of the feedback we’ve received so far has directly shaped what we’re shipping next.

This is a quick update on a few things we’ve just released.

New model support: Qwen3-14B, Qwen3.5 & GPT-OSS 20B

We’ve added support for two new Qwen models and OpenAI’s GPT-OSS 20B, giving teams access to more frontier-level intelligence models.

Qwen3.5-397B-A17B
Pricing: $0.3 / $1.80 and $0.15 / $1.2
Qwen3.5 is a powerful step up for multimodal background agents and OpenClaw-style workflows. It handles richer reasoning and mixed-input tasks well, which makes it a strong choice for asynchronous agents that need to process context, images, or longer chains of work without requiring real-time responses.

One thing that’s been especially interesting here is the economics. Running Qwen3.5 via async inference dramatically changes cost profiles for large async workloads - while maintaining the same model quality.

For equivalent workloads, this can be:

  • Up to 91% cheaper vs Anthropic
  • Up to 83% cheaper vs OpenAI

Async inference with Doubleword changes the economics of agentic and asynchronous workloads - making frontier-level intelligence far more accessible.

This is a big part of why we believe async inference is becoming such an important design choice for production systems.

Qwen3-14B
Pricing: $0.03 / $0.3 and $0.02 / $0.2
This model has been performing extremely well for classification tasks, lightweight reasoning, and smaller sub-agent workflows. If you’re running background processes where speed and cost efficiency matter more than frontier-level reasoning, this is a strong default option. It’s especially useful for workloads that involve large volumes of structured decisions or simple task routing.

GPT-OSS 20B
Pricing: $0.03 / $0.2 and $0.02 / $0.15
This model designed for powerful reasoning, agentic tasks, and versatile developer use cases which increases the suite of models within Doubleword suited to agentic workloads. 

Our goal has been to offer best-in-class models at every meaningful price point - so builders can choose the right capability level for each part of their system rather than overpaying across the board.

Let us know which models you'd like to see next - email support@doubleword.ai for your suggestions.

Webhooks + notifications

We’ve also shipped webhooks and notifications.

Until now, many users were polling to check whether a batch had completed. With webhooks, you can now provide an endpoint and we’ll automatically POST updates when a batch completes or fails.

In practice, this means:

  • No more polling loops
  • Cleaner async workflows
  • Easier integration into existing pipelines
  • Simpler orchestration for agent or background systems

Try it out

If you haven’t already, try running something with the new models and see how they perform on your real workloads. We’d love to hear what works well, what doesn’t, and which models you’d like to see us support next.

Footnotes

Table of contents:

Heading 2
Heading 3
Heading 4
Heading 5
Heading 6
"
Learn more about self-hosted AI Inference
Subscribe to our newsletter
Thanks you for subscription!
Oops! Something went wrong while submitting the form.

Stop overpaying for inference.

Teams use Doubleword to run low-cost, large-scale inference pipelines for async jobs.
‍
Free credits available to get started.

Get started - Free
Doubleword logo black
AI Inference, Built for Scale.
Products
Doubleword APIDoubleword Inference Stack
Use Cases
Async AgentsSynthetic Data GenerationData Processing
Resources
Seen in the WildDocumentationPricingAsync Pipeline BuilderResource CentreTechnical BlogAI Dictionary
Company
AboutPrivacy PolicyTerms of ServiceData Usage Policy
Careers
Hiring!
Contact
© 2026 Doubleword. All rights reserved.
We use cookies to ensure you get the best experience on our website.
Accept
Deny