π―
Declarative DSL
Compose conversation blueprints like React components, turning dataset design into a predictable, readable workflow.
Build synthetic datasets
like you build UIs
β¨Declarative building blocksπ§ AI-powered variationsβοΈTypesafe pipelines
Like React, but for datasets β declarative and typesafe DSL for composing conversations and generating thousands of AI-powered training examples.
Compose reusable conversations with a declarative DSL and generate production-ready datasets in minutes.
Production feature set
Torque borrows from the best product-engineering playbooks: versioned blueprints, strict typing, reproducible runs, and streaming observability. No YAML, no bespoke tools β just code.
π―
Compose conversation blueprints like React components, turning dataset design into a predictable, readable workflow.
π
Harness Zod-powered typing with end-to-end inference so every AI training workflow stays safe, reliable, and maintainable.
π
Generate with OpenAI, Anthropic, DeepSeek, vLLM, LLaMA.cpp, or any model supported by your composable data pipeline.
π€
Blend handcrafted prompts with AI dataset generation to produce realistic multi-message conversations at scale.
π°
Reuse context across runs to minimize token spend while keeping dataset scaling fast, efficient, and reproducible.
β‘
Stream progress in a beautiful CLI with concurrent workers, deterministic seeds, and instant feedback loops.
The challenge
Every team building AI products hits the same roadblocks when trying to create high-quality training data at scale.
π
Sometimes you don't have enough real data
β±οΈ
Maintaining quality and consistency across thousands of examples is extremely time consuming
π§
Tool calling patterns require intricate message sequences and are errorβprone
π
Generating different conversation flows means rewriting everything or creating various hard to maintain scripts
π²
Designing generators that are random yet reproducible is surprisingly complex
π§©
Getting AI to understand complex composition scenarios (nested variations, conditional flows) takes significant prompt engineering time
The gaps in today's tooling
Torque is built for product teams that expect the same polish from their AI training data as they do from their production apps. First, let's take a quick look at the pain that keeps teams from shipping.
How Torque helps
Torque replaces sprawling prompts and brittle scripts with a composable TypeScript toolkit. Every phase is automated, typed, and observable.
βTorque is the missing layer between your product engineers and the models they rely on.β
01Stage 1
Describe each conversation as a reusable component. Torque keeps prompts, metadata, and policy checks together in strongly typed modules.
Ship new flows faster with versionable building blocks, not ad-hoc scripts.
02Stage 2
Spin up tens of workers, hydrate examples with AI, and cache deterministic seeds so teams get repeatable datasets without wasting tokens.
Every run logs audit trails, token spend, and policy results β plug it straight into CI.
03Stage 3
Replay production traffic through model providers, enforce quality gates, and ship regressions straight into Slack or GitHub checks.
Close the loop with built-in evaluations and targeted retries when policies fail.
Join developers building the future of LLM training data