Synthetic data with receipts.
Four production-grade products. One cryptographically-sealed evidence chain. From a one-line prompt to a fully-synthesised, provenance-stamped dataset — and an autonomous agent (or plain English) to run the whole pipeline.
- the flagship engine fidelity
- 95.69 %
- ADS tools
- 43
- ADS modules
- 48
- Connectors
- 14
| data_clean | drop_threshold=0.05 | ~5c |
| encode_categorical | method=onehot | ~3c |
| train_predictive | model=randomforest | ~12c |
| shap_explain | sample=200 | ~3c |
Four products. One contract. One evidence chain.
Pick the entry point that matches the data you have today — the underlying evidence pipeline is identical, so you move between them without changing your downstream tooling.
Six stages from prompt to sealed bundle.
Every dataset shipped from any product on this pillar passes through the same six stages. The bundle that lands in your bucket can be verified offline by anyone with the open-source evidence verifier CLI.
the sealed contract
Schema, constraints, intent and seed sealed into a JSON artefact before any data is generated.
Engine run
Mock, Synthesize, or ADS-driven pipeline executes against the sealed contract; per-step I/O recorded.
Quality gates
K-S, Pearson, χ², constraint satisfaction, per-column drift checked. Fail-closed: regression aborts.
Cryptographic chain
Each step's inputs and outputs hashed and chained. Tamper-evident, verifiable offline.
Evidence bundle
Signed .tar.zst with contract, run-log, quality report, artefact manifest, engine SBOM.
Tenant isolation
Per-tenant Fernet keys, per-tenant artefact prefixes, per-tenant evidence keys.
Numbers we’ll defend.
Validated under an independent third-party QA harness. Full reproducibility certificate published on the /verify page.
From data_clean to shap_explain — same engines callable from chat or SDK.
Snowflake, BigQuery, Databricks, Postgres, S3, GCS, Azure Blob — all Fernet-vaulted.
Connector creds auto-hoisted to the encrypted vault, never written to logs.
Bring a CSV. We’ll show you the evidence bundle.
30-minute working session: you upload (or we mock) a representative dataset, we run it through Synthesize and the Autonomous Data Scientist, and you keep the signed evidence bundle and quality report.
| data_clean | drop_threshold=0.05 | ~5c |
| encode_categorical | method=onehot | ~3c |
| train_predictive | model=randomforest | ~12c |
| shap_explain | sample=200 | ~3c |