Type inference + coercion
Strings that should be dates become dates. Numbers with commas become numbers. Boolean-like fields normalize to true/false. Per-column inferred type returned alongside the cleaned bytes.
Drop a messy CSV, get a clean structured table back. Snake-case headers, type inference + coercion, whitespace trim, dropped empty rows. Plus Smart Tables: per-column PII tags and outlier counts.
Norm is the on-ramp. The same pipeline every other DAVA product runs on — but exposed directly so you can clean a file and walk away. Sandboxed adapters, deterministic output, hash-chained to your audit log.
Each capability runs in the shared engine — the Norm pipeline, the Trust audit chain, the Decisioning mode toggle. Same substrate as the other four products.
Strings that should be dates become dates. Numbers with commas become numbers. Boolean-like fields normalize to true/false. Per-column inferred type returned alongside the cleaned bytes.
Every column comes back with a sensitivity hint — email, phone, SSN, IBAN, credit card (Luhn-checked), DOB, name-like. Surfaces in the dashboard so customers can mask before exporting.
IQR + z-score, configurable. Surfaced as a count so analysts can decide whether to investigate or coerce.
The cleaned bytes ship with a SHA-256 of input, the adapter version that ran, and a chain pointer. Every subsequent operation on this file is linked to its origin.
The Python SDK is the most mature. TypeScript follows the same shape. Both ship with strict types and async-first APIs.
Same auth as the rest of DAVA: bearer API key on Authorization: Bearer dava_live_… or session cookie + CSRF for browser flows.
We'll run your hardest dataset through DAVA Norm during a 5-day pilot. You keep the cleaned output and the evidence pack either way.