← Back
Research

Research

Papers, benchmarks, evals, and long-form technical posts from the neww.ai team. We publish what we learn — successes and failures.

Paper2026-04

Engine Registry: a pluggable architecture for multi-product AI platforms

neww.ai Research

We describe the engine-registry pattern used at neww.ai to unify 53+ AI engines behind a single routing layer, and present empirical results on end-to-end latency and quality across 30 downstream products.

Read
Post2026-03

Retrieval at scale: lessons from a multi-product RAG pipeline

neww.ai Research

What we learned running RAG across millions of documents — chunking, hybrid retrieval, re-ranker cost-benefit, and eval harness design.

Read
Benchmark2026-03

The neww.ai routing benchmark

neww.ai Research

A public benchmark evaluating model selection policies across cost, latency, and quality for production tool-use workloads.

Eval2026-02

Prompt-injection resistance in multi-agent systems

neww.ai Research

An evaluation of prompt-injection defenses when agents chain across untrusted inputs and tools, with recommendations for practitioners.

Work with us on research

We partner with academic groups and independent researchers on evals, benchmarks, and open problems.

research@neww.ai