// RESEARCH

Engineered first.
Published second.

We publish what we build. Three research papers on arXiv, fifteen filed patents (ten at PCT stage), and 47 internally documented patentable innovations across the platform. Research as accountability — not marketing.

arXiv papers
3
PCT filed
10
Filed
15
Innovation pipeline
47

47 innovations.
Filed and rising.

We undertook a structured IP-mining exercise in 2024. Fifteen patents filed. Ten already at PCT phase. The remaining are within the standard 12-month PCT filing window. Filing cadence is accelerating.

// PATENT PIPELINE

From idea → filing → PCT.

47PATENTABLE INNOVATIONSidentified internally (2024)15PATENTS FILEDsubmitted to patent office10PCT PHASEinternational filing in progress
PCT 202441088473 (Optimized Real-Time NLP on Edge Devices) is the foundation patent in this portfolio.
202441088473

Optimized Real-Time NLP on Edge Devices in Resource-Constrained Environments

PCT filed Nov 2025

Tensulator + Tensor Codec — Dynamic Neural Compression in ExSLerate V2

PCT filed · Indian Patent Office

Hybrid KV-Cache Reuse Architecture for LLM Inference

Provisional

Spatial Programming Compiler for AI Accelerators

Provisional

HaluMon — Hallucination Detection via Multi-Metric Scoring

Provisional
// IP COVERAGE BY LAYER

Every layer of the stack protected.

The 4 representative patents above span the full stack — from silicon-aware compiler design (L01) to hallucination detection at the platform layer (L03). The remaining 11 filed (and 32 in pipeline) distribute across all five layers.

DATA FLOWL05ApplicationsIRA · LINGO · IDPIN PIPELINEL04Foundation ModelsSHAKTI · NEXONS1NAMED — PCT FILEDL03PlatformHALUMON · LINGOFORGE1NAMED — PROVISIONALL02EdgeMatrixRUNTIME · KV-CACHE1NAMED — PROVISIONALL01Krsna SoCSPATIAL COMPILER · IP1NAMED — PROVISIONAL
Counts shown reflect the 4 named patents above. Full portfolio: 47 innovations identified · 15 filed · 10 PCT — distributed across all five layers.
// BENCHMARK DISCLOSURE

Reproducible. By design.

Every performance number we publish includes the model version, hardware, batch size, sequence length, and methodology. Comparisons are run on public datasets with peer-reviewed harnesses (DeepEval, lm-evaluation-harness, OpenCompass).

Hardware reference

NVIDIA A100 (80 GB), L40s, H100 · AMD MI300 · Intel Gaudi · Krsna simulator

Frameworks compared

vLLM 0.10.2, TensorRT-LLM 1.0.0, SGLang, FlashInfer, llama.cpp

Benchmark suites

MMLU, GSM8K, MATH, HumanEval, MedQA, MMMU, DocVQA, OCRBench, ChartQA

Methodology

Q4_KM quantization · 5-shot for aggregate · 0-shot for chain-of-thought · 3-run average

// LET'S BUILD

Want the full benchmarks?