Each architecture family wants different things from CORE — KV cache for transformers, scan kernels for SSMs, recurrent loops for RWKV, ODE solvers for LFMs, conv passes for CNNs. CORE handles the dispatch end to end.
Transformers
PRODUCTIONLlama · Qwen · Mistral · Shakti · GPT-class · DeepSeek
The dominant LLM architecture today. ISE supports the full attention-mechanism family with hybrid KV-cache reuse — the optimization that drives EdgeMatrix's +73% throughput lift on L40s.
Vision-Language Models (VLM)
PRODUCTIONShakti-VLM · Qwen2-VL · LLaVA · MiniCPM-V
Multimodal architectures combining vision encoders with transformer decoders. ISE handles VLM workloads natively, where vLLM and TensorRT-LLM still require workarounds.
State Space Models
PRODUCTIONMamba · Mamba-2 · Jamba (hybrid)
Linear-time alternatives to attention. SambaASR — our Mamba-based speech model — runs natively on ISE with 4× throughput vs Whisper. Jamba (Mamba+Transformer hybrid) handles long-context workloads.
Linear Attention (RWKV)
PRODUCTIONRWKV-4 · RWKV-5 (Eagle) · RWKV-6 (Finch) · RWKV-7 (Goose)
RNN-style models with constant memory and linear time complexity. ISE supports the RWKV family for ultra-long-context and on-device workloads where transformer memory cost is prohibitive.
Liquid Foundation Models
BETALFM-1B · LFM-3B · LFM-40B (Liquid AI)
Continuous-time neural network architectures. ISE has end-to-end support for the LFM family — important for time-series, control, and reasoning workloads where Liquid AI is pushing state-of-the-art.
Convolutional Networks (CNN)
PRODUCTIONResNet · EfficientNet · YOLO · MobileNet · ConvNeXt
Classical vision architectures. Most AI chips serve EITHER CNNs OR LLMs — ISE serves both on the same runtime. Critical for device-makers who need vision today and LLMs tomorrow without replacing silicon.
Mixture of Experts (MoE)
PRODUCTIONMixtral 8×7B · DeepSeek-V3 · DBRX
Sparse routing architectures where only a fraction of parameters activate per token. ISE handles expert routing and KV-cache across MoE layouts without bespoke per-model engineering.
Diffusion & Image Generation
SUPPORTEDStable Diffusion · SDXL · Flux · custom UNet
Iterative denoising architectures for image and video generation. ISE's runtime supports diffusion, but customer deployments are still pending. Architectural note: preprocessing pipelines (T5-XXL, CLIP encoders) carry large memory footprints — diffusion is well-suited to data-center silicon (NVIDIA, AMD, Intel) but not to true-edge processors where the preprocessor alone exceeds the memory budget.