Materialized Intelligence API Documentation#
Materialized Intelligence is an API for large-scale, latency-insensitive inference. We’re currently in beta with a focus on batch inference workloads, enabling data-heavy use cases. This includes large-scale data processing, augmentation, and generation.
Our first product is a hosted batch inference API that enables developers to run workloads of arbitrary size easily and inexpensively. You can expect up to 80-90% decrease in costs relative to online inference solutions, extreme throughput capabilities, and white-glove support. If you have a use case that may benefit from our tools, please reach out to us at team@materialized.dev to discuss your requirements.
Contents
- Introduction
- Installation
- Quickstart
- Examples
- Batch API Reference
- Stage API Reference
- Python SDK and CLI Reference
- Available Models
- Text-to-Text Models
- Reasoning Models (text-to-text)
- Embedding Models
- Custom Models