DeepReasoning
StaleDescription
A high-performance LLM inference API and Chat UI that integrates DeepSeek R1's chain-of-thought reasoning traces with Anthropic Claude models.
A high-performance LLM inference API and Chat UI that integrates DeepSeek R1's chain-of-thought reasoning traces with Anthropic Claude models.
Fast, flexible LLM inference engine built in Rust — supports multiple model architectures and quantization schemes for high-performance local LLM deployment.
A high-throughput and memory-efficient inference and serving engine for LLMs, featuring PagedAttention, continuous batching, and optimized KV cache management for production deployments.
Open Vision Agents by Stream. Build voice and vision agents quickly with any model or video provider, using Stream's edge network for ultra-low latency realtime interactions.
A deep research agent framework optimized for complex research and prediction tasks, with MiroThinker-1.7 and MiroThinker-H1 models achieving 74.0 and 88.2 on BrowseComp benchmark, supporting multi-step reasoning and information retrieval.