Blog
-
Serving gpt-oss-120b at 5.8 ms TPOT with two RNGD cards: compiler optimizations in practice
Our Viewpoints -
Introducing Furiosa NXT RNGD Server: Efficient AI inference at data center scale
News -
Furiosa SDK 2025.3 boosts RNGD performance with multichip scaling and more
Technical Updates -
FuriosaAI and OpenAI showcase the future of sustainable enterprise AI
News -
Furiosa presents papers about improving AI performance at ICML 2025 and ACL 2025
News -
Announcing FuriosaAI’s $125M Series C bridge funding to scale sustainable AI compute
News -
LG AI Research taps FuriosaAI to achieve 2.25x better LLM inference performance vs. GPUs
News -
Furiosa SDK 2025.2.0 is here: Hugging Face Hub integration, reasoning model support, enhanced APIs, and more
Technical Updates -
Furiosa to bring RNGD to Microsoft’s Azure Marketplace
News -
Accelerating RAG applications with RNGD: A technical walkthrough
Technical Updates -
AI Agents Explained: Core Concepts and Key Capabilities
White Papers -
Furiosa releases Furiosa SDK 2025.1.0
Technical Updates -
Demonstrating High-Speed Inference Throughput with the Furiosa SDK
Our Viewpoints -
The Future of AI is Efficient Inference
Our Viewpoints -
FuriosaAI presents papers at ECCV, COLM conference to improve AI image generation
News -
The next chapter of Kubernetes: Enabling ML inference at scale
Our Viewpoints