Market Cap: $3.4491T 2.49%
Volume(24h): $195.6881B -32.40%
  • Market Cap: $3.4491T 2.49%
  • Volume(24h): $195.6881B -32.40%
  • Fear & Greed Index:
  • Market Cap: $3.4491T 2.49%
Cryptos
Topics
Cryptospedia
News
CryptosTopics
Videos
Top News
Cryptos
Topics
Cryptospedia
News
CryptosTopics
Videos
bitcoin
bitcoin

$103094.926080 USD

3.95%

ethereum
ethereum

$3398.208576 USD

6.43%

tether
tether

$0.999971 USD

-0.04%

xrp
xrp

$2.326205 USD

9.96%

bnb
bnb

$947.145845 USD

4.46%

solana
solana

$160.315987 USD

7.54%

usd-coin
usd-coin

$1.000014 USD

0.01%

tron
tron

$0.288163 USD

2.37%

dogecoin
dogecoin

$0.164881 USD

5.50%

cardano
cardano

$0.536519 USD

7.14%

hyperliquid
hyperliquid

$40.526327 USD

6.62%

chainlink
chainlink

$14.898178 USD

5.68%

bitcoin-cash
bitcoin-cash

$483.923206 USD

4.44%

ethena-usde
ethena-usde

$0.999280 USD

0.02%

stellar
stellar

$0.276354 USD

6.32%

Cryptocurrency News Articles

Nebius, Token Factory, and AI Inference: A New Era of Open-Source AI?

Nov 05, 2025 at 09:48 pm

Explore how Nebius's Token Factory is revolutionizing AI inference by empowering open-source models with enterprise-grade scalability and cost-efficiency.

Nebius, Token Factory, and AI Inference: A New Era of Open-Source AI?

Nebius, Token Factory, and AI Inference: A New Era of Open-Source AI?

The AI landscape is rapidly evolving, with inference costs becoming a major factor. Nebius's Token Factory offers a production inference platform that enables companies to deploy and optimize open-source AI models, potentially revolutionizing the economics of AI inference.

Unveiling Nebius Token Factory

Nebius has launched the Nebius Token Factory, a platform designed to democratize AI inference. By supporting major open-source models like NVIDIA Nemotron, DeepSeek, GPT-OSS by OpenAI, Llama, and Qwen, Token Factory empowers AI companies and enterprises to leverage the flexibility of open models without the complexities of managing them in production.

Key Features and Benefits

Nebius Token Factory stands out due to its ability to deliver sub-second latency, autoscaling throughput, and 99.9% uptime. The platform's architecture is optimized for efficiency, reducing inference costs and latency by up to 70%. Key features include:

  • Support for major open-source models: Seamlessly deploy and optimize various AI models.
  • Enterprise-grade reliability: Benefit from high availability and consistent performance.
  • Cost-efficiency: Reduce inference costs through optimized infrastructure.
  • Teams and Access Management: Enhance collaboration and ensure compliance with granular access control.

Real-World Impact

Early adopters are already seeing significant benefits. Prosus, for example, has achieved up to 26x cost reductions compared to proprietary models. Higgsfield AI relies on Nebius for on-demand and autoscaling inference, enabling faster and more cost-efficient AI in production. Hugging Face is collaborating with Nebius to improve access and scalability for developers.

NVIDIA's Blackwell Platform and InferenceMAX

NVIDIA's Blackwell platform is emerging as a frontrunner in AI inference. According to the InferenceMAX v1 benchmark, a $5 million NVIDIA GB200 NVL72 system could generate about $75 million in token revenue, a 15x return on investment. This platform delivers 10x more throughput per megawatt and cuts cost per million tokens by 15x compared to the previous generation. NVIDIA's full-stack approach optimizes model performance through collaborations with OpenAI, Meta, and DeepSeek AI, along with software tweaks like the TensorRT LLM library.

The Rise of AI Factories

The AI industry is shifting from pilot projects to AI factories. Nebius Token Factory, along with NVIDIA's Blackwell platform, is playing a crucial role in this transformation by providing the infrastructure needed to turn data into tokens, predictions, and business decisions in real-time.

Final Thoughts

With Nebius Token Factory and advancements in platforms like NVIDIA Blackwell, the future of AI inference looks bright. Open-source models are becoming more accessible and cost-effective, empowering organizations to innovate and scale their AI initiatives. Who knows? Maybe one day, AI will be so efficient, it'll write its own blog posts. Until then, we'll keep you updated!

Original source:seekingalpha

Disclaimer:info@kdj.com

The information provided is not trading advice. kdj.com does not assume any responsibility for any investments made based on the information provided in this article. Cryptocurrencies are highly volatile and it is highly recommended that you invest with caution after thorough research!

If you believe that the content used on this website infringes your copyright, please contact us immediately (info@kdj.com) and we will delete it promptly.

Other articles published on Nov 07, 2025