![]() |
|
![]() |
|
![]() |
|
![]() |
|
![]() |
|
![]() |
|
![]() |
|
![]() |
|
![]() |
|
![]() |
|
![]() |
|
![]() |
|
![]() |
|
![]() |
|
![]() |
|
加密貨幣新聞文章
Meta Holds Its First-Ever Event for AI Developers, LlamaCon, Announcing That It's Ready to Compete with ChatGPT
2025/04/30 04:15
Meta held its first-ever event for AI developers, LlamaCon, at the company’s headquarters in Menlo Park, where it announced that it was ready to compete with ChatGPT from OpenAI, as well as Google, AWS, and AI-as-a-service startups.
Meta Founder and CEO Mark Zuckerberg was joined by Co-Founder and CEO of Databricks, Ali Ghodsi.
This is really a big deal for Meta and the AI industry as the maker of the popular open-source Llama LLM seeks to directly monetize the incredible adoption Meta has realized. Developers just access the model from the cloud; no hardware or software to install.
But it is also a big deal for Cerebras and Groq, the two startups selected by Meta for serving fast tokens, many times faster than a GPU. (Nvidia, Cerebras and Groq are all clients of Cambrian-AI Research.) Meta did not disclose pricing, as access to the API is currently in preview, and access to Groq and Cerebras is only available by request. This is the first time either startup has landed a foothold at a hyper-scale Cloud Service Provider (CSP). And Meta has made it super easy to use; developers just select Groq or Cerebras in the API call.
Cerebras is the industry's fastest inference processor by far (~18X) but Grok is also 5-fold faster ... More than any GPU.
“Cerebras is proud to make Llama API the fastest inference API in the world,” said Andrew Feldman, CEO and co-founder of Cerebras. “Developers building agentic and real-time apps need speed. With Cerebras on Llama API, they can build AI systems that are fundamentally out of reach for leading GPU-based inference clouds.”
Microsoft Confirms $1.50 Windows Security Update Hotpatch Fee Starts July 1
Microsoft Confirms Password Spraying Attack — What You Need To Know
Google’s Gmail Upgrade—Why You Need To Change Your App
Llama on Cerebras is far faster than on Google TPUs or Nvidia GPUs.
Andrew’s point is important. Obtaining inferences at some 100 tokens per second is faster than a human can read, so “one-shot” inference requests for a service like ChatGPT runs just fine on GPUs. But multi-model agents and reasoning models can increase computational requirements by some 100-fold, opening an opportunity for faster inference from companies like Cerebras, Groq. Meta did not mention the third fast-inference company, Samba Nova, but indicated that they are open to other compute options in the future.
It will be interesting to see how well these two new options fare in the tokens-as-a-service world.
免責聲明:info@kdj.com
所提供的資訊並非交易建議。 kDJ.com對任何基於本文提供的資訊進行的投資不承擔任何責任。加密貨幣波動性較大,建議您充分研究後謹慎投資!
如果您認為本網站使用的內容侵犯了您的版權,請立即聯絡我們(info@kdj.com),我們將及時刪除。
-
-
- io.net(IO)Q1 2025市場分析
- 2025-05-23 22:35:14
- IO.NET(IO)是一個分散的圖形處理單元(GPU)和中央處理單元(CPU)網絡,旨在提供可擴展有效的計算資源訪問。
-
-
- 正式驗證是計算機科學的理論領域之一。
- 2025-05-23 22:30:15
- 從歷史上看,這一領域是晦澀難懂的,但是AI的最新進展可能會將其帶到前面和中心。
-
- CME Group的XRP期貨的推出使看漲的海浪通過市場
- 2025-05-23 22:25:14
- 隨著全球加密貨幣動量圍繞XRP激增,CME集團的XRP期貨的推出使Bullish Waves遍及市場。
-
-
-
-
- PI Network(PI)價格慢慢從5月初的大幅下降中恢復過來
- 2025-05-23 22:15:14
- 在5月12日至5月17日之間下降了約50%之後,代幣在過去四天中備份了20%以上,估價為0.83美元