Market Cap: $3.476T -3.070%
Volume(24h): $112.3363B 8.220%
  • Market Cap: $3.476T -3.070%
  • Volume(24h): $112.3363B 8.220%
  • Fear & Greed Index:
  • Market Cap: $3.476T -3.070%
Cryptos
Topics
Cryptospedia
News
CryptosTopics
Videos
Top News
Cryptos
Topics
Cryptospedia
News
CryptosTopics
Videos
bitcoin
bitcoin

$102418.358867 USD

-1.97%

ethereum
ethereum

$3298.096549 USD

1.21%

xrp
xrp

$3.048127 USD

-1.30%

tether
tether

$0.999866 USD

-0.01%

solana
solana

$231.464380 USD

-2.61%

bnb
bnb

$675.655067 USD

-0.56%

usd-coin
usd-coin

$0.999928 USD

-0.01%

dogecoin
dogecoin

$0.327988 USD

-0.25%

cardano
cardano

$0.945324 USD

-1.12%

tron
tron

$0.256233 USD

0.65%

chainlink
chainlink

$25.471085 USD

1.61%

avalanche
avalanche

$34.603954 USD

-1.17%

stellar
stellar

$0.416369 USD

-2.01%

sui
sui

$4.058447 USD

-3.89%

toncoin
toncoin

$4.893106 USD

1.10%

Cryptocurrency News Articles

Hugging Face Unveils Inference-as-a-Service Powered by NVIDIA NIM at SIGGRAPH 2024

Jul 30, 2024 at 03:33 am

Open Source platform Hugging Face is offering developers Inference-as-a-Service that will be powered by NVIDIA's NIM. The new service provides 5x better token efficiency for AI models and allows immediate access to NIM microservices running on NVIDIA DGX Cloud.

Hugging Face Unveils Inference-as-a-Service Powered by NVIDIA NIM at SIGGRAPH 2024

Open Source platform Hugging Face is now offering developers Inference-as-a-Service that will be powered by NVIDIA’s NIM. The new service provides 5x better token efficiency for AI models and allows immediate access to NIM microservices running on NVIDIA DGX Cloud.

The new inference-as-a-service was announced at the ongoing SIGGRAPH 2024, a premier conference and exhibition on computer graphics and interactive techniques in Denver, Colorado. The new service will let developers deploy powerful LLMs like Llama 2, Mistral AI models and many more with optimisation from NVIDIA NIM microservices. Hugging Face Enterprise Hub users can access serverless inference for increased flexibility and minimal infrastructure overhead with NVIDIA NIM.

When accessed as a NIM, large models like the 70-billion-parameter version of Llama 3 will deliver up to 5x higher throughput when compared with off-the-shelf deployment on NVIDIA H100 Tensor Core GPU-powered systems.

The new inference service also supports Train on DGX Cloud, an AI training service that is already available on Hugging Face.

Enter NVIDIA NIM

NVIDIA NIM is a set of AI microservices, including NVIDIA AI foundation models and open-source community models, that has been optimised for inference with standard APIs. It improves token processing efficiency and enhances the NVIDIA DGX Cloud infrastructure, accelerating AI applications. This setup provides faster, more robust results.

The NVIDIA DGX Cloud platform is tailored for generative AI, offering developers reliable, accelerated computing infrastructure for faster production readiness. It supports AI development from prototyping to production without requiring long-term commitments.

Hugging Face to the Fore

The new announcement banks on an existing partnership between both tech companies and is only going to foster the developer community further. Interesting recent announcements from Hugging Face include its profitability with a 220-member team and the release of SmolLM, a series of small language models.

Disclaimer:info@kdj.com

The information provided is not trading advice. kdj.com does not assume any responsibility for any investments made based on the information provided in this article. Cryptocurrencies are highly volatile and it is highly recommended that you invest with caution after thorough research!

If you believe that the content used on this website infringes your copyright, please contact us immediately (info@kdj.com) and we will delete it promptly.

Other articles published on Feb 02, 2025