Jump to content

Nvidia breaks records in training and inference for real-time conversational AI


NelsonG

Recommended Posts

Nvidia’s GPU-powered platform for developing and running conversational AI that understands and responds to natural language requests has achieved some key milestones and broken some records that have big implications for anyone building on their tech – which includes companies large and small, since much of the code they’ve used to achieve these advancements is open source, written in PyTorch and easy to run.

The biggest achievements Nvidia announced today include its breaking the hour mark in training BERT, one of the world’s most advanced AI language models and a state-of-the-art model widely considered a good standard for natural language processing. Nvidia’s AI platform was able to train the model in under an hour, a record-breaking achievement at just 53 minutes, and the trained model could then successfully infer (ie, actually applying the learned capability achieved through training to achieve results) in under 2 milliseconds (10 milliseconds is considered a high-water mark in the industry), another record.

Nvidia’s breakthroughs aren’t just cause for bragging rights – these advances scale and provide real-world benefits for anyone working with their NLP conversational AI and GPU hardware. Nvidia achieved its record-setting times for training on one of its SuperPOD systems which is made up of 92 Nvidia DGX-2H systems runnings 1,472 V100 GPUs, and managed the inference on Nvidia T4 GPUs running Nvidia TensorRT – which beat the performance of event highly optimized CPUs by many orders of magnitude. But it’s making available the BERT training code, and TensorRT optimized BERT Sample via GitHub for all to leverage.

Alongside these milestones, Nvidia’s Research wing also build and trained the largest ever language model based on ‘Transformers,’ which is the tech that underlies BERT, too. This custom model includes a massive 8.3 billion parameters, making it 24 times the size of BERT-Large, the largest current core BERT model. Nvidia has cheekily titled this model ‘Megatron,’ and also offered up the PyTorch code it used to train this model so that others can also train their own similar, massive Transformer-based language models.

Techcrunch?d=2mJPEYqXBVI Techcrunch?d=7Q72WNTAKBA Techcrunch?d=yIl2AUoC8zA Techcrunch?i=g72Wn_orMAI:OWn1Cy6_hmQ:-BT Techcrunch?i=g72Wn_orMAI:OWn1Cy6_hmQ:D7D Techcrunch?d=qj6IDK7rITs
g72Wn_orMAI

View the full article

Link to comment
Share on other sites

Please sign in to comment

You will be able to leave a comment after signing in



Sign In Now
  • Recently Browsing   0 members

    • No registered users viewing this page.
  • Our picks

    • Wait, Burning Man is going online-only? What does that even look like?
      You could have been forgiven for missing the announcement that actual physical Burning Man has been canceled for this year, if not next. Firstly, the nonprofit Burning Man organization, known affectionately to insiders as the Borg, posted it after 5 p.m. PT Friday. That, even in the COVID-19 era, is the traditional time to push out news when you don't want much media attention. 
      But secondly, you may have missed its cancellation because the Borg is being careful not to use the C-word. The announcement was neutrally titled "The Burning Man Multiverse in 2020." Even as it offers refunds to early ticket buyers, considers layoffs and other belt-tightening measures, and can't even commit to a physical event in 2021, the Borg is making lemonade by focusing on an online-only version of Black Rock City this coming August.    Read more...
      More about Burning Man, Tech, Web Culture, and Live EventsView the full article
      • 0 replies
    • Post in What Are You Listening To?
      Post in What Are You Listening To?
    • Post in What Are You Listening To?
      Post in What Are You Listening To?
    • Post in What Are You Listening To?
      Post in What Are You Listening To?
    • Post in What Are You Listening To?
      Post in What Are You Listening To?
×
×
  • Create New...