AI Business is part of the Informa Tech Division of Informa PLC

This site is operated by a business or businesses owned by Informa PLC and all copyright resides with them. Informa PLC's registered office is 5 Howick Place, London SW1P 1WG. Registered in England and Wales. Number 3099067.

AI Leaders

GTC 2022: Nvidia launches large language model tools

Article ImageNeMo LLM and BioNeMo to launch in beta

Nvidia is launching tools for customizing AI applications based on large language models that makes training processes exponentially faster.

Large language models (LLM) form the basis of many AI applications people interact with daily. The likes of GPT-3, LaMDA and OPT-175B perform everything from voice assistants to search engines.

To help developers work with such systems, Nvidia plans to launch new offerings to cover language models ranging in size from three billion parameters, all the way to its own Megatron 530B, which is among the largest monolithic transformer-based language models.

The NeMo Large Language Model Service is designed for developers to tailor several pretrained foundation models via prompt learning on Nvidia-managed infrastructure.

BioNeMo enables the deployment of customized AI applications for uses such as content generation, text summarization and chatbots. The service can also be applied to science-focused applications, like drug discovery.

The new tools enable training processes on related models to take minutes to hours compared with weeks or months, according to Nvidia.

“Large language models hold the potential to transform every industry,” said Jensen Huang, founder and CEO of Nvidia, in a statement. “The ability to tune foundation models puts the power of LLMs within reach of millions of developers who can now create language services and power scientific discoveries without needing to build a massive model from scratch.”

NeMo LLM and BioNeMo services will be available in early access from October. Developers are required to apply to access the offerings.

The beta release of the NeMo Megatron framework is available from Nvidia NGC and is optimized to run on Nvidia DGX Foundry and DGX SuperPOD, as well as accelerated cloud services from AWS, Microsoft Azure and Oracle Cloud Infrastructure.

Related stories:

5 takeaways from the Nvidia GTC 2022 conference

Nvidia GTC 2022: AI to revolutionize gaming

GTC 2022: Nvidia’s first SaaS offering and Omniverse updates

GTC 2022: Nvidia, Booz Allen launch AI-enabled cybersecurity platform

GTC 2022: Nvidia unveils industrial-grade edge AI computing platform

GTC 2022: Nvidia invests $15M in autonomous building platform

Trending Stories
All Upcoming Events

Upcoming Webinars

More Webinars

Latest Videos

More videos


More EBooks

Research Reports

More Research Reports
AI Knowledge Hub

Newsletter Sign Up

Sign Up