Nvidia enables broader usage of AI with LLM cloud services



Nvidia announced a series of new cloud services that aim to enable more organizations and individuals to create, train and benefit from giant AI models. The cloud offerings include the Nvidia NeMo LLM Service and the Nvidia BioNeMo LLM Service. “We are seeing an explosion of research, applying transformer models to all kinds of use cases this year,” said Paresh Kharya, senior director of accelerated computing products at Nvidia


Nvidiaは、より多くの組織や個人が巨大なAIモデルの作成、訓練、利益を可能にすることを目的とした一連の新しいクラウドサービスを発表しました。クラウドサービスには、Nvidia Nemo LLMサービスとNvidia Bionemo LLMサービスが含まれます。Nvidiaの加速コンピューティング製品のシニアディレクターであるParesh Kharya氏は、次のように述べています。


In recent years, large language models (LLMs) have become a foundational form of artificial intelligence (AI) models.


The challenge, however, has been that creating and training new LLMs is far from a trivial exercise.


At the Nvidia GTC conference today, the company made a long list of announcements spanning the full spectrum of AI operations across multiple industries.

本日のNVIDIA GTC会議で、同社は複数の業界でAIオペレーションの全範囲にまたがる発表の長いリストを作成しました。

One of the key announcements that Nvidia made is about a series of new LLM capabilities, including a pair of cloud services that aim to enable more organizations and individuals to create, train and benefit from LLMs.


“We are announcing NeMo LLM Service to enable customization and inference of giant AI models,” Paresh Kharya, senior director of accelerated computing products at Nvidia, told VentureBeat.

NVIDIAのAccelerated Computing ProductsのシニアディレクターであるParesh Kharyaは、VentureBeatに、「巨大なAIモデルのカスタマイズと推論を可能にするために、Nemo LLMサービスを発表しています」と語った。

“Just like how LLMs can understand the human language, they’ve also been trained to understand the language of biology and chemistry.” LLMs are based on AI transformer architecture and are widely used to support a growing number of use cases.


Kharya explained that with a transformer, the AI model can understand which parts of a sentence, an image or even very disparate data points are relevant to each other.


Unlike convolutional neural networks (CNNs), which typically look at only the immediate neighboring relationships, transformers are designed to train on more distant relationships as well, which Kharya said is very important for use cases like natural language processing (NLP).


“Transformers also enable us to train on unlabeled datasets, and that greatly expands the volume of data,” he said.


“We are really seeing an explosion of research, applying transformer models to all kinds of use cases this year.


We are expected to have 11,000 papers on transformers, actually seven times more than five years ago.” The GPT-3 LLM has helped to increase awareness and adoption of LLMs for a variety of use cases, including summation and text generation.

トランスに関する11,000の論文があり、実際には5年以上前に7倍になると予想されています。」GPT-3 LLMは、合計やテキスト生成を含むさまざまなユースケースのLLMSの認識と採用を増やすのに役立ちました。

An LLM is also at the foundation of the DALL-E text-to-image generation technology.


“Today, we are seeing LLMs being applied to predict protein structures from sequences of amino acids or for understanding and generating art by learning the relationship between pixels,” Kharya said.


As with any type of AI model, context matters.


What might make sense for one audience or use case will not be appropriate for another.


Training entirely new LLMs for every type of use case is a time-consuming process.


Kharya said that an emerging approach of providing context to LLMs for specific use cases is a technique known as prompt learning.


He explained that with prompt learning, a companion model is trained that learns to provide the context to the pretrained large language model, using what’s called a prompt token.  The companion model can learn the context by using as few as 100 examples of queries with the right responses.


At the end of the prompt learning training, a token is generated that can then be used together with the query, which will provide the context required from the LLM.


The new NeMo LLM Service is an effort to make it easier to enable customization and inference of giant AI models.  The giant AI models that the service will support include a 5 billion- and a 20 billion-parameter GPT-based model, as well as one based on the Megatron 530-billion parameter LLM.

新しいNEMO LLMサービスは、巨大なAIモデルのカスタマイズと推論を容易にするための取り組みです。サービスがサポートする巨大なAIモデルには、50億および20億パラメーターのGPTベースのモデルと、530億ポラメーターLLMに基づいたモデルが含まれます。

As part of the service, Nvidia is also supporting prompt learning–based tuning to rapidly enable context-specific use cases.


Kharya said that the NeMo LLM Service will also include the option to use both ready-made models and custom models through a cloud-based API experience.

Kharyaは、Nemo LLMサービスには、クラウドベースのAPIエクスペリエンスを通じて、既製モデルとカスタムモデルの両方を使用するオプションも含まれると述べました。

Going a step further, Nvidia is also launching a specific LLM capability for life sciences with the BioNeMo Service.


“Just like how an LLM can understand the human language, they’ve also been trained to understand the language of biology and chemistry,” Kharya said.  Kharya said that, for example, DNA is the language basically written in the alphabet of nucleic acid and the language of protein structures is written in the alphabet of amino acids.  Overall the goal with the new LLM services is to further expand the use of AI.


“The promises and possibilities are really immense and it’s the access to large language models and the ability to customize them easily that was not there before,” Kharya said.


“So what the NeMo Large Language Model Service does is it removes that barrier and it now enables everyone to access and experiment with [LLMs] for their use cases.” VentureBeat’s mission is to be a digital town square for technical decision-makers to gain knowledge about transformative enterprise technology and transact.

「だから、Nemo Large Language Model Serviceが行うことは、その障壁を取り除くことであり、これにより、誰もがユースケースの[LLM]にアクセスして実験できるようになりました。」VentureBeatの使命は、技術的な意思決定者が変革的なエンタープライズテクノロジーと取引に関する知識を得るためのデジタルタウンスクエアになることです。