Choose, train

and deploy AI

with ResetData

Host AI workloads or launch new AI capabilities in minutes

NIM Models

llama-3.2-11b-vision-instruct

visual qa

Meta /

llama-3.2-11b-vision-instruct

The Meta Llama 3.2 Vision collection of multimodal large language models (LLMs) is a collection of pre-trained and instruction-tuned image reasoning generative models in 11B and 90B sizes (text + images in / text out). The Llama 3.2 Vision instruction-tuned models are optimized for visual recognition, image reasoning, captioning, and answering general questions about an image. The models outperform many of the available open source and closed multimodal models on common industry benchmarks.

Image-Text Retrieval
Visual Grounding
Visual QA
Image-to-Text

NIM Models

llama-4-scout-17b-16e-instruct

reasoning

meta /

llama-4-scout-17b-16e-instruct

The Llama 4 collection of models are natively multimodal AI models that enable text and multimodal experiences. These models leverage a mixture-of-experts architecture to offer industry-leading performance in text and image understanding. These Llama 4 models mark the beginning of a new era for the Llama ecosystem. We are launching two efficient models in the Llama 4 series, Llama 4 Scout, a 17 billion parameter model with 16 experts, and Llama 4 Maverick, a 17 billion parameter model with 128 experts.

Language generation
Image-to-text
Vision assistant
Reasoning
Visual question answering

NIM Models

nvidia-llama-3.2-nv-rerankqa-1b-v2

retrieval augmented generation

NVIDIA /

nvidia-llama-3.2-nv-rerankqa-1b-v2

The NVIDIA NeMo Retriever Llama3.2 reranking model is optimized for providing a logit score that represents how relevant a document(s) is to a given query. The model was fine-tuned for multilingual, cross-lingual text question-answering retrieval, with support for long documents (up to 8192 tokens).

nemo retriever
reranking
Retrieval Augmented Generation

NIM Models

llama-3.3-70b-instruct

reasoning

Meta /

llama-3.3-70b-instruct

The Meta Llama 3.3 multilingual large language model (LLM) is a pretrained and instruction tuned generative model in 70B (text in/text out). The Llama 3.3 instruction-tuned text-only model is optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.

Instruction following
Math
Text-to-Text
Reasoning
Code Generation

New AI Factory capacity is now available. Reserve yours for AI and GPUaaS workloads today.

OUR DIFFERENCE

We eliminate the latency, heat and efficiency bottlenecks to making new things possible with AI.

Zero

wastewater

0%

fewer emissions

0%

lower costs

NVIDIANVIDIADELLDELL

You can't fight physics. As data centres (DCs) add Graphical Processing Units (GPU) and chips exceeding the 400W limit of air-cooling identified by ASHRAE standards, some DCs spend half their energy on cooling, not computing.

AI factories with liquid immersion cooling up to 10X as efficient as legacy designs can slash cloud costs 40% and emissions 45% to deliver more sustainable AI cloud operations.