Featured Energy Sessions at NVIDIA GTC 2025


Learn from energy leaders using HPC and AI to boost exploration, production, and fuel delivery, while enhancing power grid reliability and resiliency.
Learn from energy leaders using HPC and AI to boost exploration, production, and fuel delivery, while enhancing power grid reliability and resiliency.
Meta recently released its Llama 3.2 series of vision language models (VLMs), which come in 11B parameter and 90B parameter variants. These models are multimodal, supporting both text and image inputs. In addition, Meta has launched text-only small language model (SLM) variants of Llama 3.2 with 1B and 3B parameters. NVIDIA has optimized the Llama 3.2 collection of models for great performance andβ¦
NVIDIA has built three computers and accelerated development platforms to enable developers to create physical AI.
Neuromodulation is a technique that enhances or restores brain function by directly intervening in neural activity. It is commonly used to treat conditions like Parkinsonβs disease, epilepsy, and depression. The shift from open-loop to closed-loop neuromodulation strategies enables on-demand modulation, improving therapeutic effects while reducing side effects. This could lead to significantβ¦
Expanding the open-source Meta Llama collection of models, the Llama 3.2 collection includes vision language models (VLMs), small language models (SLMs), and an updated Llama Guard model with support for vision. When paired with the NVIDIA accelerated computing platform, Llama 3.2 offers developers, researchers, and enterprises valuable new capabilities and optimizations to realize theirβ¦
Vision-language models (VLMs) combine the powerful language understanding of foundational LLMs with the vision capabilities of vision transformers (ViTs) by projecting text and images into the same embedding space. They can take unstructured multimodal data, reason over it, and return the output in a structured format. Building on a broad base of pretraining, they can be easily adapted forβ¦