Get answers to your most pressing questions from NVIDIA technology experts.
Connect With the Experts sessions are a unique opportunity to meet in person with the people behind NVIDIA products and research at an interactive area of the San Jose Convention Center.
Each 50-minute session will give you a chance to get your questions answered and learn about the latest NVIDIA technologies—including generative AI, agentic AI, data center, robotics, data processing, accelerated computing, and more.
We'll provide a brief overview of the NVIDIA Nsight™ family of tools and debuggers. You’ll learn about the latest features for debugging, profiling, and optimizing NVIDIA® CUDA® compute and graphics applications on the newest platforms. Several experts from the tools development and management teams will be available to talk about getting started, best practices, and advanced techniques for application debugging and optimization with NVIDIA developer tools. You can ask questions about anything from high-level concepts to specific tools details, and also discuss plans and suggestions for improving the tools in the future.
Key Takeaways:
Join us and learn about quantum computing at NVIDIA! We'll show you how to use CUDA-Q and cuQuantum to accelerate your quantum computing research, from simulating algorithms and processors to running hybrid quantum-classical workloads. In this interactive session, we'll discuss how to take full advantage of NVIDIA’s most powerful GPUs to explore the benefits of quantum computing. Join NVIDIA quantum software experts to learn more about the CUDA-Q open-source platform for programming quantum processing units (QPUs), GPUs, and CPUs seamlessly, and how it integrates with cuQuantum to accelerate applications simulation. Discuss your ideas with the team and give us feedback on your experience using NVIDIA’s quantum computing software.
Key Takeaways:
Join NVIDIA experts to explore the field of digital health. Discuss how NVIDIA NIM™ microservices and blueprints empower developers, researchers, and enterprises to quickly create generative AI solutions in healthcare incorporating clinical documentation and agentic health workflows. Dive into the latest advancements in generative AI foundational models, learning about their applications in understanding, summarization, and how they intersect in real-world clinical applications. And engage directly with our product managers, developer relations, and solution architects in discussions that address your biggest challenges.
Key Takeaways:
Real-time edge AI and robotic systems are transforming medical workflows and creating revolutionary devices that are advancing the standard of care for disease detection, diagnosis, and interventions. NVIDIA's platforms—MONAI, Holoscan, Isaac™, Omniverse™, and Metropolis—span model training, digital twins/simulations, and runtime deployment. Together, they give developers an end-to-end, accelerated, full-stack infrastructure that enables rapid prototyping and scalable development of real-time AI applications. Learn from NVIDIA experts on how our platforms can save you time and money in bringing AI-enabled systems to market—from smart diagnostic devices and innovative imaging systems to surgical robots trained using digital twin technologies.
Key Takeaways:
Do you want to learn how to get started accelerating large language models and other state-of-the-art deep learning models? Join this interactive support session with the cuDNN engineering staff to get direct help and surface your questions, issues, and ideas specific to your use case.
Key Takeaways:
Visit us to discuss ways to get started or deepen your understanding of NVIDIA tools that support vision AI applications. Our topic experts are offering one-on-one dialogue on topics related to training, fine-tuning, and deploying a range of vision AI models. Ask about NVIDIA Metropolis, VLMs, foundation models, NIM microservices, reference workflows, DeepStream, TAO, TensorRT™, Triton™ Inference Server, Jetson™, and more.
Key Takeaways:
Join NVIDIA experts to explore the field of digital biology. We’ll discuss the latest advancements in accelerated computing and foundation models for drug discovery and genomics—from bioinformatics tooling to protein structure prediction, molecular dynamics to molecular generation, and much more. Learn how NVIDIA BioNeMo™ and Parabricks, together with NVIDIA NIM microservices and blueprints, empower developers, researchers, and enterprises to quickly create generative AI solutions across chemistry, biology, and genomics. Engage directly with our product managers, developer relations, and solution architects in discussions that address your challenges and seek answers to your questions.
Key Takeaways:
The IDC has predicted that by 2025, 80% of all data will be unstructured. Join us for an interactive session and learn how you can use GPUs to perform semantic search and data mining at scale on embeddings created from unstructured data. We'll dive into the cuVS library and provide an overview of the various databases and libraries that have integrated it to achieve state-of-the-art performance for building and searching vector indexes.
Key Takeaways:
Did you know GPUs can reduce the cost of your Apache Spark workloads? Come join this session to find out how. Whether you’re new to Apache Spark or a long-time user, we're here to have a conversation about large-scale data processing, your production workloads, optimal configurations, and how to introduce GPUs to your environment. We can also answer questions about job qualification and tuning with GPUs. Get questions answered about potential performance gains, cost savings, or power usage reduction possible with GPU-accelerated Spark. Learn about the migration journey others have already undertaken.
Key Takeaways:
Join this engaging session led by NVIDIA experts who collaborate closely with the global academic research community, supporting leading researchers in using NVIDIA software technologies to drive scientific advancements across diverse fields. We'll provide insights into how various NVIDIA teams, representing different functional areas, can assist in accelerating your scientific discoveries at multiple levels. The team will outline their approach to working with individual researchers and groups, share success stories, and introduce a range of developer support programs, including but not limited to the Academic Grant Program, Deep Learning Institute courses and Teaching Kits, and Researcher Advisory Councils. We'll also focus on aligning discussions with specific research domains such as robotics, biology, generative AI, data processing, and climate science—while showcasing tools and resources available to help researchers enhance research outcomes.
Key Takeaways:
Just as critical to the performance of AI model training, tail latency can negatively affect real-time AI inference, financial trading, and large-scale data analytics, where even microsecond delays can affect outcomes. The approach to zero-tail latency and accelerated networking with enhanced programmable I/O represent a paradigm shift in the performance and scalability of modern computing infrastructures. This session will open the discussion to dive into the significance of accelerated networking technologies in delivering zero-tail latency, ensuring that every data transfer meets demanding performance requirements even beyond 100k GPU scale.
Key Takeaways:
Join this interactive session to deep dive into the world of large language model (LLM) inference. Learn about techniques to reduce computational overhead and discover strategies for scaling your applications to handle high-concurrency demands. Whether you’re exploring best practices for application deployment with a NIM or low-level model optimizations with Tensor-RT-LLM to accelerate inference and minimize latency, this session will cover it all.
Key Takeaways:
Next-generation robots will need to sense, plan, and act autonomously. A "sim-first" approach is required to ensure that the entire robot stack performs. Connect with NVIDIA experts to learn how you use NVIDIA Isaac Sim™ for your robotics workflows, including generating synthetic data for perception model training and software-in-loop testing to validate your robot stack.
Key Takeaways:
We'll introduce the newest TensorCore programming guide, explaining how to build up NVIDIA Blackwell architecture TensorCore GEMM, use new low-precision TensorCore, map the above concepts to CUTLASS, and better use CUTLASS. We'll also share techniques to optimize memory bandwidth and maximize latency hiding. A straightforward rule of thumb can predict the bandwidth use of a kernel, simplifying the allocation of staging buffers in shared memory.
Key Takeaways:
Deploying AI models for real-world applications demands efficient inference strategies to handle their intense memory and compute requirements. Learn the key phases of LLM inference and provide practical techniques for developers and data scientists to optimize model performance, reduce costs, and achieve scalable deployment. We'll also cover state-of-the-art attention mechanism improvements, such as multi-query and flash attention, to optimize memory use during inference. Additionally, we'll explore model-serving strategies like in-flight batching and speculative inference, which can significantly enhance GPU utilization while minimizing latency in production environments.
Key Takeaways:
Compute for humanoid robotics is crucial to deliver optimized performance and advanced AI capabilities. NVIDIA Jetson™ Thor for humanoid robotics features the next-gen NVIDIA Blackwell architecture with integrated safety features and high-performance computing to ensure safe and efficient interactions in real-world environments. Connect with our experts to understand how to unlock power and performance for humanoid robotics applications from software, AI models, and libraries to compute. Join us to accelerate your humanoid robotics projects and drive innovation in embodied AI.
Key Takeaways:
NVIDIA Grace™ Hopper and Grace Blackwell are a new breed of memory-converged CPU-GPU superchips engineered to meet the challenges of AI, high-performance computing, and data processing. Learn about the unique architecture of these superchips and their versatile form factors that simplify data center deployment. We'll address such questions as how the 900 GB/s NVLink™ chip-to-chip interconnect enables the CPU and GPU to access all system-allocated memory, how you can seamlessly migrate your x86-based workloads to the Arm-based Grace CPU, and the performance gains you can expect to unlock for your applications.
Key Takeaways:
Get unique insights from NVIDIA's Kaggle Grandmasters, who have competed and earned top honors in hundreds of international data science competitions. The team brings deep expertise in generative AI, computer vision, natural language processing, forecasting, recommender systems, tabular data, signals, and graphs. Discover the latest tools and techniques winning competitions today, hear best practices for GPU acceleration across diverse use cases, and explore how these experts approach complex challenges. Bring your questions and learn directly from data science champions on optimizing model performance for real-world applications.
Key Takeaways: