Wenwen Gao

Wenwen Gao is a senior product manager for NeMo at NVIDIA, focusing on LLM training framework and microservices. Her past experience include LLM inference (NIM) and recommender systems (Merlin). She holds a B.S. in computer science from the University of Toronto and an M.B.A. from the MIT Sloan School of Management.
Wenwen Gao

Posts by Wenwen Gao

Generative AI

Reinforcement Learning with NVIDIA NeMo-RL: Reproducing a DeepScaleR Recipe Using GRPO

Reinforcement learning (RL) is the backbone of interactive AI. It is fundamental for teaching agents to reason and learn from human preferences, enabling... 5 MIN READ
Generative AI

Scaling to Millions of Tokens with Efficient Long-Context LLM Training

The evolution of large language models (LLMs) has been marked by significant advancements in their ability to process and generate text. Among these... 7 MIN READ
Generative AI

Run Hugging Face Models Instantly with Day-0 Support from NVIDIA NeMo Framework

As organizations strive to maximize the value of their generative AI investments, accessing the latest model developments is crucial to continued success. By... 6 MIN READ
A multi-data center illustration.
Data Center / Cloud

Turbocharge LLM Training Across Long-Haul Data Center Networks with NVIDIA Nemo Framework

Multi-data center training is becoming essential for AI factories as pretraining scaling fuels the creation of even larger models, leading the demand for... 6 MIN READ
Decorative image of a datacenter with floating icons overlaid.
Generative AI

LLM Inference Benchmarking Guide: NVIDIA GenAI-Perf and NIM

This is the second post in the LLM Benchmarking series, which shows how to use GenAI-Perf to benchmark the Meta Llama 3 model when deployed with NVIDIA NIM. ... 11 MIN READ
Generative AI

LLM Inference Benchmarking: Fundamental Concepts

This is the first post in the large language model latency-throughput benchmarking series, which aims to instruct developers on common metrics used for LLM... 15 MIN READ