CUDA
CUDA® is a parallel computing platform and programming model developed by NVIDIA for general computing on graphical processing units (GPUs). With CUDA, developers are able to dramatically speed up computing applications by harnessing the power of GPUs.
In GPU-accelerated applications, the sequential part of the workload runs on the CPU – which is optimized for single-threaded performance – while the compute intensive portion of the application runs on thousands of GPU cores in parallel. When using CUDA, developers program in popular languages such as C, C++, Fortran, Python and MATLAB and express parallelism through extensions in the form of a few basic keywords.
The CUDA Toolkit from NVIDIA provides everything you need to develop GPU-accelerated applications. The CUDA Toolkit includes GPU-accelerated libraries, a compiler, development tools and the CUDA runtime.
Learn more
NVIDIA Confidential Computing
NVIDIA Confidential Computing secures data in use, protecting AI models and workloads as they execute, by leveraging hardware-based trusted execution environments built into NVIDIA Hopper and Blackwell architectures and supported platforms. It enables enterprises to deploy AI training and inference, whether on-premises, in the cloud, or at the edge, with no changes to model code, while ensuring the confidentiality and integrity of both data and models. Key features include zero-trust isolation of workloads from the host OS or hypervisor, device attestation to verify that only legitimate NVIDIA hardware is running the code, and full compatibility with shared or remote infrastructure for ISVs, enterprises, and multi-tenant environments. By safeguarding proprietary AI models, inputs, weights, and inference activities, NVIDIA Confidential Computing enables high-performance AI without compromising security or performance.
Learn more
Cody
Cody, Sourcegraph’s AI code assistant goes beyond individual dev productivity, helping enterprises achieve consistency and quality at scale with AI.
Unlike traditional coding assistants, Cody understands the entire codebase, enabling deeper contextual awareness for smarter autocompletions, refactoring, and AI-driven code suggestions. It integrates with IDEs like VS Code, Visual Studio, Eclipse, and JetBrains, providing inline editing and chat without disrupting workflows. Cody also connects with tools like Notion, Linear, and Prometheus to enhance development context. Powered by advanced LLMs like Claude Sonnet 4 and GPT-4o, it optimizes speed and performance based on enterprise needs, and is always adding the latest AI models. Developers report significant efficiency gains, with some saving up to six hours per week and doubling their coding speed.
Learn more
NVIDIA TensorRT
NVIDIA TensorRT is an ecosystem of APIs for high-performance deep learning inference, encompassing an inference runtime and model optimizations that deliver low latency and high throughput for production applications. Built on the CUDA parallel programming model, TensorRT optimizes neural network models trained on all major frameworks, calibrating them for lower precision with high accuracy, and deploying them across hyperscale data centers, workstations, laptops, and edge devices. It employs techniques such as quantization, layer and tensor fusion, and kernel tuning on all types of NVIDIA GPUs, from edge devices to PCs to data centers. The ecosystem includes TensorRT-LLM, an open source library that accelerates and optimizes inference performance of recent large language models on the NVIDIA AI platform, enabling developers to experiment with new LLMs for high performance and quick customization through a simplified Python API.
Learn more