NVIDIA ACE#
NVIDIA ACE is a suite of real-time AI solutions for end-to-end development of interactive avatars and digital human applications at-scale. Its customizable microservices offer the fastest and most versatile solution for bringing avatars to life at-scale, based on NVIDIA’s Unified Compute Services, full-stack AI platform and RTX technology. Leverage ACE to seamlessly integrate NVIDIA AI into your application, including Riva for speech and translation AI, Voice Font for transferring your voice to a digital avatar, NeMo LLM for natural language understanding, Audio2Emotion, Animation Graph, Omniverse Renderer or A2F-2D or A2F-3D for AI powered 2D and 3D character animation. Or, further optimize for cloud scale with microservices and domain-specific AI workflows like NVIDIA Maxine, NVIDIA Tokkio, Animation Pipeline, and NVIDIA Kairos, to deliver advanced avatars across video conferencing customer service use cases and gaming.
- Animation Graph Microservice
- Omniverse Renderer Microservice
- Audio2Face-3D
- Overview
- Getting Started
- Architecture overview
- Audio2Face-3D Microservice
- Observability
- Sample application connecting to Audio2Face-3D
- Performance
- Fetch current configuration from Audio2Face-3D
- gRPC with Audio2Face-3D
- Unreal Engine Interaction
- A2F-3D NIM Manual Container Deployment and Configuration
- Prerequisites
- Configuration files
- How to use configuration files
- Model caching
- Starting the A2F-3D NIM with custom entrypoint
- Changing Configuration - The Shortest Way
- Changing Configuration - The Flexible Way
- Advanced Stylization
- Configuration files for Unreal Engine Metahuman
- Parameter Tuning Guide
- Environment variables
- Volumes
- Quick Deployment of Audio2Face-3D Microservices
- Kubernetes Deployment
- Support Matrix
- Migrating from 1.0 to 1.2
- Sharing Audio2Face-3D Compute Resources
- Network Customization and Accessibility
- Glossary of Terms
- Security
- Release Notes
- Troubleshooting
- ACE Agent
- Getting Started
- Architecture
- Deployment
- Tutorials
- Building a Bot using Colang 2.0 and Event Interface
- Building LangChain-Based Bots
- Building a Low Latency Speech-To-Speech RAG Bot
- Customizing a Bot
- Selecting an LLM Model
- Using an On-Premise LLM Model Deployed via NIM
- Creating a New Custom Action
- Using a Custom NLP Model
- Customizing ASR Recognition with Word Boosting
- Customizing ASR Recognition for Long Pause Handling
- Customizing TTS Pronunciation using IPA
- Using 3rd Party Text-to-Speech (TTS) Solutions
- Sample Bots
- Configuration Guide
- User Guide
- API Guide
- Best Practices
- Best Practices Introduction
- Planning your Bot
- Step 1: Choose your Bot’s Domain of Expertise
- Step 2: Determine the Conversation Flow and Knowledge Bases
- Step 3: Determine the Language Model
- Step 4: Determine the Dialog Modeling Language
- Step 5: Plan More Modalities
- Step 6: Determine the Frontend Channel
- Step 7: Determine the Deployment Infrastructure
- Logging and Debugging Issues
- Security Considerations
- Reference
- Riva ASR
- Riva TTS
- Riva Translation
- Audio2Face-2D Microservice
- Voice Font Microservice
- Resource Downloader
- Audio2Face-3D Authoring
- Before you Begin
- Overview
- Quick start
- WSL Setup Guide
- Architecture overview
- Audio2Face-3D Authoring Microservice
- Sample applications connecting to A2F-3D Authoring Microservice
- gRPC directly with Audio2Face-3D Authoring
- Container Deployment
- AudioClip Database
- Security
- Troubleshooting
- Low-FPS and latency Troubleshooting
- Changelog
- Licenses
- Unreal Renderer Microservice