Adaptable and Real-Time
AI Inference Acceleration

Overview

Optimal Artificial Intelligence Inference from Edge to Cloud

The Vitis™ AI development environment is Xilinx’s development platform for AI inference on Xilinx hardware platforms, including both edge devices and Alveo cards. It consists of optimized IP, tools, libraries, models, and example designs. It is designed with high efficiency and ease of use in mind, unleashing the full potential of AI acceleration on Xilinx FPGA and ACAP.  
 

Vitis AI Deployment Features

How your development works with AI:

  • Supports mainstream frameworks and the latest models capable of diverse deep learning tasks
  • Provides a comprehensive set of pre-optimized models that are ready to deploy on Xilinx devices. You can find the closest model and start re-training for your applications!
  • Provides a powerful open source quantizer that supports pruned and unpruned models quantization, calibration, and fine tuning.
  • The AI profiler provides layer by layer analysis to help with bottlenecks
  • The AI library offers open source high-level C++ and Python APIs for maximum portability from edge to cloud.
  • Efficient and scalable IP cores can be customized to meet your needs of many different applications from a throughput, latency, power and lower-precision perspective.

Explore All the Possibilities with Vitis AI

Vitis AI Model Zoo

Vitis AI Model Zoo

 

AI Optimizer

With world-leading model compression technology, we can reduce model complexity by 5x to 50x with minimal accuracy impact. Deep Compression takes the performance of your AI inference to the next level.

Artificial Intelligence Optimizer Block Diagram

Artificial Intelligence Quantizer Block Diagram

AI Quantizer

By converting the 32-bit floating-point weights and activations to fixed-point like INT8, the AI Quantizer can reduce the computing complexity without losing prediction accuracy. The fixed-point network model requires less memory bandwidth, thus providing faster speed and higher power efficiency than the floating-point model.


AI Compiler

Maps the AI model to a high-efficient instruction set and data flow. Also performs sophisticated optimizations such as layer fusion, instruction scheduling, and reuses on-chip memory as much as possible.

Artificial Intelligence Compiler Block Diagram

AI Profiler

The performance profiler allows programmers to perform an in-depth analysis of the efficiency and utilization of your AI inference implementation.


AI Library

The runtime provides a lightweight set of C++ and Python APIs. enabling easy application development. It also provides efficient task scheduling, memory management, and interrupt handling.

Artificial Intelligence Library Block Diagram
Documentation

Vitis AI Documentation

Default Default Title Document Type Date