
Automotive
Computer vision models help propel self-driving cars and also help recognize signage, pedestrians, and other vehicles to be avoided. Natural-language processing models can help recognize spoken commands to in-car telematics.
AMD EPYC™ 9004 processor-based servers and cloud instances enable fast, efficient AI inference close to your Enterprise data, driving transformative business performance.
Learn how AMD EPYC™ 9004 Series processors empower you to deploy CPU-based AI processing that is efficient for small AI models, many classical ML and inference workloads, plus traditional compute workloads being augmented by AI.
AI Inference uses a trained AI model to make predictions on new data. AMD offers a range of solutions for AI inference depending on your model size and application requirements. AMD EPYC™ CPUs excel for small to medium AI models and workloads where proximity to data matters. AMD Instinct™ GPUs shine for large models and dedicated AI deployments demanding very high performance and scale. Both offer impressive performance and efficiency, letting you choose the right solution for your needs.
Model Size | Type | Pros | Cons | Typical Use Cases |
Small | Classical |
|
|
|
Generative |
|
|
|
Model Size | Type | Pros | Cons | Typical Use Cases |
Medium | Classical |
|
|
|
Predictive |
|
|
|
Model Size | Type | Pros | Cons | Typical Use Cases |
Large | Generative |
|
|
|
Classical |
|
|
|
|
Predictive |
|
|
|
AI models integrated within computer vision, natural language processing, and recommendation systems have significantly impacted businesses across multiple industries. These models help companies recognize objects, classify anomalies, understand written and spoken words, and make recommendations. By accelerating the development of these models, businesses can reap the benefits, regardless of their industry.
Whether deployed as CPU only or used as a host for GPUs executing larger models, AMD EPYC™ 9004 Series processors are designed with the latest open standard technologies to accelerate Enterprise AI inference workloads.
Up to 128 AMD “Zen 4” Cores with AVX-512 instruction support deliver great parallelism for AI inference workloads, reducing the need for GPU acceleration.
Exceptional Power Efficiency: AMD EPYC processors power the most energy efficient servers, delivering exceptional performance and helping reduce energy costs.1
Fast Processing and I/O: 14% generational increase in instructions per clock cycle, DDR5 memory, and PCIe® Gen 5 I/O for fast data processing.2
Framework Support: AMD supports the most popular AI frameworks, including TensorFlow, PyTorch, and ONNX Runtime, covering diverse use cases like image classification and recommendation engines.
Open Source and Compatibility: Optimizations are integrated into popular frameworks offering broad compatibility and open-source upstream friendliness. Plus, AMD is working with Hugging Face to enable their open-source models out of the box with ZenDNN.
ZenDNN Plug-in: These plug-ins accelerate AI inference workloads by optimizing operators, leveraging microkernels, and implementing efficient multithreading on AMD EPYC cores.
As the use of digitization, cloud computing, AI, and other emerging technologies fuel the growth of data, the need for advanced security measures becomes even more pressing. This heightened need for security is further amplified by the increased global emphasis on privacy regulations and severe penalties for breaches, highlighting the unparalleled value of data amid rising security risks.
Built-in at the silicon level, AMD Infinity Guard offers the advanced capabilities required to defend against internal and external threats and help keep your data safe.3