Hardware and software for Transformer Model Inference, designed for generative AI systems. High performance, low power consumption, and improved total cost of ownership.
Scalability & efficiency- inference accelerators, designed for high power efficiency, strong performance-per-dollar, and optimized memory bandwidth, key factors for large-scale inference deployment.