Graphics processing units (GPUs) are massively parallel processors originally designed for rendering graphics that are now widely used to accelerate AI and machine learning inference workloads. For inference, GPUs execute large numbers of matrix and tensor operations concurrently, dramatically reducing latency and increasing throughput versus general‑purpose CPUs. They matter because they underpin most production-scale deep learning services, from recommendation systems to generative AI, enabling cost-effective, high-performance deployment of trained models.
by NVIDIA (dominant ecosystem vendor; GPUs also produced by AMD, Intel, others)BigTech
GPUs are sold as hardware devices (cards, servers, appliances) with pricing varying by model and vendor; cloud providers offer GPU instances billed per second/hour. Additional software (e.g., enterprise support, management suites) may be licensed separately. Pricing is not standardized and depends on configuration, volume, and contracts.
No use cases found for this technology.
Browse all technologies