graphics processing unit technology Actions Cite verifiedCite While every effort has been made to follow citation style rules, there may be some discrepancies. Please refer to the appropriate style manual or other sources if you have any questions. Select Citation Style MLA APA Chicago Manual of Style Copy Citation Share Share Share to social media Facebook X URL https://www.britannica.com/technology/graphics-processing-unit Give Feedback Feedback Corrections? Updates? Omissions? Let us know if you have suggestions to improve this article (requires login). Feedback Type Select a type (Required) Factual Correction Spelling/Grammar Correction Link Correction Additional Information Other Your Feedback Submit Feedback Thank you for your feedback Our editors will review what you’ve submitted and determine whether to revise the article.
Print Cite verifiedCite While every effort has been made to follow citation style rules, there may be some discrepancies. Please refer to the appropriate style manual or other sources if you have any questions. Select Citation Style MLA APA Chicago Manual of Style Copy Citation Share Share Share to social media Facebook X URL https://www.britannica.com/technology/graphics-processing-unit Feedback Also known as: GPU Written by Nick Tabor Nick Tabor is a freelance journalist and the author of Africatown: America's Last Slave Ship and the Community It Created. Nick Tabor Fact-checked by The Editors of Encyclopaedia Britannica Encyclopaedia Britannica's editors oversee subject areas in which they have extensive knowledge, whether from years of experience gained by working on that content or via study for an advanced degree. They write new content and verify and edit content received from contributors. The Editors of Encyclopaedia Britannica Last Updated: Aug 19, 2024 • Article History Table of Contents
NVIDIA's GeForce 6600 GPU, released in 2004NVIDIA first released a GPU (and coined the name graphics processing unit) in 1999 with its GeForce 256.(more)graphics processing unit (GPU), electronic circuit board that can quickly perform many mathematical calculations. The technology was originally designed to speed up 3-D graphics rendering. Since its introduction in the 1990s, the graphics processing unit (GPU) has transformed computer software and video games, allowing programmers to produce vivid and realistic images on screens. More recently, GPUs have been used beyond computer graphics in areas including high-performance computing, machine learning, artificial intelligence (AI), weather forecasting, and cryptocurrency mining.
Technology made to render 3-D graphics and video had been used for decades before the GPU was invented. The first electronic stored-program computer, the Small Scale Experimental Machine (or, more informally, “Manchester Baby”) was created in 1948 and displayed images with a cathode-ray tube. The Whirlwind computer, a project funded by the U.S. military to aid with tasks such as aircraft simulation and air traffic control, was built at the Massachusetts Institute of Technology (MIT) between 1948 and 1951 and became the first computer to display video. Video technology continued to advance through the 1960s and ’70s, when personal computers and video games became more ubiquitous. The electronics brand RCA’s “Pixie” chip, which was introduced in 1976, featured a newly improved graphics resolution of 64 by 128 pixels.
Another major advance in chip technology was the invention of Pixel-Planes technology in 1980 at the University of North Carolina. Pixel-Planes was graphics hardware that allocated one processor per pixel, which meant that many parts of the on-screen image were generated simultaneously. Such technology made it possible to produce graphics at much faster speeds.
The popularity of video games helped drive graphics technology in the 1990s. Sony first used the acronym GPU in 1994, when it introduced the first PlayStation consoles; strictly speaking, the technology being referred to was a geometry transformation engine (GTE). Though the technology was still used to perform calculations, it was less powerful than a modern GPU. Many components of the GTE were integrated into the GPU when the technology came out. The 3-D add-in card (which some consider the first modern GPU) was introduced in 1995 by a small company called 3Dlabs. However, GPU was still not a widely known term at the time.
The technology company NVIDIA, under the leadership of Taiwanese American entrepreneurJensen Huang, coined the term graphics processing unit for the launch of the GeForce 256 graphics card in 1999. NVIDIA explained to consumers that such cards could handle intensive graphics functions, taking the strain off a computer’s central processing unit (CPU) and thus allowing for greater processing speeds.
Since the 1990s GPUs have become more powerful and sophisticated as graphics processors but have been incorporated into products other than personal computers and video games. In 2010 NVIDIA and the car manufacturer Audi announced that NVIDIA’s GPUs were being used to power the navigation and entertainment systems in all new Audi vehicles worldwide. Features powered by the new GPUs included full 3-D navigation and the capacity to play two videos simultaneously on different screens.
Get a Britannica Premium subscription and gain access to exclusive content. Subscribe Now Smartphones have also depended on GPUs since they were first released. The original Apple iPhone, made available in 2007, used a PowerVR MBX graphics core, manufactured by the UK company Imagination Technologies.
“The GPU is reducing the time it takes researchers and engineers to analyze, design, and diagnose problems and challenges that in the past would have taken days to weeks, in some cases like protein-folding, maybe months. But is it still a graphics processing engine? Clearly not.”—industry analyst Jon Peddie in a 2018 IEEE Computer Society article
ChatGPT is an example of a large language model (LLM)GPUs are essential in powering LLMs, such as ChatGPT. The artificial intelligence–based technology requires about 30,000 GPUs.(more)In 2006 NVIDIA released Compute Unified Device Architecture (CUDA), a software layer that allows GPUs to process multiple data values in parallel. Developers then began to use the technology for other computer-intensive applications. In 2016, for example, an NVIDIA engineering team used GPU technology to build a self-driving car. Simulations used in molecular dynamics, weather forecasting, and astrophysics can also be performed with GPUs. GPUs have also played a major role in harnessing the benefits of AI, as such technology requires the powerful processing speeds provided by GPUs. ChatGPT, for one, requires about 30,000 of NVIDIA’s powerful GPUs to operate effectively.