Recent from talks
Nothing was collected or created yet.
GDDR5 SDRAM
View on Wikipedia| Type of RAM | |
GDDR5 chips on a Nvidia GeForce GTX 980 Ti | |
| Developer | JEDEC |
|---|---|
| Type | Synchronous dynamic random-access memory |
| Generation | 5th generation |
| Predecessor | GDDR4 SDRAM |
| Successor | GDDR6 SDRAM |
Graphics Double Data Rate 5 Synchronous Dynamic Random-Access Memory (GDDR5 SDRAM) is a type of synchronous graphics random-access memory (SGRAM) with a high bandwidth ("double data rate") interface designed for use in graphics cards, game consoles, and high-performance computing.[1] It is a type of GDDR SDRAM (graphics DDR SDRAM).
Overview
[edit]Like its predecessor, GDDR4, GDDR5 is based on DDR3 SDRAM memory, which has double the data lines compared to DDR2 SDRAM. GDDR5 also uses 8-bit wide prefetch buffers similar to GDDR4 and DDR3 SDRAM.
GDDR5 SGRAM conforms to the standards which were set out in the GDDR5 specification by the JEDEC. SGRAM is single-ported. However, it can open two memory pages at once, which simulates the dual-port nature of other VRAM technologies. It uses an 8N-prefetch architecture and DDR interface to achieve high performance operation and can be configured to operate in ×32 mode or ×16 (clamshell) mode which is detected during device initialization. The GDDR5 interface transfers two 32-bit wide data words per write clock (WCK) cycle to/from the I/O pins. Corresponding to the 8N-prefetch, a single write or read access consists of a 256-bit wide two CK clock cycle data transfer at the internal memory core and eight corresponding 32-bit wide one-half WCK clock cycle data transfers at the I/O pins.
GDDR5 operates with two different clock types. A differential command clock (CK) as a reference for address and command inputs, and a forwarded differential write clock (WCK) as a reference for data reads and writes, that runs at twice the CK frequency. Being more precise, the GDDR5 SGRAM uses a total of three clocks: two write clocks associated with two bytes (WCK01 and WCK23) and a single command clock (CK). Taking a GDDR5 with 5 Gbit/s data rate per pin as an example, the CK runs with 1.25 GHz and both WCK clocks at 2.5 GHz. The CK and WCKs are phase aligned during the initialization and training sequence. This alignment allows read and write access with minimum latency.
A single 32-bit GDDR5 chip has about 67 signal pins and the rest are power and grounds in the 170 BGA package.
Commercialization of GDDR5
[edit]GDDR5 was revealed by Samsung Electronics in July 2007. They announced that they would mass-produce GDDR5 starting in January 2008.[2]
Hynix Semiconductor introduced the industry's first 60 nm class "1 Gb" (10243 bit) GDDR5 memory in 2007.[3] It supported a bandwidth of 20 GB/s on a 32-bit bus, which enables memory configurations of 1 GB at 160 GB/s with only 8 circuits on a 256-bit bus. The following year, in 2008, Hynix bested this technology with its 50 nm class "1 Gb" GDDR5 memory.
In November 2007, Qimonda, a spin-off of Infineon, demonstrated and sampled GDDR5,[4] and released a paper about the technologies behind GDDR5.[5] As of May 10, 2008, Qimonda announced volume production of 512 Mb GDDR5 components rated at 3.6 Gbit/s (900 MHz), 4.0 Gbit/s (1 GHz), and 4.5 Gbit/s (1.125 GHz).[6]
On November 20, 2009, Elpida Memory announced the opening of the company's Munich Design Center, responsible for Graphics DRAM (GDDR) design and engineering. Elpida received GDDR design assets from Qimonda AG in August 2009 after Qimonda's bankruptcy. The design center has approximately 50 employees and is equipped with high-speed memory testing equipment for use in the design, development and evaluation of Graphics memory.[7][8] On July 31, 2013, Elpida became a fully owned subsidiary of Micron Technology and based on current public LinkedIn professional profiles, Micron continues to operate the Graphics Design Center in Munich.[9][10]
Hynix 40 nm class "2 Gb" (2 × 10243 bit) GDDR5 was released in 2010. It operates at 7 GHz effective clock-speed and processes up to 28 GB/s.[11][12] "2 Gb" GDDR5 memory chips will enable graphics cards with 2 GB or more of onboard memory with 224 GB/s or higher peak bandwidth. On June 25, 2008, AMD became the first company to ship products using GDDR5 memory with its Radeon HD 4870 video card series, incorporating Qimonda's 512 Mb memory modules at 3.6 Gbit/s bandwidth.[13][14]
In June 2010, Elpida Memory announced the company's 2 Gb GDDR5 memory solution, which was developed at the company's Munich Design Center. The new chip can work at up to 7 GHz effective clock-speed and will be used in graphics cards and other high bandwidth memory applications.[15]
"4 Gb" (4 × 10243 bit) GDDR5 components became available in the third quarter of 2013. Initially released by Hynix, Micron Technology quickly followed up with their implementation releasing in 2014. On February 20, 2013, it was announced that the PlayStation 4 would use sixteen 4 Gb GDDR5 memory chips for a total of 8 GB of GDDR5 @ 176 Gbit/s (CK 1.375 GHz and WCK 2.75 GHz) as combined system and graphics RAM for use with its AMD-powered system on a chip comprising 8 Jaguar cores, 1152 GCN shader processors and AMD TrueAudio.[16] Product teardowns later confirmed the implementation of 4 Gb based GDDR5 memory in the PlayStation 4.[17][18]
In February 2014, as a result of its acquisition of Elpida, Micron Technology added 2 Gb and 4 Gb GDDR5 products into the company's portfolio of graphics memory solutions.[19]
As of January 15, 2015, Samsung announced in a press release that it had begun mass production of "8 Gb" (8 × 10243 bits) GDDR5 memory chips based on a 20 nm fabrication process. To meet the demand of higher resolution displays (such as 4K) becoming more mainstream, higher density chips are required in order to facilitate larger frame buffers for graphically intensive computation, namely PC gaming and other 3D rendering. Increased bandwidth of the new high-density modules equates to 8 Gbit/s per pin × 170 pins on the BGA package x 32-bits per I/O cycle, or 256 Gbit/s effective bandwidth per chip.[20]
On January 6, 2015, Micron Technology President Mark Adams announced the successful sampling of 8 Gb GDDR5 on the company's fiscal Q1-2015 earnings call.[21][22] The company then announced, on January 25, 2015, that it had begun commercial shipments of GDDR5 using a 20 nm process technology.[23][24][25] The formal announcement of Micron's 8 Gb GDDR5 appeared in the form of a blog post Archived 2015-09-07 at the Wayback Machine by Kristopher Kido on the company's website September 1, 2015.[26][27]
GDDR5X
[edit]In January 2016, JEDEC standardized GDDR5X SGRAM.[28] GDDR5X targets a transfer rate of 10 to 14 Gbit/s per pin, twice that of GDDR5.[29] Essentially, it provides the memory controller the option to use either a double data rate mode that has a prefetch of 8n, or a quad data rate mode that has a prefetch of 16n.[30] GDDR5 only has a double data rate mode that has an 8n prefetch.[31] GDDR5X also uses 190 pins per chip (190 BGA).[30] By comparison, standard GDDR5 has 170 pins per chip; (170 BGA).[31] It therefore requires a modified PCB. QDR (quad data rate) may be used in reference to the write command clock (WCK) and ODR (Octal Data Rate) in reference to the command clock (CK).[32]
GDDR5X commercialization
[edit]
Micron Technology began sampling GDDR5X chips in March 2016,[33] and began mass production in May 2016.[34]
Nvidia officially announced the first graphics card using GDDR5X, the Pascal-based GeForce GTX 1080 on May 6, 2016.[35] Later, the second graphics card to use GDDR5X, the Nvidia Titan X (Pascal) on July 21, 2016,[36] the GeForce GTX 1080 Ti on February 28, 2017,[37] and Nvidia Titan Xp on April 6, 2017.[38]
See also
[edit]References
[edit]- ^ Micron TN-ED-01: GDDR5 SGRAM Introduction. Archived 2015-09-18 at the Wayback Machine Accessed April 11, 2014
- ^ Pancescu, Alexandru (July 18, 2007). "Samsung Pushes The GDDR5 Standard Forward". Softpedia. Retrieved 18 September 2019.
- ^ "History: 2000s". SK Hynix. Archived from the original on 6 August 2020. Retrieved 8 July 2019.
- ^ Register report Archived 2008-07-06 at the Wayback Machine. Retrieved November 2, 2007.
- ^ Qimonda GDDR5 Archived 2016-08-26 at the Wayback Machine White Paper
- ^ GDDR5 in Production, New Round of Graphics Cards War Imminent., retrieved May 11, 2008
- ^ Topalov, Milan. "Elpida officially opens Munich Design Center". www.fabtech.org. Archived from the original on 2016-01-17. Retrieved 2015-09-09.
- ^ "Elpida Opens High Speed DRAM Test Laboratory at Munich Design Center | Business Wire". www.businesswire.com. Retrieved 2015-09-09.
- ^ "Micron (MU) Completes Elpida Memory, Rexchip Purchases". Retrieved 2015-09-09.
- ^ "Markus Balb | LinkedIn".
- ^ Hynix 1H '11 Product Catalog, page 8. Archived 2014-03-13 at the Wayback Machine Accessed July 24, 2014.
- ^ Hynix H5GQ2H24AFR Product Overview. Archived 2014-07-23 at the Wayback Machine Accessed July 24, 2014.
- ^ Qimonda Press Release. May 21, 2008 Archived September 16, 2008, at the Wayback Machine
- ^ AMD Press Release. June 25, 2008
- ^ Pop, Sebastian. "Elpida Starts Making GDDR5 Graphics Memory, Delivers 2Gb Chip". Retrieved 2015-09-09.
- ^ "Interview with PS4 system architect". 2013-04-01.
- ^ "PlayStation 4 Teardown". Retrieved 2015-09-09.
- ^ teardown.com. "Sony PlayStation 4 Teardown : Board & Chip Shots and Images (Digital Home Teardown)". www.techinsights.com. Archived from the original on 2015-10-02. Retrieved 2015-09-09.
- ^ "Micron Technology, Inc.—GDDR5 | DRAM". www.micron.com. Archived from the original on 2016-03-20. Retrieved 2016-09-06.
- ^ "Samsung Electronics Starts Mass Producing Industry's First 8-Gigabit Graphics DRAM (GDDR5)". 2015-01-15.
- ^ "Micron Technology's (MU) CEO Mark Durcan on Q1 2015 Results—Earnings Call Transcript". Seeking Alpha. Retrieved 2015-09-09.
- ^ "Micron: We are sampling 8Gb GDDR5 for 8GB graphics cards". Retrieved 2015-09-09.
- ^ "Micron Technology's (MU) CEO Mark Durcan on Q3 2015 Results—Earnings Call Transcript". Seeking Alpha. Retrieved 2015-09-09.
- ^ "Micron begins commercial shipments of 20nm GDDR5 chips". Retrieved 2015-09-09.
- ^ "Micron delivers GDDR5 memory on 20 nm". www.hitechreview.com. Retrieved 2015-09-09.
- ^ "Micron Starts Shipping 8Gb GDDR5 Memory For Next Generation Graphics Cards | HotHardware". Retrieved 2015-09-09.
- ^ "Micron Technology, Inc.—Next-Gen Graphics Products Get Extreme Speed From Latest Graphics Memory Solutions". www.micron.com. Archived from the original on 2015-09-07. Retrieved 2015-09-09.
- ^ "JEDEC Announces Publication of GDDR5X Graphics Memory Standard". JEDEC. 2016-01-26. Retrieved 2016-02-10.
- ^ "JEDEC Publishes GDDR5X Specifications – Double the Bandwidth of GDDR5 With Lowered Power Consumption". Retrieved 6 June 2016.
- ^ a b "GDDR5X SGRAM: MT58K256M32 – 16 Meg x 32 I/O x 16 banks, 32 Meg x 16 I/O x 16 banks" (PDF). Micron Technology. May 2016. Archived from the original (PDF) on February 7, 2017. Retrieved May 29, 2016.
- ^ a b "GDDR5 SGRAM: MT51J256M32 – 16 Meg x 32 I/O x 16 banks, 32 Meg x 16 I/O x 16 banks" (PDF). Micron Technology. November 2015. Archived from the original (PDF) on February 7, 2017. Retrieved May 29, 2016.
- ^ Smith, Ryan. "Micron Spills on GDDR6X: PAM4 Signaling For Higher Rates, Coming to NVIDIA's RTX 3090". www.anandtech.com. Archived from the original on August 14, 2020.
- ^ Shilov, Anton (March 29, 2016). "Micron Begins to Sample GDDR5X Memory, Unveils Specs of Chips". AnandTech. Archived from the original on March 30, 2016. Retrieved 16 July 2019.
- ^ Shilov, Anton (May 12, 2016). "Micron Confirms Mass Production of GDDR5X Memory". AnandTech. Archived from the original on May 13, 2016. Retrieved 16 July 2019.
- ^ Newsroom, NVIDIA. "A Quantum Leap in Gaming: NVIDIA Introduces GeForce GTX 1080". NVIDIA Newsroom Newsroom.
{{cite web}}:|last=has generic name (help) - ^ "The New NVIDIA TITAN X: The Ultimate. Period. - The Official NVIDIA Blog". nvidia.com. 21 July 2016.
- ^ Newsroom, NVIDIA. "NVIDIA Introduces the Beastly GeForce GTX 1080 Ti -- Fastest Gaming GPU Ever". NVIDIA Newsroom Newsroom.
{{cite web}}:|last=has generic name (help) - ^ "The New Titan Is Here: NVIDIA TITAN Xp - NVIDIA Blog". nvidia.com. 6 April 2017.
External links
[edit]GDDR5 SDRAM
View on GrokipediaIntroduction
Overview
GDDR5 SDRAM, or Graphics Double Data Rate 5 Synchronous Graphics Random Access Memory, is a specialized type of dynamic random-access memory designed for high-bandwidth applications in graphics processing units (GPUs), including graphics cards, game consoles, and high-performance computing systems. It operates as a single-ported SGRAM with a double data rate interface, enabling simultaneous data transfers on both rising and falling clock edges to achieve efficient throughput for graphics-intensive workloads.[5][6] Developed as the successor to GDDR3 SDRAM (with GDDR4 seeing limited adoption), GDDR5 prioritizes enhanced bandwidth and power efficiency to meet the escalating demands of visual rendering and parallel computing tasks, while maintaining compatibility with existing GPU architectures. Its core design incorporates an 8n-prefetch architecture, which prefetches eight times the I/O data width, allowing for burst transfers of eight 32-bit words (256 bits total) in a single access cycle for x32 configurations and thereby optimizing data delivery to the GPU core.[6][7] The GDDR5 standard is defined and maintained by the JEDEC Solid State Technology Association, which establishes the functional, electrical, and timing requirements to promote interoperability among semiconductor manufacturers. This standardization ensures reliable performance across diverse implementations. An extension known as GDDR5X builds on GDDR5 to deliver further performance gains in ultra-high-end graphics scenarios.[8]Historical Development
The development of GDDR5 SDRAM emerged as a response to the limitations of its predecessor, GDDR4, which offered only marginal improvements in bandwidth over GDDR3 and lacked robust on-die error correction mechanisms necessary for sustaining higher data rates in graphics applications.[9] GDDR4's short-lived adoption highlighted the need for a more substantial leap in performance, prompting the industry to prioritize GDDR5 to address escalating demands for video memory throughput without compromising reliability at elevated speeds.[10] In early 2007, Qimonda, a key player in memory technology, announced its focus on GDDR5, bypassing GDDR4 entirely, and actively contributed to the JEDEC standardization process, with expectations for the standard's finalization by summer 2007.[10] The JEDEC committee formalized the GDDR5 specification (JESD212) in December 2009, building upon DDR3 architecture but optimizing it for graphics workloads through enhanced prefetch buffering and signaling tailored to high-bandwidth needs. Samsung Electronics led early prototyping efforts, revealing initial GDDR5 developments in July 2007 as the first to produce functional prototypes, setting the stage for subsequent industry advancements.[3] Key milestones followed rapidly in late 2007. Qimonda began sampling 512 Mb GDDR5 chips in November, demonstrating the technology's viability ahead of broader standardization.[11] Shortly thereafter, Hynix Semiconductor introduced the industry's first 1 Gb GDDR5 device using a 66 nm process, capable of delivering up to 20 GB/s of bandwidth per chip, which enabled processing of over 20 hours of DVD-quality video in real time.[12] These prototypes underscored GDDR5's potential for superior graphics performance. Qimonda continued contributing to refinements in the standard until its bankruptcy in 2009, after which its assets, including GDDR-related patents, were acquired by competitors like Elpida.[13] GDDR5 made its commercial debut in the AMD Radeon HD 4870 GPU, launched on June 25, 2008.[4]Technical Specifications
Architecture and Organization
GDDR5 SDRAM features a single-ported design, in which each bank supports only one active row at a time, but simulates dual-port behavior for concurrent read and write operations by utilizing its multi-bank structure to allow independent accesses across banks. This approach enables efficient handling of graphics workloads that frequently require simultaneous data retrieval and updates without blocking operations in a single bank. The architecture is adapted from DDR3 SDRAM but optimized for the high-bandwidth demands of graphics processing, emphasizing parallel bank access over general-purpose computing efficiency.[5] The memory supports configurable data widths of ×32 or ×16 modes to accommodate various graphics card configurations, with 32-bit data transfers per write clock (WCK) cycle in the standard ×32 mode to maximize throughput. Chips are organized in a row-and-column array within banks, facilitating burst accesses typical of graphics rendering tasks where large blocks of data are fetched or stored sequentially. This organization prioritizes quick row activation and column multiplexing to support the prefetch mechanisms inherent to GDDR5's double data rate interface.[14] Available memory densities range from 512 Mb to 8 Gb per chip, structured hierarchically with rows, columns, and banks to scale capacity while maintaining access speed. For instance, a 512 Mb device is typically arranged as 2M rows × 32 columns × 8 banks in ×32 mode, while larger 1 Gb devices expand to 16 banks for improved concurrency. The bank architecture consists of 8 or 16 banks per die, often divided into four bank groups to reduce inter-bank conflicts and enable parallel operations suited to the random access patterns in graphics applications.[5] GDDR5 includes CRC-8 for error detection on read/write operations at the interface level, which triggers retries for detected transmission errors, ensuring data integrity. This complements the error detection code (EDC) mechanism without the overhead of full system-level ECC.[15]Interface and Signaling
GDDR5 SDRAM utilizes a dual-clock system to manage the timing of commands, addresses, and data transfers efficiently at high speeds. The command/address clock (CK), implemented as a differential pair (CK_t and CK_c), operates at half the effective data rate, with commands registered on the rising edge in single data rate (SDR) mode and addresses captured on both rising and falling edges in double data rate (DDR) mode. Complementing this, the write clock (WCK), also a differential pair (WCK_t and WCK_c), runs at the full data rate—nominally twice the frequency of CK—and functions as a forwarded strobe from the memory controller specifically for write operations to ensure precise data alignment. This architecture decouples command timing from data strobe requirements, enabling robust operation up to data rates of 8 Gbps per pin.[7][16] The physical interface of GDDR5 chips is standardized in a 170-ball fine-pitch ball grid array (FBGA) package, typically measuring 12 mm × 14 mm, which supports direct attachment to graphics processing unit substrates without intermediate modules. This compact, lead-free package employs an outer data, inner control (ODIC) pinout, dividing the 32-bit data bus into four bytes across quadrants for optimized signal routing and reduced crosstalk. Signaling employs true differential pairs for the CK and WCK clocks to suppress common-mode noise and improve signal integrity at high frequencies. In contrast, data signals (DQ) use pseudo-differential signaling via pseudo-open drain (POD) drivers with on-die termination (ODT), featuring nominal 60 Ω and 120 Ω impedances calibrated via the ZQ pin and terminated to the supply voltage (V_DDQ), which enhances eye opening and reduces reflections without requiring fully differential data lines.[7][16] Read and write protocols in GDDR5 leverage source-synchronous timing to maintain synchronization between the memory controller and device. During read operations, data outputs are aligned to the CK clock edges in a source-synchronous manner, allowing the controller to capture data using the forwarded clock for minimal skew. Write operations, however, rely exclusively on the WCK clock as a data strobe, where the controller generates and forwards WCK to clock incoming DQ signals into the device, ensuring accurate latching independent of CK variations. This protocol supports the inherent 8n-prefetch buffer architecture, facilitating fixed burst lengths of eight words to match graphics workload patterns and contribute to overall bandwidth without introducing variable latency complexities.[7][16] Voltage specifications for GDDR5 emphasize compatibility with high-performance graphics while offering flexibility for power-sensitive designs. The core and I/O supplies (V_DD and V_DDQ) operate at a nominal 1.5 V ±3%, providing the drive strength needed for data rates up to 7 Gbps. An optional low-voltage mode at 1.35 V ±3% is supported through dynamic voltage scaling, allowing reduced power draw in applications where maximum speed is not required, while maintaining full functional compatibility.[7][16]Key Operational Features
GDDR5 SDRAM incorporates an 8n-prefetch buffer architecture that enables the retrieval of 8 words—equivalent to 256 bits in ×32 device configurations—per single command, effectively doubling the data throughput relative to 4n-prefetch designs in prior graphics memory standards. This prefetch mechanism aligns with the double data rate (DDR) interface, where data is transferred on both rising and falling edges of the forwarded clock (WCK), optimizing burst transfers for high-bandwidth graphics workloads. By prefetching multiple words in advance, the architecture reduces latency in accessing sequential data, enhancing overall system efficiency in GPU memory controllers. To ensure robust signal integrity at high speeds, GDDR5 supports write-leveling and read-training modes during device initialization, allowing calibration of data strobe (DQS) timing relative to the clock edges. Write-leveling adjusts the controller's output timing to align with the memory device's input window, while read-training fine-tunes sampling points to center data eyes and mitigate skew across the bus. These modes, activated via specific mode register settings, are essential for reliable operation over point-to-point connections in graphics cards, where trace lengths and loading can introduce timing variations. Dynamic on-die termination (ODT) in GDDR5 minimizes signal reflections by dynamically enabling termination resistors at the memory device ends during read and write operations, configurable through mode registers for nominal, write, and park values. This feature adapts termination strength based on the transaction type—such as RTT_NOM for reads or RTT_WR for writes—reducing crosstalk and improving eye quality on the high-speed data bus without requiring external components. By supporting dynamic ODT, GDDR5 maintains signal quality across varying bus configurations, critical for multi-device topologies in graphics subsystems. GDDR5 SDRAM provides a fixed burst length of 8 cycles, tailored to the access patterns common in graphics processing, such as texture mapping and frame buffer updates in GPUs. This burst length maximizes throughput for sequential data streams, leveraging the prefetch buffer to deliver 32 bytes per operation in ×32 mode.[5] For enhanced reliability under thermal stress, GDDR5 includes temperature-compensated self-refresh (TCSR), which adjusts the internal refresh rate according to detected operating temperature ranges to prevent data corruption while minimizing power consumption. Enabled through mode register configuration, TCSR divides the temperature spectrum into bins—typically normal, extended, and high—with corresponding refresh intervals that lengthen at lower temperatures where retention times are longer. This feature ensures stable operation in thermally variable environments like high-performance graphics cards, without external temperature sensors.Performance Characteristics
Data Rates and Bandwidth
GDDR5 SDRAM supports per-pin data rates ranging from an initial 3.6 Gbit/s in 2008 to a maximum of 8 Gbit/s by 2015, enabling significant throughput improvements for graphics applications.[17][18] The effective bandwidth of a GDDR5 memory module is calculated as (data rate per pin in Gbit/s × total data pins across chips) / 8, yielding results in GB/s. For instance, a 1 Gbit/s per-pin rate on a 256-bit bus—typically implemented with eight 32-bit chips—delivers 32 GB/s.[19] Data rates evolved progressively with revisions, tied to the quarter-rate clock (CK) frequency, where the effective rate equals four times the CK due to the quad data rate signaling on the write clock (WCK). Early implementations reached 4 Gbit/s at 1 GHz CK, advancing to 5 Gbit/s at 1.25 GHz CK and 6 Gbit/s at 1.5 GHz CK, with later variants achieving 7–8 Gbit/s through refined timing and signaling optimizations.[7][20]| Data Rate (Gbit/s) | CK Frequency (GHz) | Example Year/Implementation |
|---|---|---|
| 4 | 1.0 | 2009 early GPUs |
| 5 | 1.25 | 2010 mainstream |
| 6 | 1.5 | 2012 revisions |
| 7–8 | 1.75–2.0 | 2015 high-end |
