← Back to Library

ISSCC 2026: Nvidia & Broadcom CPO, HBM4 & LPDDR6, TSMC active LSI, logic-based SRAM, UCIe-S and more

Dylan Patel has delivered a rare piece of semiconductor journalism that cuts through the usual hype to reveal the gritty engineering realities shaping the next decade of AI infrastructure. While most coverage fixates on marketing slides, this analysis of ISSCC 2026 exposes how memory vendors are fundamentally rewriting the rules of physics to keep pace with computational demands. The most startling revelation isn't just that speeds are increasing, but that the very architecture of memory stacks is being dismantled and rebuilt to solve problems that seemed unsolvable just a year ago.

The Architecture of Speed

Patel identifies a seismic shift in how High Bandwidth Memory (HBM) is constructed, moving away from the monolithic approaches of the past. "The most obvious architectural change from HBM3E to HBM4 is the process technology split between the core DRAM dies and the base die," Patel writes, noting that Samsung is now using advanced logic nodes for the base while keeping the core on DRAM processes. This is not a minor tweak; it is a strategic pivot that allows for higher transistor density and significantly lower power consumption. By decoupling these layers, the industry can finally address the thermal and electrical bottlenecks that have plagued AI accelerators.

ISSCC 2026: Nvidia & Broadcom CPO, HBM4 & LPDDR6, TSMC active LSI, logic-based SRAM, UCIe-S and more

The data presented supports a bold claim: Samsung's new HBM4 stack is not merely competitive but, in specific metrics, superior to the current market leader. "Samsung demonstrated a 36 GB, 12-high HBM4 stack featuring 2048 IO pins and 3.3 TB/s of bandwidth," Patel notes, highlighting that this performance significantly exceeds the official industry standard. The implication is clear: the JEDEC standard, which sets the baseline for interoperability, is already being outpaced by the cutting edge of R&D. This creates a fascinating dynamic where the "standard" becomes a floor rather than a ceiling, forcing the entire ecosystem to accelerate.

Samsung's implementation significantly exceeds the baseline specification of the official JEDEC HBM4 standard, reaching 13 Gb/s per pin and delivering 3.3 TB/s of bandwidth.

However, this aggressive performance comes with a significant caveat that Patel does not shy away from. The move to advanced logic nodes for the base die introduces cost and yield challenges. "The 1c front-end manufacturing process has proved challenging for Samsung throughout 2025... Front-end yields for the 1c node were only around 50% last year," Patel explains. Critics might argue that prioritizing raw speed over yield stability is a dangerous gamble, especially when the alternative competitors are using more mature, lower-cost processes. Yet, Patel's framing suggests that in the AI arms race, performance is the only currency that matters, even if it erodes margins in the short term.

Solving the Timing Nightmare

As memory stacks grow taller and channel counts double, the physical reality of signal propagation becomes a major hurdle. Patel dives deep into the concept of "tCCDR," the minimum interval required between consecutive read commands, which directly impacts how fast an AI model can access data. "As the stack heights and channel counts increase, the variation between the dies accumulates, causing larger timing mismatches across channels and dies," Patel writes. This is a profound engineering challenge; without a solution, the theoretical bandwidth of the memory would be useless due to synchronization errors.

Samsung's response is a sophisticated calibration system that Patel describes as a game-changer. "To address this issue, Samsung introduces a 'per-channel TSV RDQS timing auto-calibration scheme,'" he explains. This system measures delay variations and compensates for them in real-time, effectively aligning the timing across the entire stack. The result is a tangible performance boost: "This scheme alone increased data rates from 7.8 Gb/s to 9.4 Gb/s." This level of granular detail is what separates this commentary from generic industry reports. It moves beyond the "what" to explain the "how," showing that the gains are not magic but the result of solving specific, complex timing problems.

The article also touches on the broader context of testing, noting that traditional methods are insufficient for these new architectures. "By moving the base die to Samsung Foundry's SF4 logic process, Samsung enables a fully programmable testing framework capable of running complex test algorithms," Patel argues. This shift from fixed-pattern testing to programmable, real-world simulation is critical for yield learning. It mirrors the evolution seen in through-silicon via (TSV) technology, where early adoption required similar leaps in testing methodology to ensure reliability in 3D stacked chips. Without these advanced testing capabilities, the yield penalties Patel mentioned earlier would be catastrophic.

The Mobile Memory Revolution

While HBM grabs the headlines for AI, Patel ensures the reader understands that the mobile market is undergoing a similar transformation with LPDDR6. The analysis of Samsung's LPDDR6 chip reveals a clever trade-off between power and area. "LPDDR6 adopts a 2 sub-channel per die architecture... However, there is a latency penalty for accessing data in the secondary sub-channel," Patel writes. This dual-channel design doubles the peripheral circuitry, leading to a "5% of the total die area" penalty. For a busy executive, this is a crucial data point: the industry is willing to sacrifice silicon real estate to gain the power efficiency required for next-generation mobile devices.

The power-saving techniques described are equally impressive. "By carefully choosing which peripheral logic is using which power domain, read power has been reduced by 27% and write power reduced by 22%," Patel notes. This is not just about battery life; it is about thermal management in increasingly dense devices. The use of wide NRZ signaling and Data Bus Inversion (DBI) further illustrates the complexity of modern memory design. "DBI is a power-saving and signal integrity mechanism... It limits the maximum number of simultaneous switching outputs to half the bus width, reducing power consumption and supply noise," Patel explains.

This approach represents a notable departure from HBM3E, enabling much more robust debugging and better yield learning for HBM stacks.

Patel's analysis of the density metrics offers a sobering counterpoint to the speed gains. "The memory density described leads us to believe that this prototype LPDDR6 chip was manufactured on their 1b process," he suggests, noting that the density is actually lower than previous generations. This raises a valid question: are we trading density for speed and power efficiency? A counterargument worth considering is that this trade-off might limit the scalability of mobile memory in the long run, forcing device manufacturers to choose between larger chips or lower capacities. However, given the immediate demand for power efficiency in edge AI, this seems a calculated risk.

Bottom Line

Patel's coverage of ISSCC 2026 succeeds by refusing to treat semiconductor advancements as black boxes, instead illuminating the specific architectural choices that will define the next generation of computing. The strongest part of the argument is the detailed breakdown of how splitting the base die process is enabling performance leaps that defy current standards. The biggest vulnerability remains the yield and cost implications of these aggressive manufacturing strategies, which could delay mass adoption. Readers should watch closely to see if Samsung's high-risk, high-reward approach forces competitors to follow suit or if the market corrects toward more cost-effective solutions.

Deep Dives

Explore these related deep dives:

  • Through-silicon via

    Understanding the physical limitations and thermal challenges of TSVs is essential to grasping why Samsung's move to a logic-based base die in HBM4 represents a critical architectural shift rather than just a process node upgrade.

  • Shmoo plot

    The article mentions 'Shmoo Plot' as a key data visualization; knowing this specific engineering tool reveals how manufacturers rigorously map the voltage-frequency boundaries that determine if a chip like HBM4 can actually meet the strict power constraints of the Rubin platform.

Sources

ISSCC 2026: Nvidia & Broadcom CPO, HBM4 & LPDDR6, TSMC active LSI, logic-based SRAM, UCIe-S and more

by Dylan Patel · SemiAnalysis · Read full article

There are three major semiconductor conferences each year, IEDM, VLSI and finally ISSCC. We have covered the former two in great detail over the past few years. Today, we finally complete the trinity with our roundup on ISSCC 2026.

Compared to IEDM and VLSI, ISSCC has a much bigger focus on integration and circuits. Almost every paper comes with some form of circuit diagram, together with clear measurements and data.

In past years, ISSCC findings have been hit or miss when it comes to industry impact. This year was different, a significant number of papers and presentations were directly relevant to market trends. Topics covered range from the latest advancements in HBM4, LPDDR6, GDDR7, and NAND, to co-packaged optics, advanced die-to-die interfaces, and advanced processors from the likes of MediaTek, AMD, Nvidia, and Microsoft.

In this roundup, we will cover major categories such as Memory, Optical Networking, High-Speed Electrical Interconnects, Processors.

Memory.

One key theme that caught our attention at this year’s ISSCC was memory, including Samsung HBM4, Samsung and SK Hynix LPDDR6, and SK Hynix GDDR7. Other than DRAM, logic-based SRAM and MRAM also piqued our interest.

Samsung HBM4 - Paper 15.6.

Samsung was the only one among the top three memory vendors to present a technical paper on HBM4. Before ISSCC, we noted in our Accelerator & HBM model that Samsung had made great improvements in their HBM4 generation over their HBM3E. The data presented at ISSCC confirmed our analysis, with Samsung posting best-in-class performance - we have also detailed this development months ago, in a model update note.The technical details presented at ISSCC, combined with industry chatter we have gathered, clearly demonstrate that Samsung’s HBM4 is competitive with its peers. Notably, it can meet the pin speed required for Rubin while staying below 1V. While Samsung still lags SK Hynix in terms of reliability and stability, the company has made meaningful progress in closing the gap on the technology front and could challenge SK Hynix’s dominance in HBM. Their 1c-based HBM4 paired with an SF4 logic base die appears to deliver stronger performance in pin speed.

Samsung demonstrated a 36 GB, 12-high HBM4 stack featuring 2048 IO pins and 3.3 TB/s of bandwidth, built using 6th-generation 10nm-class (1c) DRAM core dies paired with an SF4 logic base die.

The most obvious architectural change from HBM3E to HBM4 is the process technology split between the core DRAM dies ...