ISPD ’25 TOC

Proceedings of the 2025 International Symposium on Physical Design

 Full Citation in the ACM Digital Library

SESSION: Session 1: Opening Session and First Keynote

Towards Designing and Deploying Ising Machines

  • Sachin S. Sapatnekar

Today, NP-complete or NP-hard combinatorial problems are often solved on classical computers, using heuristics with no optimality guarantees or approximation algorithms with loose optimality bounds. Ising computation provides a new paradigm for solving these problems using networks of coupled oscillators. In contrast with traditional Ising machines that use supercooled chips, recent approaches have proposed the use of coupled CMOS ring oscillators, reducing the power dissipation of these systems by several orders of magnitude. This talk will overview the Ising model, discuss the challenges of building CMOS Ising machines, including issues related to layout and timing, and point to directions that are helping deploy these methods to solve ever-larger combinatorial problems.

SESSION: Session 2: Placement and DTCO

GOALPlace: Begin with the End in Mind

  • Anthony Agnesina
  • Rongjian Liang
  • Geraldo Pradipta
  • Anand Rajaram
  • Haoxing Ren

Co-optimizing placement with congestion is integral to achieving high-quality designs. This paper presents GOALPlace, a learning-based approach to improving placement congestion by controlling cell density. It efficiently learns from an EDA tool’s post-route optimized results and uses an empirical Bayes technique to adapt the target to a specific placer’s solutions, effectively beginning with the end in mind. Our method enhances correlation with the tool’s router and timing-opt engine, while solving placement globally without expensive incremental congestion estimation and mitigation methods. A statistical analysis with hierarchical netlist clustering establishes the importance of density and the potential for an adequate cell density target across placements. Our experiments show that our method, when integrated into an academic GPU-accelerated global placer, consistently produces macro and standard cell placements that match or exceed the quality of commercial tools. Our empirical Bayes methodology also shows a substantial quality improvement over leading academic mixed-size placers, achieving up to 10× fewer design rule check (DRC) violations, a 5% decrease in wirelength, and a 30% and 60% reduction in worst and total negative slack (WNS/TNS).

Invited: Scaling Standard Cell Layout Using Track Height Compression and Design Technology Co-optimization

  • Chung-Kuan Cheng
  • Byeonggon Kang
  • Bill Lin
  • Yucheng Wang

Moore’s law scaling is approaching physical limits, as indicated by the technology roadmap. Recent standard cell layout reductions rely on track height compression, which increases pin density and routing congestion. To address these challenges, design technology co-optimization (DTCO) was introduced. This paper explores how much track height can be compressed and how DTCO features can sustain layout scaling. To support this exploration, we developed an SMT-based cell synthesis tool that integrates gear ratio, M1 metal grid offset, local-interconnect source-drain (LISD) merging, adjustable gate cut lengths, and double-height architecture with pass-throughs, and various power delivery options.

In our exploration, we compress horizontal track numbers from four to two. Our synthesis tool enables flexible gear ratio options through a graph-based data structure that allocates vertical rout- ing resources at a smaller pitch than the contacted poly pitch. For double-height architecture, we allow pass-through options to alleviate routing congestion. The empirical results identify critical design strategies to meet scaling demands and overcome pin density challenges. Overall, the study demonstrates the design options that lead to potential scaling capability in the near future.

Invited: Physical Design Challenges for Design Technology Co-optimization

  • Taewhan Kim

Design technology co-optimization (DTCO) is the process of optimizing design and process technology together to enhance performance, power efficiency, chip utilization, and manufacturing cost/yield. Through DTCO, we are able to evaluate technologies, design rules, and cell architectures using block-level PPA (performance, power, area) analysis, which greatly helps semiconductor fabs reduce cost and shorten time-to-market in advanced process development with substantial architectural innovation.

The parameters that DTCO targets to evaluate include design rules (e.g., gate poly pitch, M1 pitch, side-to-top spacing rule, via-enclosure rule), cell architectures (e.g., single-row or multi-row-height, 2D or 1D M1, preference of metal direction), and technologies (e.g., Fin-FET, Nanosheet-FET, Complementary-FET), which are collectively called DTCO parameters.

LEGALM: Efficient <u>Leg</u>alization for Mixed-Cell-Height Circuits with Linearized <u>A</u>ugmented <u>L</u>agrangian <u>M</u>ethod

  • Jing Mai
  • Chunyuan Zhao
  • Zuodong Zhang
  • Zhixiong Di
  • Yibo Lin
  • Runsheng Wang
  • Ru Huang

Advanced technologies increasingly adopt mixed-cell-height circuits due to their superior power efficiency, compact area usage, enhanced routability, and improved performance. However, the complex constraints of modern circuit design, including routing challenges and fence region constraints, increase the difficulty of mixed-cell-height legalization. In this paper, we introduce LEGALM, a state-of-the-art mixed-cell-height legalizer that can address routability and fence region constraints more efficiently. We propose an augmented Lagrangian formulation coupled with a block gradient descent method that offers a novel analytical perspective on the mixed-cell-height legalization problem. To further enhance efficiency, we develop a series of GPU-accelerated kernels and a triplefold partitioning technique with minor quality overhead. Experimental results on ICCAD-2017 and modified ISPD-2015 benchmarks show that our approach significantly outperforms current state-of-the-art legalization algorithms in both quality and efficiency.

SESSION: Session 3: Acceleration

Cypress: VLSI-Inspired PCB Placement with GPU Acceleration

  • Niansong Zhang
  • Anthony Agnesina
  • Noor Shbat
  • Yuval Leader
  • Zhiru Zhang
  • Haoxing Ren

The scale of printed circuit board (PCB) designs has increased significantly, with modern commercial designs featuring more than 10,000 components. However, the placement process heavily relies on manual efforts that take weeks to complete, highlighting the need for automated PCB placement methods. The challenges of PCB placement arise from its flexible design space and limited routing resources. Existing automated PCB placement tools have achieved limited success in quality and scalability. In contrast, very large-scale integration (VLSI) placement methods have proven to be scalable for designs with millions of cells and delivering high-quality results. Therefore, we propose Cypress, a scalable, GPU-accelerated PCB placement method inspired by VLSI. It incorporates tailored cost functions, constraint handling, and optimized techniques adapted for PCB layouts. In addition, there is an increasing demand for realistic and open-source benchmarks to (1) enable meaningful comparisons between tools and (2) establish performance baselines to track progress in PCB placement technology. To address this gap, we present a PCB benchmark suite synthesized from real commercial designs. We evaluate our method against state-of-the-art commercial and academic PCB placement tools with the benchmark suite. Our approach demonstrates a 1-5.9X higher routability on the proposed benchmarks. For fully routed designs, Cypress achieves 1-19.7X shorter routed track lengths. With GPU acceleration, Cypress delivers up to 492.3X speedup in run time. Finally, we demonstrate scalability to real commercial designs, a capability unmatched by existing tools.

GPU-Accelerated Inverse Lithography Towards High Quality Curvy Mask Generation

  • Haoyu Yang
  • Haoxing Ren

Inverse Lithography Technology (ILT) has emerged as a promising solution for photo mask design and optimization. Relying on multi-beam mask writers, ILT enables the creation of free-form curvilinear mask shapes that enhance printed wafer image quality and process window. However, a major challenge in implementing curvilinear ILT for large-scale production is mask rule checking, an area currently under development by foundries and EDA vendors. Although recent research has incorporated mask complexity into the optimization process, much of it focuses on reducing e-beam shots, which does not align with the goals of curvilinear ILT. In this paper, we introduce a GPU-accelerated ILT algorithm that improves not only contour quality and process window but also the precision of curvilinear mask shapes. Our experiments on open benchmarks demonstrate a significant advantage of our algorithm over leading academic ILT engines. Source code will be available at https://github.com/phdyang007/curvyILT.

Invited: Trailblazing the Future: Innovative Chip Design in the Era of Pervasive AI

  • Sudipto Kundu

Few engineering challenges are as complex and arduous as chip design, which typically requires multiple teams of experts and months of dedicated and diligent PPA (Performance, Power and Area) exploration work to achieve the desired goal. In the era of pervasive AI, Chip design automation tools are witnessing a seismic shift, emerging as a powerful intelligent AI agent that orchestrates various decision-making process at every aspect of chip design flow by leveraging multiple computes to explore different PPA strategies. This radical shift demands an AI ecosystem that connects data storage, compute resources, real time data analytics and efficient search space navigation technologies. In this talk, we will explore how physical design implementation tools are embracing AI agents to drive PPA exploration by using novel reinforcement learning techniques powered by deep insights of design and flow execution experience of prior run. The core of such system is built on continuous learning paradigm so that with each successive iteration of the same design, the system learns and adapts and delivers consistent PPA improvement.

SESSION: Session 4: Panel on Heterogenous Integration

Invited: Automatic Die-to-Die Routing with Shielding

  • Sheng-Yu Hsiao
  • Yu-Yueh Chang
  • Jeong-Tying Li

Die-to-die routing is a special layout problem in 3DIC advanced packaging, whose primary productivity bottleneck is routing. The requirements for high wiring density, high shielding rate, complex via stacking and design rules present challenges to create good routing results. In this paper, we describe our approaches to assign signal nets to various routing layers, insert VSS wires to shield signal nets as much as possible and achieve a high completion rate without design rule violations.

Invited: Streamlining and Automating Routing of Multi-Chiplet Technologies

  • Ksenia Roze

As designers transition to heterogeneously integrated multi-chiplet architectures, one of the largest bottlenecks they face is die-to-die (D2D) and die-to-substrate (D2S) signal routing. In some cases, routing can be up to three-quarters of the total design cycle. A next generation routing solution is required to reduce the routing bottleneck. In an effort to reduce this bottleneck, Cadence has been working on new strategies and solutions for automated routing of multi-chiplet technologies. This work led to the development of the industry’s first automated constraint-driven router that can meet the complex D2D and D2S routing requirements, including shielding and teardrop insertion.

This presentation will outline the routing strategies deployed to minimize signal delay, improve yield and ensure that the design meets all electrical and mechanical requirements. Advanced substrate routing automation involves breaking down the routing process into smaller, more controllable tasks such as pin escapes, breakout definitions, layer estimation, through-hole placement, detail routing and multi-layer congestion analysis. Electrical constraints like delay and minimum length must also considered while routing the signals. D2D routing involves developing a compact model with shielding to ensure maximum signal integrity and reliability. This often requires multiple channels for a given D2D region, and global planning of each channel space is necessary to optimize the routing design. Good routing ordering and escape definition achieve uniform routing across different channels.

This talk will present how, by strategically approaching each routing stage and defining topologies based on the complexity of the design, a significant amount of the routing workload can be automated. Our new advanced auto router enables full automated routing solution for multi-chiplet technologies resulting in significant productivity gains and brings new level of innovation to solving most complex D2D and D2S routing challenges.

SESSION: Session 5: Emerging Technologies

ML-QLS: Multilevel Quantum Layout Synthesis

  • Wan-Hsuan Lin
  • Jason Cong

Quantum Layout Synthesis (QLS) plays a crucial role in optimizing quantum circuit execution on physical quantum devices. As we enter the era where quantum computers have hundreds of qubits, optimal OLS tools face scalability issues, while heuristic methods suffer significant optimality gap due to the lack of global optimization. To address these challenges, we introduce a multilevel framework, which is an effective methodology for solving large-scale problems in VLSI design. In this paper, we present ML-QLS, the first multilevel quantum layout tool with a scalable refinement operation integrated with novel cost functions and clustering strategies. Our clustering provides valuable insights into generating a proper problem approximation for quantum circuits and devices. The experimental results demonstrate that ML-QLS can scale up to problems involving hundreds of qubits and achieve a remarkable 69% performance improvement over leading heuristic QLS tools for large circuits, which underscores the effectiveness of multilevel frameworks in quantum applications.

LiDAR: Automated Curvy Waveguide Detailed Routing for Large-Scale Photonic Integrated Circuits

  • Hongjian Zhou
  • Keren Zhu
  • Jiaqi Gu

As photonic integrated circuit (PIC) designs advance and grow in complexity, largely driven by innovations in photonic computing and interconnects, traditional manual physical design processes have become increasingly cumbersome. Available PIC layout automation tools are mostly schematic-driven, which has not alleviated the burden of manual waveguide planning and layout drawing for engineers. Previous research in automated PIC routing largely relies on off-the-shelf algorithms designed for electrical circuits, which only support high-level route planning to minimize waveguide crossings. It is not customized to handle unique photonics-specific routing constraints and metrics, such as curvy waveguides, bending, port alignment, and insertion loss. These approaches struggle with large-scale PICs and cannot produce real layout geometries without design-rule violations (DRVs). This highlights the pressing need for electronic-photonic design automation (EPDA) tools that can streamline the physical design of modern PICs. In this paper, for the first time, we propose an open-source automated PIC detailed routing tool, dubbed LiDAR, to generate DRV-free PIC layout for large-scale real-world PICs. LiDAR features a grid-based curvy-aware A* engine with adaptive crossing insertion, congestion-aware net ordering and objective, and crossing-waveguide optimization scheme, all tailored to the unique property of PIC. On large-scale real-world photonic computing cores and interconnects, LiDAR generates a DRV-free layout with 14% lower insertion loss and 6.25x speedup than prior methods, paving the way for future advancements in the EPDA toolchain. Our codes are open-sourced at https://github.com/ScopeX-ASU/LiDAR.

Invited: Physical Design for Systolic Array-Based Integrated Circuits

  • Jiang Hu

Systolic arrays have become a popular hardware architecture for machine learning computing, which is a key driver for the growth of the semiconductor industry. Unlike many other circuits, systolic arrays exhibit distinct 2D regularity, which holds significant potential for improving physical design quality. However, this regularity is largely overlooked in existing physical design methodologies.

Recent studies have demonstrated that leveraging this regularity can significantly enhance placement quality for FPGAs [1,3], cell placement [2], and mixed-size placement [4]. For instance, utilizing the regularity has resulted in over 20% wirelength reduction in FPGA placement compared to an industrial tool and a remarkable 53% wirelength reduction in mixed-size placement compared to a commercial tool.

Despite these advantages, exploiting the regularity is not as simple as duplicating the schematic. FPGA DSP architectures, typically column-based, often fail to align with the 2D regularity of systolic arrays. Additionally, in cell and mixed-size placement, the placement of IO and control logic can disrupt this regularity. This invited talk will highlight techniques to address these challenges. Beyond placement, the discussion will extend to other opportunities in physical design for systolic arrays, including routing, routability prediction, clock network synthesis, and lithographic hot-spot prediction.

SESSION: Session 6: Reliability

Photonic Side-Channel Analyzer: Enabling Security-Aware Physical Design Methodology

  • Meizhi Wang
  • Yi-Ru Chen
  • S. S. Teja Nibhanupudi
  • Elham Amini
  • Antonio Saavedra
  • Yinan Wang
  • Daniel Wasserman
  • Jean-Pierre Seifert
  • Jaydeep P. Kulkarni

Photon Emission (PE) from Integrated Circuits (IC) is an emerging non-invasive side channel that poses a serious security risk to modern System-on-Chips (SoCs). These emissions, generated during transistor switching, are determined by circuit operations and can be exploited in Side-Channel Analysis (SCA). Furthermore, physical design choices, such as standard cell placement and routing, affect how these emissions propagate and are detected. This makes it crucial to assess and mitigate such risks during the design phase. This paper presents a novel photonic side-channel analysis framework that integrates directly into the physical design flow. The framework enables designers to assess security vulnerabilities in digital ASIC designs by generating both time-resolved and accumulated PE maps at the standard-cell gate level. These PE maps can be applied to various side-channel analysis methods to identify vulnerable regions in the circuit.

We demonstrate the framework by applying it to a 40nm 128-bit Advanced Encryption Standard (AES) core, where we employ localized Correlation PE Attacks (CPEA) on simulated time-resolved PE maps. This approach pinpoints regions with high side-channel leakage. The results showcase the framework’s effectiveness in providing early detection and allow designers to enhance the overall security of the design against PE-related vulnerabilities. To validate our simulation framework, we compared the simulated accumulated PE maps with real-world measurements from a 40nm AES test chip. The close alignment between simulated and measured data confirms the accuracy of our simulator in predicting photon emission behavior across the chip.

Multi-Stage CSM Timing Waveform Propagation Accelerated by NLDM Assistance

  • Shih-Kai Lee
  • Pei-Yu Lee
  • Iris Hui-Ru Jiang

Static timing analysis (STA) is essential for timing closure. To address the complicated effects emerging at advanced technology nodes, the Current Source Model (CSM) has been developed to compute timing waveforms for timing propagation. Compared with Non-Linear Delay Model (NLDM), CSM provides superior accuracy but suffers from the efficiency and scalability issue. In this paper, we propose a multi-stage CSM timing propagation framework with three acceleration techniques with the assistance of NLDM. Our acceleration techniques are general and compatible with any CSM-based STA engine. Experimental results demonstrate the effectiveness of our acceleration techniques: Compared with CSM-based analysis, we achieve 4× speedups with only 0.4% accuracy loss.

SESSION: Session 7: Retrospective and Prospective of Physical Design

Invited: The Future of Functional ECO Automation and Logical Equivalence Checking for Advanced Digital Design Flows

  • Zhuo Li
  • David Stratman

Logical equivalence checking (LEC, or EC) is critical to design implementation and for decades has allowed cost-efficient RTL-level functional testing to be the dominant type of verification done on a project. Test once, then formally prove that the subsequent design stages later in the implementation flow are 100% logically equivalent. But over the last 10-15 years, SoCs have grown 100X in complexity, creating new challenges.

Concurrently, the use of functional ECOs to shortcut long design implementation cycles has skyrocketed. While automated approaches have greatly improved the ECO process quality and accelerated this overall trend, the setup can be challenging for inexperienced designers.

In order to significantly speed up and simplify EC and improve the entire functional ECO process, we require a new approach to both flows. This talk will highlight some of Cadence’s recent breakthrough research in this space, including the use of AI and ML to improve single-run results and multiply designer productivity while gathering insights and leveraging learnings across the duration of a project.

Invited: Toward an ML EDA Commons: Establishing Standards, Accessibility, and Reproducibility in ML-driven EDA Research

  • Vidya A. Chhabria
  • Jiang Hu
  • Andrew B. Kahng
  • Sachin S. Sapatnekar

Machine learning (ML) is transforming electronic design automation (EDA), offering innovative solutions for designing and optimizing integrated circuits (ICs). However, the field faces significant challenges in standardization, accessibility, and reproducibility, limiting the impact of ML-driven EDA (ML EDA) research. To address these barriers, this paper presents a vision for an ML EDA Commons, a collaborative open ecosystem designed to unify the community and drive progress through establishing standards, shared resources, and stakeholder-based governance. The ML EDA Commons focuses on three objectives: (1) Maturing existing EDA infrastructure to support ML EDA research; (2) Establishing standards for benchmarks, metrics, and data quality and formats for consistent evaluation via governance that includes key stakeholders; and (3) Improving accessibility and reproducibility by providing open datasets, tools, models, and workflows with cloud computing resources, to lower barriers to ML EDA research and promote robust research practices via artifact evaluations, canonical evaluators, and integration pipelines. Inspired by successes of ML and MLCommons, the ML EDA Commons aims to catalyze transparency and sustainability in ML EDA research.

Invited: Mapping Two Decades of Innovation: Lessons from 25 Years of ISPD Research

  • Gona Rahmaniani
  • Matthew Guthaus
  • Laleh Behjat

The design automation research community has driven the evolution of integrated circuits from a handful of transistors in the 1960s to billions today. The International Symposium on Physical Design (ISPD) has been instrumental in tackling challenges like scaling complexities, hardware security, and the exponential growth in transistor counts. This study conducts a comprehensive bibliometric analysis of ISPD publications using Natural Language Processing, machine learning, and network analysis. It explores research themes, collaboration dynamics, and global contributions through citation networks, co-authorship graphs, geographical and spatial mapping, and topic modeling. Key areas of focus include Physical Design Optimization, Power Efficiency, and Emerging Technologies, with prominent topics such as placement, routing, clock skew, lithography, machine learning, and hardware security. The analysis highlights the evolution of foundational techniques like placement and routing while identifying emerging trends such as AI-driven design automation. These insights provide a roadmap for sustaining innovation in physical design over the next 25 years.

SESSION: Session 8: Second Keynote

Invited: How Automotive Functional Safety is Disrupting Digital Implementation

  • Chuck J. Alpert
  • Vitali Karasenko
  • Connie O’Shea

The automotive industry is experiencing transformative disruption as the demand for vehicle electrification, connectivity, and autonomy drives manufacturers toward creating a ”datacenter on wheels.” As a result, the cost of silicon in vehicles is projected to rise significantly in the coming years, attracting many semiconductor companies to the market. However, unlike smartphones or data centers, safety is paramount in the automotive sector, prompting widespread adoption of the ISO 26262 functional safety standard. Meeting this standard introduces additional design time, rigorous processes, and increased silicon costs.

In design implementation, safety can be achieved through inserting safety mechanisms such as parity, triple-voting flops, and dual-core lockstep. However, the silicon cost of implementing safety can significantly increase chip area (e.g., from 30-80%), so design teams need advanced methodologies to achieve safety with minimum pain, but also minimum area and power. In particular, the Dual Core Lock Step is a popular safety mechanism since it provides excellent safety coverage for the logic to which it is applied. However, having numerous DCLS modules in a single design can become a floorplanning nightmare, leading to massive congestion, area bloat, and overall performance degradation. We propose a novel methodology for DCLS insertion during logic synthesis and digital implementation to address these issues.

SESSION: Session 9: AI for Chip Design

HeLO: A <u>He</u>terogeneous <u>L</u>ogic <u>O</u>ptimization Framework by Hierarchical Clustering and Graph Learning

  • Yuan Pu
  • Fangzhou Liu
  • Zhuolun He
  • Keren Zhu
  • Rongliang Fu
  • Ziyi Wang
  • Tsung-Yi Ho
  • Bei Yu

Modern very large-scale integration (VLSI) designs usually consist of modules with various topological structures and functionalities. To better optimize such large and heterogeneous logic networks, it is essential to identify the structural and functional characteristics of its modules, and represent them with appropriate DAG types (such as AIG, MIG, XAG, etc.) for logic optimization. This paper proposes HeLO, a hetero-DAG logic optimization framework empowered by hierarchical clustering and graph learning. HeLO leverages a hierarchical clustering algorithm, which splits the original Boolean network into sub-circuits by considering both topological and functional characteristics. A novel graph neural network model is customized to generate the topological-functional embedding (used for distance calculation in hierarchical clustering) and predict the best-fit DAG type of each sub-circuit. Experimental results demonstrate that HeLO outperforms LSOracle, the SOTA heterogeneous logic optimization framework, in terms of node-depth product (for technology-independent logic optimization) and delay-area product (for technology mapping) by 8.7% and 6.9%, respectively.

GraphCAD: Leveraging Graph Neural Networks for Accuracy Prediction Handling Crosstalk-affected Delays

  • Fangzhou Liu
  • Guannan Guo
  • Yuyang Ye
  • Ziyi Wang
  • Wenjie Fu
  • Weihua Sheng
  • Bei Yu

As chip fabrication technology advances, the capacitive effects between wires have become increasingly pronounced, making crosstalk-induced incremental delay a serious issue. Traditional static timing analysis involves complex and iterative calculations through timing windows, requiring precise alignment of aggressor and victim nets, along with delay and slew estimations, which significantly increase runtime and licensing costs. In our work, we develop a Graph Neural Network framework to predict crosstalk-affected delays, focusing on the impacts of the coupling effect and overlapping nets. Moreover, we employ a curriculum learning strategy that gradually integrates aggressors with victims, improving model convergence through progressively complex scenarios. Experimental results show that our framework precisely predicts crosstalk-affected delays, matching commercial tools’ performance with a fivefold speedup.

Invited: AI-assisted Routing

  • Qijing Wang
  • Liang Xiao
  • Evangeline F.Y. Young

Routing is an important but complicated step in physical synthesis. Considering the potential of leveraging AI to seek higher efficiency and better quality in solving routing problems, we study in this work the methodology of AI-assisted routing in a systematic way. Decoupling the functionalities of different routing components will give a high flexibility in determining where and how AI can be used in an effective manner, while maintaining a high degree of interpretability. Two applications along this direction are presented, aiming at tackling the difficulties in routing with AI assistance. These provide examples of how to implement the methodology in practice, while revealing its effectiveness and potential.

SESSION: Session 10: LLM for Chip Design

DRC-Coder: Automated DRC Checker Code Generation Using LLM Autonomous Agent

  • Chen-Chia Chang
  • Chia-Tung Ho
  • Yaguang Li
  • Yiran Chen
  • Haoxing Ren

In the advanced technology nodes, the integrated design rule checker (DRC) is often utilized in place and route tools for fast optimization loops for power-performance-area. Implementing integrated DRC checkers to meet the standard of commercial DRC tools demands extensive human expertise to interpret foundry specifications, analyze layouts, and debug code iteratively. However, this labor-intensive process, requiring to be repeated by every update of technology nodes, prolongs the turnaround time of designing circuits. In this paper, we present DRC-Coder, a multi-agent framework with vision capabilities for automated DRC code generation. By incorporating vision language models and large language models (LLM), DRC-Coder can effectively process textual, visual, and layout information to perform rule interpretation and coding by two specialized LLMs. We also design an auto-evaluation function for LLMs to enable DRC code debugging. Experimental results show that targeting on a sub-3nm technology node for a state-of-the-art standard cell layout tool, DRC-Coder achieves perfect F1 score 1.000 in generating DRC codes for meeting the standard of a commercial DRC tool, highly outperforming standard prompting techniques (F1=0.631). DRC-Coder can generate code for each design rule within four minutes on average, which significantly accelerates technology advancement and reduces engineering costs.

LEGO-Size: LLM-Enhanced GPU-Optimized Signoff-Accurate Differentiable VLSI Gate Sizing in Advanced Nodes

  • Yi-Chen Lu
  • Kishor Kunal
  • Geraldo Pradipta
  • Rongjian Liang
  • Ravikishore Gandikota
  • Haoxing Ren

On-Chip Variation (OCV)-aware and Path-Based Analysis (PBA) accurate timing optimization achieved by gate sizing (including Vth-assignment) remains a pivotal step in modern signoff. However, in advanced nodes (e.g., 3nm), commercial tools often yield suboptimal results due to the intricate design demands and the vast choices of library cells that require substantial runtime and computational resources for exploration. To address these challenges, we introduce LEGO-Size, a generative framework that harnesses the power of Large Language Models (LLMs) and GPU-accelerated differentiable techniques for efficient gate sizing. LEGO-Size introduces three key innovations. First, it considers timing paths as sequences of tokenized library cells, casting gate sizing prediction as a language modeling task and solving it with self-supervised learning and supervised fine-tuning. Second, it employs a Graph Transformer (GT) with a linear-complexity attention mechanism for netlist encoding, enabling LLMs to make sizing decisions from a global perspective. Third, it integrates a differentiable Static Timing Analysis (STA) engine to refine LLM-predicted gate size probabilities by directly optimizing Total Negative Slack (TNS) through gradient descent. Experimental results on 5 unseen million-gate industrial designs in a commercial 3nm node show that LEGO-Size achieves up to 125x speed up with 37% TNS improvement over an industry-leading commercial signoff tool with minimal power and area overhead.

Invited: Artificial Netlist Generation for Enhanced Circuit Data Augmentation

  • Seokhyeong Kang

Optimizing power, performance, and area (PPA) at advanced nodes has become an increasingly challenging and complex task. To address these challenges, approaches such as machine learning (ML) and design-technology co-optimization (DTCO) have emerged as promising solutions. However, their effectiveness is limited by the lack of diverse training data and prolonged turnaround times (TAT). Artificial data has been widely used in various fields to address the limitations of real-world data. By augmenting datasets, artificial data improve the robustness of ML models against input perturbations, leading to improved performance. Similarly, in the physical design flow, artificial data has great potential for overcoming the scarcity of real-world circuit data [1], [2], [3]. Artificial circuits proposed in previous studies are typically designed for specific applications. By developing a method to generate artificial circuit which resemble real circuits, we can address data scarcity and TAT challenges in physical design. In this talk, we will discuss how leveraging artificial circuits to explore a wide range of circuit characteristics can enhance ML model performance for unseen real-world circuits and accelerate the PPA exploration flow.

SESSION: Session 11: Cell Design

Cell-Flex Metrics for Designing Optimal Standard Cell Layout with Enhanced Cell Layout Flexibility

  • Byeonggon Kang
  • Ying Yuan
  • Yucheng Wang
  • Bill Lin
  • Chung-Kuan Cheng

As physical pitch scaling slows, efforts to match its pace by reducing standard cell height and sacrificing horizontal routing tracks have introduced placement and routing challenges, making the design of high-quality standard cell layouts increasingly crucial. However, existing cell metrics only focus on pin accessibility and are insufficient to address issues in advanced nodes (e.g., Power Delivery Networks (PDN), increased routing blockages, etc.). We propose Cell Layout Flexibility(Cell-Flex) metrics, novel metrics that evaluate flexibility of standard cell layouts. flexibility reflects the versatility of cell layouts to placement and routing demands, which influences optimizing block design. By using Cell-Flex metrics as objectives in designing cell layout, we achieve a 13.2% reduction in block area without increasing total Design Rule Violations (DRVs). We develop a Machine Learning (ML) model using Kolmogorov-Arnold Networks (KAN) that utilizes the Cell-Flex metrics as features to make DRV prediction. By adding Cell-Flex features, we improve accuracy from 0.65 to 0.79 and F1 score from 0.52 to 0.78, demonstrating that our metrics are important for DRV prediction and serve as robust indicators of cell layout quality.

Scalable CFET Cell Library Synthesis with A DRC-Aware Lookup Table to Optimize Valid Pin Access

  • Ting-Wei Lee
  • Ting-Xin Lin
  • Yih-Lang Li

With the advent of CFET technology, which stacks P and N transistors together, the number of available tracks in a cell decreases. This poses a substantial challenge of hard-to-access pins during upper-level routing, which has been addressed in previous works by lengthening IO pins and increasing the spacing between adjacent IO pins. However, upper-level routing may generate DRC violations around IO pins in a cell, which compromises these efforts to improve pin accessibility. To overcome this challenge, we propose a scalable satisfiability modulo theories-based cell routing that establishes a DRC-aware scheme to enumerate potential DRC violations, enabling pin accessibility to be improved without producing DRC violations in upper-level routing. Our experimental results demonstrate that the proposed CFET cell generator is 100 times faster than previous work on average while delivering the same or better cell quality in terms of cell area. The scalability of the proposed method allows for the synthesis of large cells, including high-driving-strength cells and multi-bit flip flop (MBFF). Moreover, compared to previous work, the proposed method reduces DRC violations by an average of 99% in upper-level routing, and reduces both wire length and via usage effectively as well.

LVFGen: Efficient Liberty Variation Format (LVF) Generation Using Variational Analysis and Active Learning

  • Junzhuo Zhou
  • Haoxuan Xia
  • Wei Xing
  • Ting-Jung Lin
  • Li Huang
  • Lei He

As transistor dimensions shrink, process variations significantly impact circuit performance, signifying the need for accurate statistical circuit analysis. In digital circuit timing analysis, the Liberty Variation Format (LVF) has emerged as an industrial leading representation of timing distributions in cell libraries at 22 nm and below. However, LVF characterization relies on the Monte Carlo (MC) method, which requires excessive SPICE simulations of cells with process variations. Similar challenges also exist for uncertainty propagation and quantification in chip manufacturing and the broader scientific communities. To resolve this foundational challenge, this paper presents LVFGen, a novel method that reduces the simulation costs of MC while generate high-accuracy LVF library. LVFGen utilizes an active learning strategy based on variational analysis to identify process variation samples that impact timing distributions more significantly. Compared to the state-of-the-art Quasi-MC method, LVFGen demonstrates an overall 2.27× speedup in LVF library generation within an accuracy level of 5k-sample MC and a 4.06× speedup within a 100k-sample MC accuracy.

Abuttable Analog Cell Library and Automatic AMS Layout

  • Tianjia Zhou
  • Cheng Chang
  • Li Huang
  • Jingyun Gu
  • Zexin Ji
  • Xiangyang Liu
  • Hailang Liang
  • Zhanfei Chen
  • Ting-Jung Lin
  • Song Wang
  • Na Bai
  • Zhengping Li
  • Lei He

The state of the art analog circuit design applies mainly a full-custom layout methodology. This demands high expertise and heavy manual workload. Additionally, neither can the resulting layout be re-used easily across different designs or different PDKs. Learning from digital standard cells, existing work has proposed stem cells that are abuttable. But stem cells have a fixed area ratio of 2 over same-sized Pcells, limiting its wide application. In this paper we develop a new type of abuttable analog cells (called Acells) for transistors and passive elements. Acells are compatible with digital standard cells and can be abutted in all directions, enabling the use of automatic digital place and route (PnR) engines. We automate Acell generation and show that the average area ratio over same-sized Pcell is 1.49 for 65nm technology and 1.3 for 28nm technology, and is expected to decrease for more advanced technologies. We then use digital PnR to automatically layout a number of analog and mixed-signal (AMS) circuits mainly in 28nm, and show that compared to Pcell-based manual layout, Acell-based layout obtains similar performance and its circuit level layout area is about 2% higher for large scale AMS circuits in our experiments.

SESSION: Session 12: 3D IC Part I

Placement-Aware 3D Net-to-Pad Assignment for Array-Style Hybrid Bonding 3D ICs

  • Pruek Vanna-iampikul
  • Junsik Yoon
  • Chaeryung Park
  • Gary Yeap
  • Sung Kyu Lim

Hybrid bonding is emerging as a key technology for 3D integration, offering finer bonding pitches that address the high interconnect density requirements of modern VLSI applications. In advanced node technologies, where metal pitches are significantly smaller than bonding pitches, 3D net assignment becomes critical for achieving optimal design performance. Existing approaches primarily focus on either ensuring the legality of the assignment or optimizing the flexibility of 3D net locations for timing purposes in isolation. This limitation restricts the performance improvements of 3D designs over traditional 2D counterparts. To overcome these challenges, we introduce AnchorGrid, a novel 3D net assignment framework designed to concurrently assign 3D nets to legal locations while supporting their movement to enhance timing optimization. By modeling 3D nets as pairs of specialized ”anchor” cells, accompanied by relative placement constraints, precise movement and alignment are achieved during the pre-route optimization phase, before final placement onto grid-based locations. Experimental results on advanced node commercial designs demonstrate that AnchorGrid achieves up to a 24.35% improvement in power, performance, and area (PPA) metrics, while reducing design rule check (DRC) violations by 90%, outperforming state-of-the-art methods.

Invited: Physical Design for Advanced 3D ICs: Challenges and Solutions

  • Yuxuan Zhao
  • Lancheng Zou
  • Bei Yu

As technology scaling predicted by Moore’s law slows down, 3D integrated circuits (3D ICs) have emerged as a promising alternative to enhance performance while maintaining cost-effectiveness. With the advancement of fabrication and bonding technologies, wafer-level 3D integration enables fine-grain 3D interconnects that maximize the benefits in power, performance, and area (PPA). However, a multitude of challenges have obstructed traditional electronic design automation (EDA) methodologies for 3D IC implementations. This paper surveys the major challenges in the physical design of advanced 3D ICs. We provide a comprehensive review of existing solutions, analyzing their advantages and disadvantages in depth. Finally, we discuss open problems and research opportunities in the development of native 3D EDA tools.

Invited: Chiplet-Based Integration – Scale-Down and Scale-Out

  • Boris Vaisband

Motivation: The demand for increased computation and memory in applications such as large language models, has increased well beyond the reticle boundaries of a system-on-chip (SoC). Chiplet-based integration is a paradigm shift that shapes the way we design our future high-performance systems. The concept is to move away from large SoCs that are limited by communication, thermal design power, and reticle size, toward a robust plug-and-play approach, where small, hardened IP heterogeneous off-the-shelf chiplets are seamlessly integrated on a single platform.

Problem statement: Recent technological breakthroughs in advanced packaging platforms, have enabled the integration of hundreds to thousands of chiplets within a single platform. Nonetheless, building a functional and efficient ultra-large-scale high-performance computation system, requires overcoming important system-level design challenges. Specifically, short- and long-range communication, power delivery and thermal management, testing, synchronization, hardware security, and others.

Approach: In this talk, we will discuss the current state-of-the-art and challenges in chiplet integration as well as the scale-down and scale-out concepts. We will introduce the silicon interconnect fabric (Si-IF), an ultra-large wafer-scale heterogeneous integration platform, for applications such as high-performance computing. We will discuss paths to address the system-level challenges for designing and integrating a high-performance computation system on the Si-IF.

SESSION: Session 13: Lifetime Achievement Session

Invited: Innovation in Times of Technology Disruption

  • Bryan Preas

I hosted Jason Cong when he was an intern at the Xerox Palo Alto Research Center in 1987. Since then, we have been friends and collaborators. I have watched his extraordinary accomplishments with pride and pleasure over the years.

Disruptive technologies are a well-studied business school topic. New, or significantly improved, technologies present new problems and allow new approaches for older challenges. These disruptions are often accompanied by substantial technical innovation and creation of new business values. In their time, electric power, automobiles and television disrupted society. More recent examples include the internet and e-commerce.

Invited: Shaping the Future of Interconnected Physical Design

  • David Z. Pan

Physical design has been a cornerstone of electronic design automation (EDA) since the early days of chip and board development, with placement and routing at its core. By the late 1980s, the field was prematurely declared ”dead,” as many believed its challenges had been resolved. However, the advent of deep submicron scaling in the 1990s revitalized physical design research, establishing it as an indispensable part of the design process. Today, with technology advancing to the 1x nm regime and the rise of 3D heterogeneous integration (3DHI), physical design remains pivotal in achieving design closure across power, performance, area, cost, and turnaround time (PPACT). Over time, physical design has transformed from its classical ”place and route” framework into a more holistic and interconnected discipline, crosscutting into physical synthesis, design for manufacturing, 3DHI, analog and RF, emerging technologies, and AI/ML. The seminal contributions of Prof. Jason Cong have been instrumental in shaping the field of interconnected physical design. The seeds he planted have grown into thriving forests, with his academic descendants emerging as key leaders driving advancements in the field. This talk will explore the synergistic aspects of interconnected physical design and highlight Prof. Cong’s profound influence and legacy in shaping the future of the field.

Invited: Coping with Interconnects

  • Jason Cong

In this paper, I review the multi-decade research on overcoming the performance bottleneck of VLSI interconnects in deep sub-micrometer and nanometer technologies that started at UCLA in the early 1990s. Our research spans from interconnect topology and geometry optimization, to wire length reduction via scalable placement, to use of novel interconnect technologies such as 3D IC and RF-interconnects, to recent work on interconnect pipelining in chiplet designs, and the shift from interconnect to entanglement in quantum computing. The latter two efforts go beyond the typical physical design space and involve space-time co-optimization. This paper is dedicated to multiple generations of Ph.D. students, postdocs, and visiting researchers who contributed to build a strong physical design research program at UCLA.

SESSION: Session 14: Third Keynote

Invited: Automation and Optimization of Heterogeneous Multi-Die Systems

  • Henry Sheng

Advances in heterogeneous integration have enabled the creation of systems built from chips and interconnects using multiple silicon node technologies, package technologies, optical technologies, thermal mitigation, and more. The scale of integration has increased non-linearly with the pitch of die-to-die interconnect such as bump and hybrid bonds. Traditionally, many systems have been constructed as an ‘assembly’ of parts using manual layout techniques. Advanced package technologies have evolved to a point where they have achieved densities that challenge the viability of assembly-based manual methods as multi-die systems scale to 5X, 8X or even 40-60X reticle size at wafer scale. The scale complexity grows both with the densification of die-to-die connections as well as the increase of allowable system footprints. Furthermore, the integration of heterogeneous components in a single design mandates a workflow that fuses previously disconnected heterogeneous workflows and competencies together. These are secular shifts that are opening new classes of problems including migration from manual assembly to automated assembly, and then from automated assembly to optimization of system QoR (quality of results). This requires design automation tools to have a unified representation and treatment of heterogeneous systems and operate on optimization at a system scale across heterogeneous components for the key system QoR metrics such as SIPI, EMIR and Thermal.

SESSION: Session 15: 3D IC Part II

ML-Based Fine-Grained Modeling of DC Current Crowding in Power Delivery TSVs for Face-to-Face 3D ICs

  • Zheng Yang
  • Zhen Zhuang
  • Bei Yu
  • Tsung-Yi Ho
  • Martin D. F. Wong
  • Sung Kyu Lim

In contrast to uniform distribution in power wires, actual currents tend to exhibit complicated crowding phenomena at the connections between Through-silicon-via (TSV) and power wires. The current crowding effect degrades power integrity and increases the difficulty of 3D IC power delivery network (PDN) analysis. Therefore, a detailed analysis of current distribution and IR drops in power TSVs within 3D IC PDN is important. This paper will explore the complicated current behavior within TSVs and PDNs of the promising face-to-face 3D IC architecture. Since existing simulation methods are computationally intensive and time-consuming, we propose a graph attention network-based (GAT-based) framework, with novel aggregation methods in the GAT models and informative fine-grained graph generation methods, to achieve efficient analysis of current crowding and IR drops in face-to-face 3D IC TSVs. For current density and voltage predictions, the proposed framework attains R2 scores of 0.9776 and 0.9952 compared to ground truth results, respectively. Our framework also demonstrates over 837× speedup than ANSYS Q3D Extractor. Furthermore, the proposed framework outperforms other machine learning-based (ML-based) methods, including the state-of-the-art method.

Invited: Modeling and Design Methodology for Backside Integration of Voltage Converters

  • Amaan Rahman
  • Hang Yang
  • Cong Hao
  • Sung Kyu Lim

As technology scales down, backside power delivery networks (BS-PDNs) are increasingly adopted to address frontside parasitics, power density, and IR-drop challenges. However, conventional BS-PDN architectures struggle with off-chip IR-drop issues caused by package-level parasitics. We introduce the first fully integrated backside voltage regulator (BS-IVR) to reduce off-chip circuitry overhead, and improve load regulation and on-chip power integrity. BS-IVR utilizes back-end-of-line compatible amorphous tungsten-doped indium oxide transistors for backside integration and our comprehensive backside PDK for design and verification. We developed an end-to-end EDA flow for BS-IVR on-chip integration and power integrity analysis. Optimizing the BS-IVR for on-chip integration achieves a compact IVR footprint of 0.0142 mm2 with a power density of 3.02 W/mm2 and efficiency of 60%. On-chip dynamic VDD IR-drop is further reduced by 46.06%.

Invited: Next-Generation Power Integrity Concepts and Applications for Physical Design

  • Emrah Acar

The rapid pace of innovation in electronics, driven by advancements in computing, machine learning, and artificial intelligence, has created an unprecedented demand for more efficient and powerful computing platforms. As integrated circuits (ICs) continue to scale and integrate into increasingly complex systems, they consume more power, leading to significant challenges in power integrity. These challenges are further exacerbated by the growing complexity of modern IC designs, necessitating more intelligent and actionable approaches to ensure robust power delivery networks.

This presentation introduces a novel methodology for addressing power integrity issues in next-generation IC designs. We propose a victim/aggressor interaction model as a foundational concept for IR drop analysis. This model enables the decomposition of IR drop in a victim instance into contributions from multiple aggressors and other components. By understanding these interactions, designers can implement corrective actions during the placement and routing stages, as well as enhance power connectivity at higher levels of the design hierarchy.

We will discuss the foundational advantages of RedHawk-SC SigmaDVD™ Technology, a cutting-edge solution for power integrity analysis and signoff. SigmaDVD™ is designed to address dynamic voltage drop (DVD) issues at advanced process nodes, providing comprehensive coverage and enabling early detection and prevention of voltage-drop-related problems. This technology is instrumental in achieving robust power integrity signoff, fixing IR violations, and ensuring timing closure with high confidence.

Key applications of SigmaDVD™ in physical design, IR/STA (Static Timing Analysis), and IR/ECO (Engineering Change Order) tools will be highlighted. The presentation will demonstrate how SigmaDVD™ is becoming the industry-leading method for avoiding DVD-induced voltage and timing problems, enabling shift-left prevention of voltage-drop issues, and delivering high-coverage power integrity signoff for advanced-node designs.

By leveraging these next-generation concepts and tools, designers can achieve more efficient, reliable, and high-performance ICs, paving the way for continued innovation in the electronics industry.

SESSION: Session 16: Contest Results and Closing Remarks

Invited: ISPD 2025 Performance-Driven Large Scale Global Routing Contest

  • Rongjian Liang
  • Anthony Agnesina
  • Wen-Hao Liu
  • Matt Liberty
  • Hsin-Tzu Chang
  • Haoxing Ren

Global routing is a critical aspect of VLSI design, significantly impacting timing, power consumption, and routability. The ISPD2024 contest focused on addressing the scalability challenges of global routing by leveraging GPU and machine learning techniques. Building on this foundation, the ISPD2025 contest introduces several important updates to better reflect real-world routing challenges. These updates include the provision of industry-standard input files for more precise modeling and integration with OpenROAD for accurate performance assessment. Collectively, these updates aim to bring the contest closer to practical routing scenarios, fostering the development of scalable and efficient solutions for large-scale chip designs.

Remembering Arvind

From Friends, Colleagues, and Students of Arvind

It is with a heavy heart that we write to share the news that on June 17th 2024, we lost our beloved colleague, mentor, and friend Arvind. Arvind passed away at 77 years old, after an illness he was being treated for took a sudden turn for the worse.

Arvind was a pioneer in computer architecture and digital systems. Over his five decades at MIT, he contributed countless foundational techniques, spanning dataflow architectures, parallel processors and programming languages, hardware description languages, and synthesis and verification of digital systems. Arvind made lasting contributions to hardware and architectural abstractions for both correctness and efficiency.

Arvind received a B.Sc. in Electrical Engineering from IIT Kanpur in 1969, followed by a M.Sc. and Ph.D. in Computer Science from the University of Minnesota in 1972 and 1973, respectively. He taught at UC Irvine from 1974-78 before moving to MIT in 1978. He received several awards during his illustrious career, including distinguished alumni awards from both his alma maters. Arvind became an IEEE Fellow in 1994 and an ACM Fellow in 2006. He was also elected to the National Academy of Engineering in 2008.

Arvind’s presence in the computer architecture community will continue to be felt via his legacy of deep technical contributions and his academic lineage. Arvind advised over 35 Ph.D. students during his career, who have all gone on to become leaders in academia and industry.

Arvind was a deep thinker with a great sense of abstraction. He would continuously seek to simplify complex systems proposed and implemented by his students, by relentlessly pushing them to explain the fundamentals to him at abstraction levels that hid the implementation complexity. Those of us who had the pleasure of writing a paper with Arvind know his single-minded devotion to really understand each result at its most fundamental level. Even after all the research was done and maybe after we all, meaning everyone except Arvind, thought the paper was done, Arvind would often force us all to understand it at yet a deeper level. Then, no matter the lateness of the hour, we had to explain this new understanding to the world to Arvind’s satisfaction. Two of his students recall working hard on a paper over many weeks but could not get it polished to Arvind’s exacting standards before the deadline. The students thought it was good enough to potentially be accepted and certainly not an embarrassment. Arvind told them “you can submit it, but take my name off of it.” This was his way of teaching students how to simplify complexity through abstraction. Invariably, these efforts would result in advancements that would have been difficult to tease out from the original description. Arvind would say that once a researcher gets used to this abstract thinking style, there is no way for them to change back. Arvind was like this in all intellectual and interpersonal pursuits. Certainly, his students were converted for life.

Arvind sincerely believed that knowledge was its own reward. One time in a discussion about the content of annual reviews, he commented that when he thinks back over the year if he can identify one (or maybe two) deep insights he gained over the year, then he considered it a successful year. That seems to be the epitome of self-actualization.

Arvind would say that a PhD advisor shouldn’t believe that they had anything to do with the brilliance of the students they graduate when the students were already bright when they came to the advisor! His sage advice to his students who went on to become professors was “The best thing an advisor can do is to keep pointing a student in the right direction and hang on for the good ride.”

Arvind treated his research group as part of the family. Arvind’s wife Gita would often take their sons to visit him at the office when they were young, as much to see the grad students as Arvind. Frequently, discussions at MIT would last into late evenings, and Arvind would call on Gita to bring food to him and his students, so they could continue the discussion. And Gita would deliver, without fail!

Arvind was deeply passionate about teaching and played a significant role in developing course content for computer architecture and hardware design courses both at MIT and several other universities via sabbaticals and collaborations. He would prepare extensively, sometimes for days, going over each slide in excruciating detail, for a lecture that he had given many times before. His love for and dedication to teaching never took a back seat even when he was ailing – he watched and critiqued the final project presentations and actively participated in final grading for the course he was teaching in Spring 2024 from a hospital bed.

On a personal level, Arvind was a wonderful person: truly kind and giving, and always smiling. He was an inspiring mentor and a true friend to countless students, colleagues, and researchers around the world. Arvind and Gita opened their Arlington home to anyone and everyone they met. Their annual Diwali parties over 3 decades were highlights for those lucky enough to attend; attendees would invariably be amazed by the myriad friends from different walks of life that Arvind and Gita had collected. They were true pillars of both the MIT and Arlington communities — and in some circles, Arvind was simply known as Gita’s husband!

For all of us, Arvind was a constant source of positive energy and wisdom. One of his favorite quotes was, “Pessimists are more often right, optimists live happier lives!.” That was certainly true of Arvind — his zest for life was infectious and he invariably lifted people’s spirits. May you rest in peace, dear Arvind. You will live forever in our hearts.

ISPD’24 TOC

ISPD ’24: Proceedings of the 2024 International Symposium on Physical Design

 Full Citation in the ACM Digital Library

SESSION: Session 1: Opening Session and First Keynote

Engineering the Future of IC Design with AI

  • Ruchir Puri

Software and Semiconductors are two fundamental technologies that have become woven into every aspect of our society, and it will be fair to say that “Software and Semiconductors have eaten the world”. More recently, advances in AI are starting to transform every aspect of our society as well. These are three tectonic forces of transformation – “AI”, -Software”, and “Semiconductors” which are colliding together resulting in a seismic shift – a future where both software and semiconductor chips themselves will be designed, optimized, and operated by AI – pushing us towards a future where “Computers can program themselves!”. In this talk, we will discuss these forces of “AI for Chips and Code” and how the future of Semiconductor chip design and software engineering is being redefined by AI.

SESSION: Session 2: Partitioning and Clustering

MedPart: A Multi-Level Evolutionary Differentiable Hypergraph Partitioner

  • Rongjian Liang
  • Anthony Agnesina
  • Haoxing Ren

State-of-the-art hypergraph partitioners, such as hMETIS, usually adopt a multi-level paradigm for efficiency and scalability. However, they are prone to getting trapped in local minima due to their reliance on refinement heuristics and overlooking global structural information during coarsening. SpecPart, the most advanced academic hypergraph partitioning refinement method, improves partitioning by leveraging spectral information. Still, its success depends heavily on the quality of initial input solutions. This work introduces MedPart, a multi-level evolutionary differentiable hypergraph partitioner. MedPart follows the multi-level paradigm but addresses its limitations by using fast spectral coarsening and introducing a novel evolutionary differentiable algorithm to optimize each coarsening level. Moreover, by analogy between hypergraph partitioning and deep graph learning, our evolutionary differentiable algorithm can be accelerated with deep graph learning toolkits on GPUs. Experiments on public benchmarks consistently show MedPart outperforming hMETIS and achieving up to a 30% improvement in cut size for some benchmarks compared to the best-published solutions, including those from SpecPart—moreover, MedPart’s runtime scales linearly with the number of hyperedges.

FuILT: Full Chip ILT System With Boundary Healing

  • Shuo Yin
  • Wenqian Zhao
  • Li Xie
  • Hong Chen
  • Yuzhe Ma
  • Tsung-Yi Ho
  • Bei Yu

Mask optimization in lithography is becoming increasingly impor- tant as the technology node size shrinks down. Inverse Lithography Technology (ILT) is one of the most performant and robust solutions widely used in the industry, yet it still suffers from heavy time con- sumption and complexity. As the number of transistors scales up, the industry currently focuses more on efficiency improvement and workload distribution. Meanwhile, most recent publications are still tangled in local pattern restoration regardless of real manufacturing conditions. We are trying to extend academia to some real industrial bottlenecks with FuILT, a practical full-chip ILT-based mask opti- mization flow. Firstly, we build a multi-level partitioning strategy with the divide-and-conquer mindset to tackle the full-chip ILT prob- lem. Secondly, we implement a workload distribution framework to maintain hardware efficiency with scalable multi-GPU parallelism. Thirdly, we propose a gradient-fusion technique and a multi-level healing strategy to fix the boundary error at different levels. Our experimental results on different layers from real designs show that FuILT is both effective and generalizable.

Slack Redistributed Register Clustering with Mixed-Driving Strength Multi-bit Flip-Flops

  • Yen-Yu Chen
  • Hao-Yu Wu
  • Iris Hui-Ru Jiang
  • Cheng-Hong Tsai
  • Chien-Cheng Wu

Register clustering is an effective technique for suppressing the increasing dynamic power ratio in modern IC design. By clustering registers (flip-flops) into multi-bit flip-flops (MBFFs), clock circuitry can be shared, and the number of clock sinks and buffers can be lowered, thereby reducing power consumption. Recently, the use of mixed-driving strength MBFFs has provided more flexibility for power and timing optimization. Nevertheless, existing register clustering methods usually employ evenly distributed and invariant path slack strategies. Unlike them, in this work, we propose a register clustering algorithm with slack redistribution at the post-placement stage. Our approach allows registers to borrow slack from connected paths, creates the possibility to cluster with neighboring maximal cliques, and releases extra slack. An adaptive interval graph based on the red-black tree is developed to efficiently adapt timing feasible regions of flip-flops for slack redistribution. An attraction-repulsion force model is tailored to wisely select flip-flops to be included in each MBFF. Experimental results show that our approach outperforms state-of-the-art work in terms of clock power reduction, timing balancing, and runtime.

SESSION: Session 3: Timing Optimization

Calibration-Based Differentiable Timing Optimization in Non-linear Global Placement

  • Wuxi Li
  • Yuji Kukimoto
  • Gregory Servel
  • Ismail Bustany
  • Mehrdad E. Dehkordi

Placement plays a crucial role in the timing closure of integrated circuit (IC) physical design. This paper presents an efficient and effective calibration-based differentiable timing-driven global placement engine. Our key innovation is a calibration technique that approximates a precise but expensive reference timer, such as a signoff timer, using a lightweight simple timer. This calibrated simple timer inherently accounts for intricate timing exceptions and common path pessimism removal (CPPR) prevalent in industry designs. Extending this calibrated simple timer into a differentiable timing engine enables ultrafast yet accurate timing optimization in non-linear global placement. Experimental results on various industry designs demonstrate the superiority of the proposed framework over the latest AMD Vivado and traditional net-weighting methods across key metrics including maximum clock frequency, wirelength, routability, and overall back-end runtime.

Novel Airgap Insertion and Layer Reassignment for Timing Optimization Guided by Slack Dependency

  • Wei-Chen Tai
  • Min-Hsien Chung
  • Iris Hui-Ru Jiang

BEOL with airgap technology is an alternative metallization option with promising performance, electrical yield and reliability to explore at 2nm node and beyond. Airgaps form cavities in inter-metal dielectrics (IMD) between interconnects. The ultra-low dielectric constant reduces line-to-line capacitance, thus shortening the interconnect delay. The shortened interconnect delay is beneficial to setup timing but harmful to hold timing. To minimize the additional manufacturing cost, the number of metal layers that accommodate airgaps is practically limited. Hence, circuit timing optimization at post routing can be achieved by wisely performing airgap insertion and layer reassignment to timing critical nets. In this paper, we present a novel and fast airgap insertion approach for timing optimization. A Slack Dependency Graph (SDG) is constructed to view the timing slack relationship of a circuit with path segments. With the global view provided by SDG, we can avoid ineffective optimizations. Our Linear Programming (LP) formulation simultaneously solves airgap insertion and layer reassignment and allows a flexible amount of airgap to be inserted. Both SDG update and LP solving can be done extremely fast. Experimental results show that our approach outperforms the state-of-the-art work on both total negative slack (TNS) and worst negative slack (WNS) with more than 89× speedup.

Parallel and Heterogeneous Timing Analysis: Partition, Algorithm, and System

  • Tsung-Wei Huang
  • Boyang Zhang
  • Dian-Lun Lin
  • Cheng-Hsiang Chiu

Static timing analysis (STA) is an integral part in the overall design flow because it verifies the expected timing behaviors of a circuit. However, as the circuit complexity continues to enlarge, there is an increasing need for enhancing the performance of existing STA algorithms using emerging heterogeneous parallelism that comprises manycore central processing units (CPUs) and graphics processing units (GPUs). In this paper, we introduce several state-of-the-art STA techniques, including task-based parallelism, task graph partition, and GPU kernel algorithms, all of which have brought significant performance benefits to STA applications. Motivated by these successful results, we will introduce a task-parallel programming system to generalize our solutions to benefit broader scientific computing applications.

SESSION: Session 4: Panel on EDA Challenges at Advanced Technology Nodes

Introduction to the Panel on EDA Challenges at Advanced Technology Nodes

  • Tung-Chieh Chen

We have gathered a panel of experts who will delve into the electronic design automation (EDA) challenges at advanced technology nodes. In this rapidly evolving field, the race towards miniaturization presents new hurdles and complexities. With advanced nodes shrinking, design and technology co-optimization becomes an increasingly intricate task. Our panelists will share their unique insights on how they approach these challenges and the impact of complicated design rules on the design process. As designs grow larger and more complex, novel strategies and methodologies are necessary to address the runtime issues of EDA tools. Our discussion will explore these strategies, including promising emerging technologies such as multi-machine or GPU acceleration that show potential in mitigating runtime challenges.

In the realm of design space exploration, effective navigation of trade-offs between different design objectives, especially power, performance, and area, is critical. Our panelists will share experiences of frontend/backend co-optimization in the advanced node design flow and discuss how machine learning techniques can be harnessed to predict and estimate various aspects of the design process. As we step into the era of “More than Moore,” the integration of diverse technologies into a unified design process presents both opportunities and challenges. Our discussion will explore anticipated advancements in advanced packaging or heterogeneous integration and identify issues at advanced technology nodes that, in the panelists’ opinion, have not yet been adequately addressed by commercial tools and/or academic research.

We will also delve into the scaling and runtime challenges at multi-million gates and beyond, the current state of EDA tools for data-sharing for machine learning/AI across tools, and the major new issues that need to be addressed for more optimal design in the 2nm to 5nm process technology nodes. Finally, we will discuss the high-priority research topics that need to be tackled to address advanced technology nodes. We look forward to an enlightening discussion filled with expert insights and thought-provoking ideas.

Panel Statement: EDA Needs at Advanced Technology Nodes

  • Andrew B. Kahng

Improvements in EDA technology become more urgent with the loss of traditional scaling levers and the growing complexity of leading-edge designs. The roadmap for device and cell architectures, power delivery, multi-tier integration, and multiphysics signoffs presents severe challenges to current optimization frameworks. In future advanced technologies, the quality of results, speed, and cost of EDA will be essential components of scaling. Each stage of the design flow must deliver predictable results for given inputs and targets. Optimization goals must stay correlated to downstream and final design outcomes.

The EDA and design ecosystem must achieve an aligned interest in scalable, efficient optimization. End-case solvers and core engines should become commoditized, freeing researchers and EDA suppliers to address new automation requirements. A rearchitected, rebuilt EDA 2.0 must achieve results in less time (multithreading, GPU) and better results in same time (cloud-native, sampling), while enabling AI/ML that steers and orchestrates optimizers within the design process. To deliver high-value system-technology pathfinding and implementation tools in time to meet industry needs, EDA must develop (learn) new optimization objectives, hierarchical solution strategies, multiphysics reduced-order models, and conditioning of instances at the interstices between chained optimizations. Last, as a community, EDA will meet the challenges of advanced technologies by investing in research and workforce development, via open infrastructure for ML EDA, and via proxy research enablements.

Large Language Models for EDA: Future or Mirage?

  • Zhuolun He
  • Bei Yu

In this paper, we explore the burgeoning intersection of Large Language Models (LLMs) and Electronic Design Automation (EDA). We critically assess whether LLMs represent a transformative future for EDA or merely a fleeting mirage. By analyzing current advancements, challenges, and potential applications, we dissect how LLMs can revolutionize EDA processes like design, verification, and optimization. Furthermore, we contemplate the ethical implications and feasibility of integrating these models into EDA workflows. Ultimately, this paper aims to provide a comprehensive, evidence-based perspective on the role of LLMs in shaping the future of EDA.

PANEL: EDA Challenges at Advanced Technology Nodes A

  • Eugene Liu

Le-Chin Eugene Liu received a Bachelor’s and a Master’s degree in electronics engineering from the National Chao Tung University, Hsinchu, Taiwan, and a Ph.D. from the Department of Electrical Engineering, University of Washington, Seattle, Washington, USA.

PANEL: EDA Challenges at Advanced Technology Nodes B

  • Guang-Wan Liao

PPA (Performance, Power, and Area) is our goal. Although most people no longer pursue top speed, it is still very important. When reducing power and area, a common side effect is that timing gets worse. Low power/area method never makes speed faster. Once the speed slows down, buffering or up-sizing is inevitable. The difference between previous tech nodes is the gap between global and detailed placement (GP and DP). More and more rules make gap growing. Once DP cannot make most instances to stay original place, expectation of global placement will be disappointed. GP or more buffering will be necessary again. Physical-rule awareness will be a necessary improvement for GP/buffering. We will list some of the challenges how advanced nodes face.

Challenges in Floorplanning and Macro Placement for Modern SoCs

  • I-Lun Tseng

Modern System-on-Chips (SoCs), such as smartphone microprocessors, are composed of billions of transistors existing in various subsystems. These subsystems can include Central Processing Units (CPUs), Graphics Processing Units (GPUs), Neural Processing Units (NPUs), Image Signal Processors (ISPs), Digital Signal Processors (DSPs), communication modems, memory controllers, and many others. For efficient Electronic Design Automation (EDA) tasks, such as those involving logic synthesis, placement, clock tree synthesis (CTS), and/or routing, these subsystems are typically broken down into smaller, more manageable circuit blocks, or circuit partitions. This subdivision strategy is crucial for keeping design times within reasonable limits.

During the top-level floorplanning phase of chip design, the dimensions, interconnect ports, and physical locations of circuit partitions are defined; the physical boundaries of these partitions are commonly designed as rectilinear shapes rather than rectangles. Partitions that are excessively large can lead to inefficient use of chip area, higher power consumption, and higher production costs. Conversely, undersized partitions can hinder subsequent physical design processes, potentially causing delays in the overall chip design schedule. Furthermore, a poor floorplan can lead to longer wire lengths and can increase feedthrough net counts in partitions, adversely affecting power, performance, and area (PPA).

In practice, the top-level floorplanning phase of chip design can involve multiple iterations of its processes. An initial iteration typically involves estimating the approximate area of each circuit partition based on various factors, such as the dimensions of macros (including SRAM macros), the number of standard cell instances, and the standard cell utilization rate, which can be projected based on the data from previous designs. These preliminary estimates are crucial for defining the initial shapes, dimensions, interconnect ports, and physical locations of the partitions. Subsequently, the downstream design processes can advance either to partition-level physical design (which includes macro placement, standard cell placement, CTS, routing, etc.) or to physical-aware logic synthesis, which uses the defined layout data to more precisely assess layout-induced effects and produce more accurate gate-level netlists.

Once the dimensions and interconnect locations of circuit partitions are defined, macro placement, which is usually followed by standard cell placement and routing processes, can be conducted. After performing these processes, PPA results may indicate that certain partitions require size adjustments due to being too small, whereas others may be identified as candidates for area reduction. Such alterations in the circuit partition areas necessitate modifications to the top-level floorplan. Furthermore, in subsequent iterations of floorplanning, certain elements (such as feedthrough nets/ports) may be added into and/or removed from partitions, prompting a reevaluation of the physical implementation feasibility for these partitions; the reevaluation stage may involve additional macro placement, cell placement, and routing activities.

Macro placement is crucial in physical design as its outcomes can substantially influence standard cell placement, CTS, routing, circuit timing, and even power consumption. However, at advanced technology nodes, macro placement outcomes produced by commercial EDA tools and reinforcement learning (RL)-based tools often require human modifications prior to practical use, which in part owing to complex design rules associated with advanced technology nodes, although these tools can rapidly generate results. Additionally, it has been observed that suboptimal macro placement can lead to issues such as IR drop and increased dynamic/static power consumption. However, these issues, which may be checked more accurately in later stages of a design flow, are frequently not addressed in a typical macro placement process. In modern SoCs, moreover, it is very common that a circuit partition contains multiple power domains. Performing macro placement on this type of circuit partition may require domain floorplanning prior to placing macros and standard cell instances within their respective power domain regions.

As described previously, the floorplanning and the macro placement are often interrelated. Early iterations of floorplanning may not achieve the best configurations for partitions in terms of PPA, leading to additional iterations in the design flow. Also, the macro placement process, along with subsequent cell placement and routing tasks, can serve as a critical and potentially fast evaluation step to assess each partition’s physical implementation feasibility, thereby driving continuous refinements in the floorplan. This iterative methodology is crucial in achieving a more refined and optimized chip design, which is especially critical at advanced technology nodes where wafer costs are significantly high.

In designing modern SoCs, the importance of performing high-quality floorplanning and high-quality macro placement cannot be overemphasized. Specifically, the floorplanning and the macro placement challenges encountered in the industry, and the obstacles preventing complete automation of these processes need to be re-examined. With ongoing advancements in EDA and AI/ML technologies, such as the application of reinforcement learning (RL) in tuning design flow parameters, coupled with enhanced computational power, we anticipate a substantial improvement and/or potential automation in the iterative aspects of these design processes. Such advancements will not only alleviate the workload of engineers but also enhance the overall quality of results (QoR) in chip designs.

PANEL: EDA Challenges at Advanced Technology Nodes C

  • Keh-Jeng Chang

EDA ecosystem’s fantastic supports and innovations have helped achieve better logic, memory, wafer-level packaging, and AI chips and systems [1] [2] and [3], for decades. We look forward to the continuous win-win collaborations among university scientists, circuit designers, semiconductor chip manufacturers and EDA companies in the foreseeable future. The critical issue in 2024 is what the major challenges are. A few months ago in the December 2023 IEDM in San Francisco, world-wide semiconductor experts are postulating that the complexities of the upcoming high-end electronic systems will soon be in the range of one trillion transistors. Advanced transistors used are either FinFET or GAA/nanosheet, or both, in need of design and technology co-optimization (DTCO). Innovations in the 3DFabric Alliance [2] are essential to the trillion-transistor trend. The leap from traditional SoC/IC designs to 3DFabric (3DIC) designs brings new benefits and opportunities. This new system-level design paradigm inevitably introduces new EDA challenges on system design, verification, thermal management, mechanical stress, and electrical-photonic compliance of the entire 3DIC assembly and reliability.

The author summarizes here an updated list of four major EDA challenges: (1) New post-FinFET, GAA, and nanosheet design and technology co-optimization (DTCO) [3] and physical design algorithms that can provide VLSI/SoC circuit designers with efficient APR and physics based SPICE models along with efficient RLC back annotations; (2) New, fast, and physics based chip-level and system-level SI/PI/EMC simulation tools and flows; (3) New concept-to-physical design methodology that can help achieve high-quality, user-friendly, and fast time-to-market EDA methodology for wafer-level packages of AI and data center solutions; (4) We expect silicon photonics will be mature within a few years and its EDA solution is challenging.

SESSION: Session 5: 3D ICs

Routing-aware Legal Hybrid Bonding Terminal Assignment for 3D Face-to-Face Stacked ICs

  • Siting Liu
  • Jiaxi Jiang
  • Zhuolun He
  • Ziyi Wang
  • Yibo Lin
  • Bei Yu
  • Martin Wong

Face-to-face (F2F) stacked 3D IC is a promising alternative for scaling beyond Moore’s Law. In F2F 3D ICs, dies are connected through bonding terminals whose positions can significantly impact routing performance. Further, there exists resource competition among all the 3D nets due to the constrained bonding terminal number. In advanced technology nodes, such 3D integration may also introduce legality challenges of bonding terminals, as the metal pitches can be much smaller than the sizes of bonding terminals. Previous works attempt to insert bonding terminals automatically using existing 2D commercial P&R tools and then consider inter-die connection legality, but they fail to take the legality and routing performance into account simultaneously. In this paper, we explore the formulation of the generalized assignment in the hybrid bonding terminal assignment problem. Our framework, BTAssign, offers a strict legality guarantee and an iterative solution. The experiments are conducted on 18 open-source designs with various 3D net densities and the most advanced bonding scale. The results reveal that BTAssign can achieve improvements in routed wirelength under all testing conditions from 1.0% to 5.0% with a tolerable runtime overhead.

Unified 3D-IC Multi-Chiplet System Design Solution

  • Wang-Tyng Lay

With the advancements in 2.5/3D fabrication offered by Foundry Technologies for unleashing computing power, EDA tools must adapt and take a direction to be more integrated and IC centric for multi-chiplet system design. 3D stacking introduces extra design and analysis requirements like full system planning, power, thermal analysis, cross-die STA and inter-die physical verification which have to be taken into account early during planning and implementation. In this paper, Cadence presents its technology that proactively looks ahead through integrated early analysis and addresses all aspects of 3D-IC design comprehensively from system planning, implementation, analysis and system level signoff capabilities.

Warpage Study by Employing an Advanced Simulation Methodology for Assessing Chip Package Interaction Effects

  • Jun-Ho Choy
  • Stéphane Moreau
  • Catherine Brunet-Manquat
  • Valeriy Sukharev
  • Armen Kteyan

A physics-based multi-scale simulation methodology that analyses die stress variations generated by package fabrication is employed for warpage study. The methodology combines coordinate-dependent anisotropic effective properties extractor with finite element analysis (FEA) engine, and computes mechanical stress globally on a package-scale, as well as locally on a feature-scale. For the purpose of mechanical failure analysis in the early stage of a package design, the warpage measurements were used for the tool’s calibration. The warpage measurements on printed circuit board (PCB), interposer and chiplet samples, during heating and subsequent cooling, were employed for calibrating the model parameters. The warpage simulation results on full package represented by PCB-interposer-chiplets stack demonstrate the overall good agreement with measurement profile. Performed study demonstrates that the developed electronic design automation (EDA) tool and methodology can be used for accurate warpage prediction in different types of IC stacks at early stage of package design.

Challenges for Automating PCB Layout

  • Wen-Hao Liu
  • Anthony Agnesina
  • Haoxing Mark Ren

Printed circuit board (PCB) design is typically semi-automated or fully manual. However, in recent years, the scale of PCB designs has rapidly enlarged, such that the engineering effort of manual design has increased dramatically. Therefore, the criticality of automation emerges. PCB houses are looking for productivity improvement that is contributed by automation. In this talk, the speaker will give a short tutorial about how a PCB design is done today and then indicate the challenges and opportunities for PCB design automation.

SESSION: Session 6: Artificial Intelligence and Machine Learning

FastTuner: Transferable Physical Design Parameter Optimization using Fast Reinforcement Learning

  • Hao-Hsiang Hsiao
  • Yi-Chen Lu
  • Pruek Vanna-Iampikul
  • Sung Kyu Lim

Current state-of-the-art Design Space Exploration (DSE) methods in Physical Design (PD), including Bayesian optimization (BO) and Ant Colony Optimization (ACO), mainly rely on black-boxed rather than parametric (e.g., neural networks) approaches to improve end-of-flow Power, Performance, and Area (PPA) metrics, which often fail to generalize across unseen designs as netlist features are not properly leveraged. To overcome this issue, in this paper, we develop a Reinforcement Learning (RL) agent that leverages Graph Neural Networks (GNNs) and Transformers to perform “fast” DSE on unseen designs by sequentially encoding netlist features across different PD stages. Particularly, an attention-based encoder-decoder framework is devised for “conditional” parameter tuning, and a PPA estimator is introduced to predict end-of-flow PPA metrics for RL reward estimation. Extensive studies across 7 industrial designs under the TSMC 28nm technology node demonstrate that the proposed framework FastTuner, significantly outperforms existing state-of-the-art DSE techniques in both optimization quality and runtime. where we observe improvements up to 79.38% in Total Negative Slack (TNS), 12.22% in total power, and 50x in runtime.

Methodology of Resolving Design Rule Checking Violations Coupled with Fully Compatible Prediction Model

  • Suwan Kim
  • Hyunbum Park
  • Kyeonghyeon Baek
  • Kyumyung Choi
  • Taewhan Kim

Resolving the design rule checking (DRC) violations at the pre-route stage is critically important to reduce the time-consuming design closure process at the post-route stage. Recently, noticeable methodologies have been proposed to predict DRC hotspots using Machine Learning based prediction models. However, little attention has been paid to how the predicted DRC violations can be effectively resolved. In this paper, we propose a pre-route DRC violation resolution methodology that is tightly coupled with fully compatible prediction model. Precisely, we devise different resolution strategies for two types of DRC violations: (1) pin accessibility (PA)-related and (2) routing congestion (RC)-related. To this end, we develop a fully predictable ML-based model for both PA and RC-related DRC violations, and propose completely different resolution techniques to be applied depending on the DRC violation type informed by the compatible prediction model such that for (1) PA-related DRC violation, we extract the DRC violation mitigating regions, then improve placement by formulating the whitespace redistribution problem on the regions into an instance of Bayesian Optimization problem to produce an optimal cell perturbation, while for (2) RC-related DRC violation, we manipulate the routing resources within the regions that have high potential for the occurrence of RC-related DRC violation. Through experiments, it is shown that our methodology is able to resolve the number of DRC violations by 26.54%, 25.28%, and 20.34% further on average over that by a conventional flow with no resolution, a commercial ECO router, and a state-of-the-art academic predictor/resolver, respectively, while maintaining comparable design quality.

AI for EDA/Physical Design: Driving the AI Revolution: The Crucial Role of 3D-IC

  • Erick Chao

3D Integrated Circuits (3D-ICs) represent a significant advancement in semiconductor technology, offering enhanced functionality in smaller form factors, improved performance, and cost reductions. These 3D-ICs, particularly those utilizing Through-Silicon Vias (TSVs), are at the forefront of industry trends. They enable the integration of system components from various process nodes, including analog and RF, without being limited to a single node. TSVs outperform wire-bonded System in Package (SiP) in terms of reduced (RLC) parasitics, offering better performance, more power efficiency, and denser implementation. Compared to silicon interposer methods, vertical 3D die stacking achieves higher integration levels, smaller sizes, and quicker design cycles. This presentation introduces a novel AI-driven method designed to tackle the challenges hindering the automation of 3D-IC design flows.

DSO.ai – A Distributed System to Optimize Physical Design Flows

  • Piyush Verma

The VLSI chip design process consists of a sequence of distinct steps like floor planning, placement, clock tree synthesis and routing. Each of these steps requires solving optimization problems that are often NP-hard, and the state-of-the art algorithms are not guaranteed to the optimal. Due to the compartmentalization of the design flow into distinct steps, these optimization problems are solved sequentially, with the output of first feeding into the next. This results in an inherent inefficiency, where the optimization goal of an early step problem is estimated using a fast and approximate surrogate model for the following steps. Consequently, any improvement in the step-specific optimization algorithm, while obvious at that step, is much smaller when measured at the end of the full design flow. For example, the placement step minimizes wire length. In the absence of routed nets, this wire length might be estimated by using a simple wire length model like the Steiner tree. Thus, any improvement in the placement algorithm is limited by the accuracy of the wire length estimate.

Recently, Reinforcement Learning (RL) has emerged as a promising alternative to the state-of-the-art algorithms used to solve optimization problems in placement and routing of a VLSI design [1, 2, 3]. The RL problem setup involves an agent exploring an unknown environment to achieve a goal. RL is based on the hypothesis that all goals can be described by the maximization of expected cumulative reward. The agent must learn to sense and perturb the state of the environment using its actions to derive maximal reward. Many problems in VLSI chip design can be represented as Markov Decision Problems (MDPs), where design optimization objectives are converted into rewards given by the environment and design variables are converted into actions provided to the environment. Recent advances in applying RL to VLSI implementation problems such as floor planning, standard cell layout, synthesis and placement have demonstrated improvements over the state-of-the-art algorithms. However, these improvements continue to be limited by the inaccuracies in the estimate of the optimization goal as described previously.

With DSO.ai, we have built a distributed system for the optimization of physical design flow, where multiple iterations of parallel runs are used to optimize enormous design parameter search spaces. In addition to a multiplicative improvement in human productivity, it has unlocked significant performance gains across a wide range of technology nodes. At the heart of DSO.ai’s decision engine is an implementation of RL that solves the sequential decision making and optimization problem spanning the entire design flow. Unlike prior works where RL is used for step-specific optimization within the chip design flow, DSO.ai’s RL algorithm wraps around the optimization steps to guide them via parameter choices that depend upon the optimization goal for the full flow. Thus, the quality of the final design generated by DSO.ai is no longer subject to the limitations of the compartmentalized design flow. DSO.ai’s RL algorithm views the full chip design flow as one optimization problem, where the design quality at the end of the flow is the only one that matters. To propagate the design through the design flow, DSO.ai makes parameter choices for the underlying optimization steps, which constitute its action space. By tracking the effect of these actions as a function of the design state, DSO.ai can find the optimal sequence of actions to meet the optimization goal at the end of the full flow.

In this presentation, we demonstrate how DSO.ai provides a flexible framework to integrate with existing design flows and serve the design quality needs throughout the design evolution cycle. We will also highlight how DSO.ai is allowing expert designers at Synopsys to package their knowledge into fully featured toolboxes ready to be deployed by novice designers. We will provide a summary of QoR gains that DSO.ai has delivered on advanced process nodes. Finally, we will show how DSO.ai’s decision engine is paving the way to automating the parameter choices in the chip design flow.

Solvers, Engines, Tools and Flows: The Next Wave for AI/ML in Physical Design

  • Andrew B. Kahng

It has been six years since an ISPD-2018 invited talk on “Machine Learning Applications in Physical Design”. Since then, despite considerable activity across both academia and industry, many R&D targets remain open. At the same time, there is now clearer understanding of where AI/ML can and cannot (yet) move the needle in physical design, as well as some of the difficult blockers and technical challenges that lie ahead. Some futures for AI/ML-boosted physical design are visible across solvers, engines, tools and flows – and in contexts that span generative AI, the modeling of “magic” handoffs at flow interstices, academic research infrastructure, and the culture of benchmarking and open-source EDA.

SESSION: Session 7: Second Keynote

Physical Design Challenges in Modern Heterogeneous Integration

  • Yao-Wen Chang

To achieve the power, performance, and area (PPA) target in modern semiconductor design, the trend to go for More-than-Moore heterogeneous integration by packing various components/dies into a package becomes more obvious as the economic advantages of More-Moore scaling for on-chip integration are getting smaller and smaller. In particular, we have already encountered the high cost of moving to more advanced technology and the high fabrication cost associated with extreme ultraviolet (EUV) lithography , mask, process, design, electronic design automation (EDA), etc. Heterogeneous integration refers to integrating separately manufactured components into a higher-level assembly (in a package or even multiple packages in a PCB) that provides enhanced functionality and improved operating characteristics. Unlike the on-chip designs with relatively regular components and wirings, the physical design problem for heterogeneous integration often needs to handle arbitrary component shapes, diverse metal wire widths, and different spacing requirements between components, wire metals, and pads, with multiple cross-physics domain considerations such as system-level, physical, electrical, mechanical, thermal, and optical effects, which are not well addressed in the traditional chip design flow. In this paper, we first introduce popular heterogeneous integration technologies and options, their layout modeling and physical design challenges, survey key published techniques, and provide future research directions for modern physical design for heterogeneous integration.

SESSION: Session 8: Analog

Fundamental Differences Between Analog and Digital Design Problems – An Introduction

  • Juergen Scheible

This article discusses fundamental differences between analog and digital circuits from a design perspective. On this basis one can understand why the design flows of these two circuit types differ so greatly, notably with regard to their degree of automation.

Layout Verification Using Open-Source Software

  • Andreas Krinke
  • Robert Fischbach
  • Jens Lienig

The design and manufacturing of integrated circuits is an expensive endeavor. The use of open-source software can lower the barrier to entry significantly, especially for smaller companies or startups. In this paper, we look at open-source software for layout verification, a crucial step in ensuring the consistency and manufacturability of a design. We show that a comprehensive design rule check (DRC) and layout versus schematic (LVS) check for commercial technologies is possible with open-source software in general and with KLayout in particular. To facilitate the use of these tools, we present our approach to automatically generate the required DRC scripts from a more abstract representation. As a result, we are able to generate nearly 74% of the over 1000 design rules of X-FABs XH018 180nm technology as a DRC script for the open-source software KLayout. This demonstrates the potential of using open-source software for layout verification and open-source process design kits (PDKs) in general.

Reinforcement Learning or Simulated Annealing for Analog Placement? A Study based on Bounded-Sliceline Grids

  • Mark Po-Hung Lin
  • Chou-Chen Lee
  • Yi-Chao Hsieh

Analog placement is a crucial phase in analog integrated circuit synthesis, impacting the quality and performance of the final circuits. This process involves determining the physical positions of analog building blocks while minimizing chip area and interconnecting wire-length. Existing methodologies often rely on the simulated-annealing (SA) approach, prioritizing constraints like symmetry-island, proximity, and well-island. We present a novel reinforcement learning (RL) based analog placement methodology on the bounded-sliceline grid (BSG) structure. Introducing a hierarchical clustering feature in BSG, we address well-island, proximity, and symmetry constraints. In experimental comparisons with the SA approach, our RL-based method exhibits superior placement quality across various analog circuits.

SESSION: Session 9: Placement

Practical Mixed-Cell-Height Legalization Considering Vertical Cell Abutment Constraint

  • Teng-Ping Huang
  • Shao-Yun Fang

Propelled by aggressive technology scaling, adopting mixed-cell-height design in VLSI circuits has made conventional single row-based cell legalization techniques obsolete. Furthermore, the vertical abutment constraint (VAC) among cells on consecutive rows emerges as an advanced design requirement, which has rarely been considered because the power/ground rails were sufficiently tall in conventional process nodes to isolate cells on different rows. Although there have been a number of studies on mixed-cell-height legalization, most of them cannot be trivially extended to well-tackle the general VAC due to the analytical optimization scheme. To address these issues, this work proposes the first mixed-cell-height legalization algorithm that addresses the general inter-row cell abutment constraint (i.e., VAC). The experimental results show that the proposed algorithm outperforms previous mixed-cell-height legalization works, even in the absence of the VAC. Upon applying the VAC, our algorithm offers superior performance and delivers promising results.

Multi-Electrostatics Based Placement for Non-Integer Multiple-Height Cells

  • Yu Zhang
  • Yuan Pu
  • Fangzhou Liu
  • Peiyu Liao
  • Kai-Yuan Chao
  • Keren Zhu
  • Yibo Lin
  • Bei Yu

A circuit design incorporating non-integer multi-height (NIMH) cells, such as a combination of 8-track and 12-track cells, offers increased flexibility in optimizing area, timing, and power simultaneously. The conventional approach for placing NIMH cells involves using commercial tools to generate an initial global placement, followed by a legalization process that divides the block area into row regions with specific heights and relocates cells to rows of matching height. However, such placement flow often causes significant disruptions in the initial placement results, resulting in inferior wirelength. To address this issue, we propose a novel multi-electrostatics-based global placement algorithm that utilizes the NIMH-aware clustering method to dynamically generate rows. This algorithm directly tackles the global placement problem with NIMH cells. Specifically, we utilize an augmented Lagrangian formulation along with a preconditioning technique to achieve high-quality solutions with fast and robust numerical convergence. Experimental results on the OpenCores benchmarks demonstrate that our algorithm achieves about 12% improvements on HPWL with 23.5X speed up on average, outperforming state-of-the-art approaches. Furthermore, our placement solutions demonstrate a substantial improvement in WNS and TNS by 22% and 49% respectively. These results affirm the efficiency and effectiveness of our proposed algorithm in solving row-based placement problems for NIMH cells.

IncreMacro: Incremental Macro Placement Refinement

  • Yuan Pu
  • Tinghuan Chen
  • Zhuolun He
  • Chen Bai
  • Haisheng Zheng
  • Yibo Lin
  • Bei Yu

This paper proposes IncreMacro, a novel approach for macro placement refinement in the context of integrated circuit (IC) design. The suggested approach iteratively and incrementally optimizes the placement of macros in order to enhance IC layout routability and timing performance. To achieve this, IncreMacro utilizes several methods including kd-tree-based macro diagnosis, gradient-based macro shifting and constraint-graph-based LP for macro legalization. By employing these techniques iteratively, IncreMacro meets two critical solution requirements of macro placement: (1) pushing macros to the chip boundary; and (2) preserving the original macro relative positional relationship. The proposed approach has been incorporated into DREAMPlace and AutoDMP, and is evaluated on several RISC-V benchmark circuits at the 7-nm technology node. Experimental results show that, compared with the macro placement solution provided by DREAMPlace (AutoDMP), IncreMacro reduces routed wirelength by 6.5% (16.8%), improves the routed worst negative slack (WNS) and total negative slack (TNS) by 59.9% (99.6%) and 63.9% (99.9%), and reduces the total power consumption by 3.3% (4.9%).

Timing-Driven Analytical Placement According to Expected Cell Distribution Range

  • Jai-Ming Lin
  • You-Yu Chang
  • Wei-Lun Huang

Since the multilevel framework with the analytical approach has been proven as a promising method to handle the very-large-scale integration (VLSI) placement problem, this paper presents two techniques including a pin-connectivity-aware cluster score function and identification of expected object distribution ranges to further improve the coarsening and refinement stages of this framework. Moreover, we extend the proposed analytical placement method to consider timing in order to speed up design convergence. To optimize timing without increasing wirelength, our approach only increases the weights of timing-critical nets, where the weight of a net is estimated according to the associated timing slack and degree. Besides, we propose a new equation to update net weights based on their historical values to maintain the stability of the net-based timing-driven placement approach. Experimental results demonstrate that the proposed analytical placement approach with new techniques can actually improve wirelength of the classic approach. Moreover, our TDP can get much better WNS and TNS than the previous timing-driven placers such as DREAMPlace4.0 and Differentiable TDP.

SESSION: Session 10: Standard Cell, Routability, and IR drop

Routability Booster ” Synthesize a Routing Friendly Standard Cell Library by Relaxing BEOL Resources

  • Bing-Xun Song
  • Ting Xin Lin
  • Yih-Lang Li

In recent years, the accessibility of pins has become a focal point for cell design and synthesis research. In this study, we propose a novel approach to improve routability in upper-level routing by eliminating one M1 track during cell synthesis. This creates space for accommodating upper-level routing, leading to improved routability. We achieve consolidated routability of transistor placement by integrating fast track assignment with dynamic programming-based transistor placement. Additionally, we introduce a hybrid routing algorithm that identifies an optimal cell routing territory for each net. This optimal territory facilitates subsequent Steiner Minimum Tree (SMT) solutions for mixed-integer linear programming (MILP) and constrains the routing region of MILP, resulting in accelerated execution. The proposed MILP approach enables concurrent routing planning and pin metal allocation, effectively resolving the chicken-or-egg causality dilemma. Experimental results demonstrate that, when using the routing-friendly synthesized cell library, the routing quality in various designs surpasses that achieved with a handcrafted cell library in ASAP7 PDK. This improvement is evident in metrics such as wirelength, number of vias, and design rule check (DRC) violations.

Novel Transformer Model Based Clustering Method for Standard Cell Design Automation

  • Chia-Tung Ho
  • Ajay Chandna
  • David Guan
  • Alvin Ho
  • Minsoo Kim
  • Yaguang Li
  • Haoxing Ren

Standard cells are essential components of modern digital circuit designs. With process technologies advancing beyond 5nm, more routability issues have arisen due to the decreasing number of routing tracks (RTs), increasing number and complexity of design rules, and strict patterning rules. The standard cell design automation framework is able to automatically design standard cell layouts, but it is struggling to resolve the severe routability issues in advanced nodes. As a result, a better and more efficient standard cell design automation method that can not only resolve the routability issue but also scale to hundreds of transistors to shorten the development time of standard cell libraries is highly needed and essential.

High quality device clustering with the considerations of routability in the layouts of different technology nodes can reduce the complexity and assist finding the routable layouts faster. In this paper, we develop a novel transformer model-based clustering methodology – training the model using LVS/DRC clean cell layouts and leveraging the personalized page rank vectors to cluster the devices with the attentions to netlist graph and learned embeddings from the actual LVS/DRC clean layouts. On a benchmark of 94 complex and hard-to-route standard cells, the proposed method not only generates 15% more LVS/DRC clean layouts, but also achieves average 12.7× faster than previous work. The proposed method can generate 100% LVS/DRC clean cell layouts over 1000 standard cells and achieve 14.5% smaller cell width than an industrial standard cell library.

Power Sub-Mesh Construction in Multiple Power Domain Design with IR Drop and Routability Optimization

  • Chien-Pang Lu
  • Iris Hui-Ru Jiang
  • Chung-Ching Peng
  • Mohd Mawardi Mohd Razha
  • Alessandro Uber

Multiple power domain design is prevalent for achieving aggressive power savings. In such design, power delivery to cross-domain cells poses a tough challenge at advanced technology nodes because of the stringent IR drop constraint and the routing resource competition between the secondary power routing and regular signal routing. Nevertheless, this challenge was rarely mentioned and studied in recent literature. Therefore, in this paper, we explore power sub-mesh construction to mitigate the IR drop issue for cross-domain cells and minimize its routing overhead. With the aid of physical, power, and timing related features, we train one IR drop prediction model and one design rule violation prediction model under power sub-meshes of various densities. The trained models effectively guide sub-mesh construction for cross-domain cells to budget the routing resource usage on secondary power routing and signal routing. Our experiments are conducted on industrial mobile designs manufactured by a 6nm process. Experimental results show that IR drop of cross-domain cells, the routing resource usage, and timing QoR are promising after our proposed methodology is applied.

SESSION: Session 11: Thermal Analysis and Packaging

Introduction of 3D IC Thermal Analysis Flow

  • Alex Hung

Thermal Challenge from modeling heterogeneous 2.5/3D IC-package is important for several reasons. Designing a large high power device, e.g. a AI or HPC processor without considering how to get the heat out is likely to lead to problems later on, resulting in a sub-optimal packaging solution from cost, size, weight and performance perspectives. Thermal simulation combines with physical verification. The benefits are enablement for automatic extraction, power map generation and simulation of the complete 3D IC assembly, viewing thermal map, and addressing hotspot. Make the IC design flow aware temperature and hotspot at the early design stage.

3Dblox: Unleash the Ultimate 3DIC Design Productivity

  • Jim Chang

The 3DIC design world is blooming with new ideas and new possibilities. With TSMC’s 3DFabricTM technology, new opportunities in architectural innovation have led to superior system performance and density. However, what comes with the new opportunities is the continuous rise in design complexity.

In this talk, we will introduce 3Dblox, our latest invention to ease the design complexity challenge. The 3Dblox is an innovative design language that modularizes the complex 3DIC structures to streamline the design flow, and is open and free to all industry participants. All 4 EDA vendors, including Cadence, Synopsys, Ansys, and Siemens have actively participated in this effort to provide a unified design ecosystem to unleash the ultimate 3DIC design productivity.

Enabling System Design in 3D Integration: Technologies and Methodologies

  • Hung-Ming Chen

3D integration solutions have been called for in the semiconductor market for a long time to possibly substitute the place of technology scaling. It consists of 3D IC packaging, 3D IC integration, and 3D silicon integration. 3D IC packaging has been in the market, but 3D IC and silicon integrations have obtained more attention and care due to modern system requirements on high performance computing and edge AI applications. In the need of further integration in electronics system development at lower cost, chip and package design are therefore evolving along the years [11].

First in technologies, except through-silicon-via (TSV) technology being applied for 3D stacking ICs, hybrid bonding (also called Cu-to-Cu direct bonding) and microbump bonding are used in silicon integration/advanced packaging, they were presented in [6, 7, 11], to name a few. They are prepared for chiplet-based (some papers called multi-die-based) integration. Manufacturing concerns such as thermal or stress-induced difficulties are still with us, how we use/adapt modern technologies wisely/cost-effectively has been our continuing mission. Besides, the materials for the integration such as glass instead of silicon for interposer can possibly cut us some slack [9].

Second in methodologies, since they are closely coupled with the technologies, tool vendors have been worked closely with foundries and design houses for more effective solutions. Among those efforts, die-to-die (D2D) implementation and interface are considered the top priority for system realization. Through the development history of redistribution layer (RDL) routing in different period, to the standard creation in universal chiplet interconnect express (UCIe) [8], we need to figure out ways to achieve more efficient and cost effective methods.

Last but not least, the focus of the integration has been system specification itself. Either the system is in edge or cloud applications will make a lot of difference. There have been some efforts in chiplet-based LLM cloud chips such as [3, 4]. Edge AI accelerators should be chipletized as well, especially for coming trends in smaller language model applications. Although there have been prior research for tools such as [1], we need more efforts on edge AI design automation. In all, system-technology co-design (STCO) concept [5, 10] is another perspective to approach the best implementation for different applications.

In our first attempt of the series manuscripts [2], we have introduced some perspectives on 3D integration for system designs; in this talk, we continue to depict the future of 3D integration as we know of, plus new observations, technologies, and methodologies such as programmable package, building-block-based multi-chiplet methodology.

SESSION: Session 12: Lifetime Achievement Session

Scheduling and Physical Design

  • Jason Cong

In a typical integrated circuit electronic design automation (EDA) flow, scheduling is a key step in high-level synthesis, which is the first stage of the EDA flow that synthesizes a cycle-accurate register transfer level (RTL) from the given behavior description, while physical design is the last stage in the EDA flow that generates the final geometric layout of the transistors and wires for fabrication. As a result, scheduling and physical design are usually carried out independently. In this paper, I discuss multiple research projects that I have been involved with, where the interaction between scheduling and physical design are shown to be highly beneficial. I shall start with my very first paper in EDA on multi-layer channel routing which benefited from an unexpected connection to the optimal two-processor scheduling algorithm, a joint work with Prof. Martin Wong, who is being honored at this conference for the 2024 ISPD Lifetime Achievement Award. Then, I shall further demonstrate how scheduling can help to overcome interconnect bottleneck, enable parallel placement and routing, and, finally, play a key role in layout synthesis for quantum computing.

Accelerating Physical Design from 1 to N

  • Evangeline F.Y. Young

Today, we have abundant parallel computing resources, while most EDA tools are still running sequentially. It is interesting to see how physical design can be advanced by leveraging this massive parallel computing power. To achieve significant speedup, it is usually not simply running the same sequential method a few copies in parallel. Innovative parallel algorithms that solve the problem from a new perspective using different mechanisms are needed. We will look at a few examples in physical design and logic synthesis in this talk to illustrate some methodologies and techniques in parallelizing design automation.

Pioneering Contributions of Professor Martin D. F. Wong to Automatic Floorplan Design

  • Ting-Chi Wang

Professor Martin D. F. Wong is well recognized as a distinguished figure in the community of physical design, owing to his numerous and noteworthy contributions. This talk aims to highlight his pioneering works in the field of automatic floorplan design. Professor Wong’s profound insights and innovative approaches have not only propelled advancements in the field but also served as an inspirational source for other researchers.

ISPD 2024 Lifetime Achievement Award Bio

  • Martin D F Wong

The 2024 International Symposium on Physical Design lifetime achievement award goes to Professor Martin D F Wong for his oustanding contributions in the field.

SESSION: Session 13: Third Keynote

Computing Architecture for Large-Language Models (LLMs) and Large Multimodal Models (LMMs)

  • Bor-Sung Liang

Large-language models (LLMs) have achieved remarkable performance in many AI applications, but they require large parameter size in their models. The parameter size ranges from several billions to trillion parameters, and results in huge computation requirements on both training and inference. General speaking, LLMs increasing more parameters are to explore “Emergent Abilities” for AI models. On the other hands, LLMs with fewer parameters are to reduce computing burden to democratize generative AI applications.

To fulfill huge computation requirement, Domain Specific Architecture is important to co-optimize AI models, hardware, and software designs, and to make trade-offs among different design parameters. Besides, there are also trade-offs between AI computation throughput and energy efficiency on different types of AI computing systems.

Large Multimodal Models (LMMs), also called Multimodal Large Language Models, integrates multiple data types as input. Multimodal information can provide rich and or environment information for LMMs to generate better user experience. LMM is also a trend for mobile devices, because mobile devices often connect with many sensors, such as video, audio, touch, gyro, navigation system, etc.

Recently, there is a trend to run smaller LLMs/LMMs (near or less than 10 billion parameters) on edge device-side, such as Llama 2, Gemini Nano, Phi-2, etc. It shines a light to apply LLMs/LMMs in mobile devices. Several companies provided experimental solutions on edge devices, such as smartphone and PC. Even LLMs/LMMs model size are reduced, they still require more computing resources than previous mobile processor workloads, and face challenges on memory size, bandwidth, and power efficiency requirements.

Besides, device-side LLMs/LMMs in mobile processors can collaborate with cloud-side LLMs/LMMs in the data center to deliver better performance. They can off-load computing from cloud-side models to provide seamless response, or to become an agent to prompt cloud-side LLMs/LMMs, or be fine-tuned locally by user data to keep privacy.

Those LLMs/LMMs trends and new usage scenarios will shape future computing architecture design. In this talk we will discuss those issues, and especially their impacts on mobile processor design.

SESSION: Session 14: Quantum and Superconducting Circuits

SMT-Based Layout Synthesis Approaches for Quantum Circuits

  • Zi-Hao Guo
  • Ting-Chi Wang

The physical qubits in current quantum computers do not all interact with each other. Therefore, in executing a quantum algorithm on an actual quantum computer, layout synthesis is a crucial step that ensures that the synthesized circuit of the quantum algorithm can run smoothly on the quantum computer. In this paper, we focus on a layout synthesis problem for quantum circuits and improve a prior work, TB-OLSQ, which adopts a transition-based satisfiability modulo theories (SMT) formulation. We present how to modify TB-OLSQ to obtain an accelerated version for runtime reduction. In addition, we extend the accelerated version by considering gate absorption for better solution quality. Our experimental results show that compared with TB-OLSQ, the accelerated version achieves 121X speedup for a set of SWAP-free circuits and 6X speedup for the other set of circuits with no increase in SWAP gates. In addition, the accelerated version with gate absorption helps reduce the number of SWAP gates by 38.9% for the circuits requiring SWAP gates, while it is also 3X faster.

Satisfiability Modulo Theories-Based Qubit Mapping for Trapped-Ion Quantum Computing Systems

  • Wei-Hsiang Tseng
  • Yao-Wen Chang
  • Jie-Hong Roland Jiang

Qubit mapping is crucial in optimizing the performance of quantum algorithms for physical executions on quantum computing architectures. Many qubit mapping algorithms have been proposed for superconducting systems recently. However, due to their limitations on the physical qubit connectivity, costly SWAP gates are often required to swap logical qubits for proper quantum operations. Trapped-ion systems have emerged as an alternative quantum computing architecture and have gained much recent attention due to their relatively long coherence time, high-fidelity gates, and good scalability for multi-qubit coupling. However, the qubit mapping of the new trapped-ion systems remains a relatively untouched research problem. This paper proposes a new coupling constraint graph with multi-pin nets to model the unique constraints and connectivity patterns in one-dimensional trapped-ion systems. To minimize the time steps for quantum circuit execution satisfying the coupling constraints for trapped-ion systems, we devise a divide-and-conquer solution using Satisfiability Modulo Theories for efficient qubit mapping on trapped-ion quantum computing architectures. Experimental results demonstrate the superiority of our approach in scalability and effectiveness compared to the previous work.

Optimization for Buffer and Splitter Insertion in AQFP Circuits with Local and Group Movement

  • Bing-Huan Wu
  • Wai-Kei Mak

Adiabatic quantum-flux parametron (AQFP) is a superconducting technology with extremely low power consumption compared to traditional CMOS structure. Since AQFP logic gates are all clocked by AC current, extra buffer cells are required for balancing the length of data paths. Furthermore, since the output current of an AQFP logic gate is too weak to drive more than one gate, splitter cells are needed for branching the output signals of multi-fanout gates. For an AQFP circuit, the total number of additional buffers and splitters may be much more than the number of logic gates (up to 9 times in the benchmark circuits after optimization), which would greatly impact the power, performance, and area of the circuit. In this paper, we propose several techniques to (i) reduce the total number of required buffers and splitters, and (ii) perturb the levels of logic gates in order to seek more optimization opportunities for buffer and splitter reduction. Experimental results shows that our approach has better quality with comparable runtime compared to a retiming-based method from ASP-DAC’23. Moreover, our approach has quality which is on equal footing with the integer linear programming-based method also from ASP-DAC’23.

SESSION: Session 15: Physical Design Challenges for Automotive

Design Automation Challenges for Automotive Systems

  • Chung-Wei Lin

As vehicular technology advances, vehicles become more connected and autonomous. Connectivity provides the capability to exchange information between vehicles, and autonomy provides the capability to make decisions and control each vehicle precisely. Connectivity and autonomy realize many evolutional applications, such as intelligent intersection management and cooperative adaptive cruise control. Electric vehicles are sometimes combined to create more use cases and business models. However, these intelligent features make the design process more complicated and challenging. In this talk, we introduce several examples of automotive design automation, which is required to improve the design quality and facilitate the design process. We mainly discuss the rising incompatibility issue, where different original equipment manufacturers and suppliers are developing systems, but the designs are confidential and thus incompatible with other players’ designs. The incompatibility issue is especially critical with autonomous vehicles because no human driver resolves incompatible scenarios. We believe that techniques and experiences in electronic design automation can provide insights and solutions to automotive design automation.

Physical Design Challenges for Automotive ASICs

  • Goeran Jerke

The design of automotive ASICs faces several key challenges that mainly arise from the harsh environmental operating conditions, specific functional loads, cost pressure, safety requirements, and the steady progress of the automotive-grade semiconductor technologies that are unique to automotive applications.

The talk first highlights these key differences between the design approaches for automotive and non-automotive ASIC designs. It also addresses why automotive ASIC designs prefer larger and more mature nodes compared to leading-edge non-automotive ASIC designs. In addition, the talk introduces several automotive-specific physical design problems and essential solutions for design implementation, direct-verification and meta-verification to address them. Finally, the talk provides an outlook of several related and yet-unsolved challenges in the physical design domain.

Solving the Physical Challenges for the Next Generation of Safety Critical & High Reliability Systems

  • Rob Knoth

Silicon systems have been part of automobiles for a long time. The physical design methodology to address the quality, reliability, and safety challenges of these systems are common knowledge in the leading automotive semiconductor companies. The rise of trends like autonomous driving (ADAS), software defined vehicles (SDV) and the electrification of our transportation network are giving rise to not only new levels of these challenges, but also many new players in the automotive semiconductor space. The same forces of opportunity which are transforming our society are also the foundation of a transformation in automotive semiconductor design: massive improvements in accelerated compute, 3DIC and chiplet based design, digital twins, and artificial intelligence (AI). We’ll discuss how these forces are helping modern automotive semiconductor design and highlight how the electronic design automation (EDA) industry can apply successful principles from earlier eras to these new challenges.

SESSION: Session 16: Contest Results and Closing Remarks

GPU/ML-Enhanced Large Scale Global Routing Contest

  • Rongjian Liang
  • Anthony Agnesina
  • Wen-Hao Liu
  • Haoxing Ren

Modern VLSI design flows demand scalable global routing techniques applicable across diverse design stages. In response, the ISPD 2024 contest pioneers the first GPU/ML-enhanced global routing competition, selecting advancements in GPU-accelerated computing platforms and machine learning techniques to address scalability challenges. Large-scale benchmarks, containing up to 50 million cells, offer test cases to assess global routers’ runtime and memory scalability. The contest provides simplified input/output formats and performance metrics, framing global routing challenges as mathematical optimization problems and encouraging diverse participation. Two sets of evaluation metrics are introduced: the primary one concentrates on global routing applications to guide post-placement optimization and detailed routing, focusing on congestion resolution and runtime scalability. Special honor is given based on the second set of metrics, placing additional emphasis on runtime efficiency and aiming at guiding early-stage planning.

Who’s Jianlie Yang

Jan 2024

Jianlie Yang

Associate Professor

Beihang University

Email:

jianlie@buaa.edu.cn

Personal webpage

https://shi.buaa.edu.cn/jianlei/zh_CN/index.htm

Research interests

EDA, Systems and Architectures for AI/DL

Short bio

Prof. Jianlei Yang received the B.S. degree in microelectronics from Xidian University, Xi’an, China, in 2009, and the Ph.D. degree in computer science and technology from Tsinghua University, Beijing, China, in 2014. He is currently an Associate Professor in Beihang University, Beijing, China, with the School of Computer Science and Engineering. From 2014 to 2016, he was a post-doctoral researcher with the Department of ECE, University of Pittsburgh, Pennsylvania, USA. Prof. Jianlei Yang was the recipient of the First/Second place on ACM TAU Power Grid Simulation Contest in 2011/2012. He was a recipient of IEEE ICCD Best Paper Award in 2013, ACM GLSVLSI Best Paper Nomination in 2015, IEEE ICESS Best Paper Award in 2017, ACM SIGKDD Best Student Paper Award in 2020.

Research highlights

His research contributions primarily encompass large-scale integrated circuit simulation and analysis algorithms, emerging memory design and computation paradigms, as well as software-hardware co-design for machine learning and graph learning applications.

  • In the field of large-scale circuit simulation solvers, he has tackled the pressing need for efficient and rapid solutions for power integrity analysis in high-performance processor chip designs. His FPS-PCG and AMG-PCG algorithms have significantly enhanced the convergence of power grid solvers [ICCAD’2011/TVLSI’2014]. These accomplishments were acknowledged with honors, including first place in the 2011 TAU Power Grid Simulation Contest and second place in the 2012 TAU Power Grid Transient Simulation Contest. Furthermore, his proposed selective inverse-based vectorless verification techniques have notably improved the efficiency of large-scale vectorless verification of power supply networks [ICCD’2013/TVLSI’2015], earning him the IEEE ICCD Best Paper Award in 2013.
  • In the realm of emerging memory design and computation paradigms, he has confronted the power consumption challenges of integrated circuits in the Post-Moore’s Law era. His systematic research has resulted in the development of reliability analysis methods for spin-based memory devices and circuits, establishing theoretical foundations and solutions for non-volatile memory reliability verification [TCAD’2016]. Additionally, his work on spin-based reconfigurable logic has tackled efficiency and reliability challenges in circuit obfuscation methods within the field of chip security [TCAD’2019]. By leveraging the inherent random flipping characteristics of spin devices, he has introduced spin-based random computing circuits for Bayesian inference computation systems, significantly enhancing inference efficiency [ASPDAC’2018/TCAD’2020]. Moreover, his innovative spin-based in-memory computing architecture has addressed data access bottlenecks in graph computing, resulting in substantial improvements in graph computation efficiency [DAC’2020].
  • In the domain of software-hardware co-design, he has delved into domain-specific custom computing architectures as a means to enhance computational efficiency. His proposed dataflow optimization techniques involve deeply customizing the dataflow execution patterns of neural network models to augment hardware-friendly characteristics [TC’2022]. He pioneered the early introduction of the dynamic gradient sparsification method to accelerate deep neural network training, in conjunction with specified architecture support for algorithm-hardware co-design [ECCV’2020/DAC’2020]. Furthermore, his dataflow customization methods and dedicated architectures for large-scale visual point cloud processing have mitigated memory access bottlenecks, resulting in substantial enhancements in the speed and energy efficiency of visual computing tasks [DAC’2019/DAC’2022].

Service

Service Awards

SIGDA has restructured its service awards, and will be giving two annual service awards.

  • Distinguished Service Award: The SIGDA Distinguished Service Award is given to individuals who have dedicated many years of their career in extraordinary services to promoting, leading, or creating ACM/SIGDA programs or events.
  • Meritorious Service Award: The SIGDA Meritorious Service Award is given to individuals who have performed professional services above and beyond traditional service to promoting, leading, or creating ACM/SIGDA programs or events.

At any given year, the number of Distinguished Service Award will be up to 2, and the number of Meritorious Service Award will be up to 4.

Nominations should consist of:

  • Award type being nominated.
  • Name, address, phone number and email of person making the nomination.
  • Name, affiliation, address, email, and telephone number of the nominee for whom the award is recommended.
  • A statement (between 200 and 500 words long) explaining why the nominee deserves the award. Note that the award is given for service that goes above and beyond traditional services.
  • Up to 2 additional letters of support. Include the name, affiliation, email address, and telephone number of the letter writer(s). Supporters of multiple candidates are strongly encouraged to compare the candidates in their letters.

Note that the nominator and reference shall come from active SIGDA volunteers. Deadline of the nomination every year: March 15 (Except 2019, May 5).

Please send all your nomination materials as one pdf file to SIGDA-Award@acm.org before the deadline.

Distinguished Service Awards

2023Tulika Mitra, National University of Singapore
For her leadership in major SIGDA conferences such asgeneral chair for ICCAD and ESWEEK“.
Patrick Groeneveld, Stanford University
For his multi-year significant contribution to the EDA community, such as DAC finance chair among many other“.
2022Vijay Narayanan, The Pennsylvania State University
For Extraordinary Dedication and Leadership to SIGDA“.
Harry Foster, Siemens EDA
For Extraordinary Dedication and Persistence in Leading DAC during Pandemic“.
2021Deming Chen, University of Illinois Urbana-Champaign
For distinguished contributions to the design automation and reconfigurable computing communities“.
Evangeline F. Y. Young, Chinese University of Hong Kong
For outstanding leadership in promoting diversity in the ACM/SIGDA community“.
2020Sri Parameswaran, University of New South Wales
“For leadership and distinguished service to the EDA community“.
2019Naehyuck Chang, Korea Advanced Institute of Science and Technology
For many years of impactful service to ACM/SIGDA in various leadership positions“.
Sudeep Pasricha, Colorado State University
For a decade of outstanding service to ACM/SIGDA in various volunteer positions
2018Chuck Alpert, Cadence Design Systems
For significant contributions to DAC”.
Jörg Henkel, Karlsruhe Institute of Technology
For leading SIGDA efforts in Europe and DATE”.
Michael ‘Mac’ McNamara, Adapt-IP
For sustained contributions to the design automation community and DAC”.
Michelle Clancy, Cayenne Communication
For sustained contributions to the community, especially DAC”.
2016Steven Levitan
In recognition of a lifetime of devoted service to ACM SIGDA and the Electronic Design Automation community.
2015Tatsuo Ohtsuki, Waseda University
Hiroto Yasuura, Kyushu University
Hidetoshi Onodera, Kyoto University
For their distinguished contributions to the Asia and South Pacific Design automation Conference (ASPDAC) as well as their many years of dedicated service on the conference’s steering committee
Massoud Pedram, University of Southern California
For his many years of service as Editor in Chief of the ACM Transactions on Design Automation of Electronic Systems (TODAES)
2014Peter Marwedel, Technical University of Dortmund
For his muiltiple years of service starting and maintaining the DATE PhD Forum
2012Joe Zamreno, Iowa State University
Baris Taskin, Drexel University
2011Peter Feldman, IBM
Radu Marculescu, CMU
Qinru Qiu, Syracuse University
Martin Wong, UIUC
Qing Wu, Air Force Rome Labs
2010Alex K. Jones
For dedicated service to ACM/SIGDA and the Design Automation Conference as director of the University Booth
Matt Guthaus
For dedicated service as director of SIGDA CADathlon at ICCAD program and Editor-in-Chief of the SIGDA E-Newsletter
Diana Marculescu
For dedicated service as SIGDA Chair, and contributions to SIGDA, DAC and the EDA Profession
2009Nikil Dutt
For contributions to ACM’s Special Interest Group on Design Automation during the past fifteen years as a SIGDA officer, coordinator of the University Booth in its early years, and most recently, as Editor-in-Chief of the ACM Transactions on Design Automation of Electronic Systems
2008SungKyu Lim
For his contributions to the DAC University Booth.”
2007Richard Goering
For his contributions as EE Times Editorial Director for Design Automation for more than two decades
Gary Smith
For his contributions as Chief EDA Analyst at Gartner Dataquest for almost two decades.”
Daniel Gajski
Mary Jane Irwin
Donald E. Thomas
Chuck Shaw

For outstanding contributions to the creation of the SIGDA/DAC University Booth, on the occasion of its 20th edition.”
Soha Hassoun
Steven P. Levitan

For outstanding contributions to the creation of the SIGDA Ph.D. Forum at DAC on the occasion of its 10th edition.”
Richard Auletta
For over a decade of service to SIGDA as University Booth Coordinator, Secretary/Treasurer, and Executive Committee Member-at-Large.
2006Robert Walker
For dedicated service as SIGDA Chair (2001 – 2005), and over a decade of service to SIGDA, DAC and the EDA profession.”
2005Mary Jane Irwin
For dedicated service as Editor in Chief of ACM Journal, TODAES (1998 – 2004), and many years of service to SIGDA, DAC, and the EDA profession.”
2004James P. Cohoon
For exemplary service to SIGDA, to ACM, to DAC, and to the EDA profession as a whole
2003James Plusquellic
For exemplary service to ACM/SIGDA and the Design Automation Conference as director of the University Booth program
2002Steven P. Levitan
For over a decade of service to ACM/SIGDA and the EDA industry — as DAC University Booth Coordinator, Student Design Contest organizer, founder and promoter of SIGDA’s web server, and most recently, Chair of ACM/SIGDA from 1997 to 2001.
Cheng-Kok Koh
For exemplary service to ACM/SIGDA and the EDA industry — as Co-director of SIGDA’s CDROM Project, as SIGDA’s Travel Grants Coordinator, and as Editor of the SIGDA Newsletter.”
2001Robert Grafton
For contributions to the EDA profession through his many years as the Program Director of NSF’s Design, Tools, and Test Program of the computer, Information Sciences & Engineering Directorate. In this position, he provided supervision, mentorship, and guidance to several generation of EDA tool designers and builders funded by grants from the National Science Foundation.”
2000Massoud Pedram
For his contributions in developing the SIGDA Multimedia Series and organizing the Young Student Support Program
Soha Hassoun
For developing the SIGDA Ph.D. Forum
1999C.L. (Dave) Liu
For his work in founding our flagship journal ACM/TODAES

Meritorious Service Awards

2023Robert Wille, Technical University of Munich
For his leading positions in major ACM SIGDA conferences, including Executive Committee of DATE, ICCAD and Chair of the PhD Forum at DAC and DATE“.
Lei Jiang, Indiana University Bloomington
For his leadership and contribution to SIGDA student research forums (SRFs) at ASP-DAC“.
Hui-Ru Jiang, National Taiwan University
For her continuous contribution to SIGDA PhD Forum at DAC and many other events“.
Jeyavijayan (JV) Rajendran, Texas A&M University
For his leadership in co-founding and organizing Hack@DAC, the largest hardware security competition in the world“.
2022Jeff Goeders, Brigham Young University
For Chairing System Design Contest @ DAC for the Past 3 Years“.
Cheng Zhuo, Zhejiang University
For the Leading Efforts to the Success of SRC@ICCAD and SDC@DAC as Chairs for the past5 years, and the Sustained Contributions to the EDA Community in China“.
Tsung-Wei Huang, University of Utah
For Chairing CADathlon and CAD Contests at ICCAD for Three Years. These Activities Have Engaged Hundreds of Students into CAD Research“.
Yiyu Shi, University of Notre Dame
For Outstanding Services in Leading SIGDA Educational Efforts“.
2021Bei Yu, Chinese University of Hong Kong
For service as SIGDA Web Chair from 2016 to 2021, SIGDA Student Research Competition Chair in 2018 and 2019, and other SIGDA activities“.
2020Aida Todri-Sanial, LIRMM/University of Montpellier
“For service as Co-Editor-in-Chief of SIGDA e-Newsletter from 2016 to 2019 and other SIGDA activities“.
Yu Wang, Tsinghua University
For service as Co-Editor-in-Chief of SIGDA e-Newsletter from 2017 to 2019 and other SIGDA activities”.
2019Yinhe Han, Chinese Academy of Sciences
For outstanding effort in promoting EDA and SIGDA events in China
Jingtong Hu, University of Pittsburgh
For contribution to multiple SIGDA education and outreach activities
Xiaowei Xu, University of Notre Dame
For contribution to the 2018 System Design Contest at ACM/IEEE Design Automation Conference
2015Laleh Behjat, University of Calgary
For service as chair of the SIGDA PhD forum at DAC
Soonhoi Ha, Seoul National University
Jeonghee Shin, Apple
For their service as co-chairs of the University Booth at DAC
1998Jason Cong
Bryan Preas
Kathy Preas
Chong-Chian Koh
Cheng-Kok Koh

For contributions in producing SIGDA CD ROM’s – Archiving the knowledge of the Design Automation Community
1997Robert Walker
For his hard work as Secretary/Treasurer and University Booth Coordinator
1996Debbie Hall
For serving as ACM Program Director for SIGDA for the past 6 years

Following are some awards no longer being given:

Technical Leadership Awards

2013Jarrod Roy
Sudeep Pasricha
Sudarshan Banerjee
Srinivas Katkoori

for running CADathlon
2012Cheng Zhuo
Steve Burns
Amin Chirayu
Andrey Ayupov
Gustavo Wilke
Mustafa Ozdal
2011Raju Balasuramanian
Zhuo Li
Frank Liu
Natarajan Viswanathan
2010Cliff Sze
For contributions to the ISPD Physical Design contest, and promoting research in physical design.”
2008Hai Zhou
For contributions to the SIGDA E-Newsletter (2005-2008)
Jing Yang
For contributions to the SIGDA Ph.D. Forum at DAC (2004-2008)
2007Geert Janssen
For contributions to the SIGDA CADathlon
Tony Givargis
For contributions to the SIGDA Ph.D. Forum at DAC (2005-2007)
Gi-Joon Nam
For contributions to the Physical Design Contest at ISPD (2005-2007)
2006Kartik Mohanram
For contributions to the SIGDA CADathlon at ICCAD (2004-2005)
Ray Hoare
For contributions to the SIGDA University Booth at DAC (2004-2006)
Radu Marculescu
For contributions to the SIGDA Ph.D. Forum at DAC (2004-2006)
Frank Liu
For contributions to the SIGDA Ph.D. Forum at DAC (2005-2006)
2005Florian Krohm
For contributions to the SIGDA CADathlon at ICCAD
R. Iris Bahar
Igor Markov

For contributions to the SIGDA E-Newsletter
2004Robert Jones
For contributions to the SIGDA Ph.D. Forum at DAC
2003Diana Marculescu
For contributions to the SIGDA Ph.D. Forum at DAC
2002Geert Janssen
For contributions to the SIGDA CADathlon at ICCAD
Pai Chou
Abe Elfadel
Olivier Coudert
Soha Hassoun

For contributions to the SIGDA Ph.D. Forum at DAC

1996 SIGDA Lifetime Achievement Award

Paul Weil
“For contributions to SIGDA for 13 years, most recently being in charge of Workshops.”

1996 SIGDA Leadership Award

Steve Levitan
“For Newsletter and Electronic Publishing”

1995 SIGDA Outstanding Leadership Award

Charlotte Acken
“In recognition of her efforts in managing the high school scholarship program”

1994 SIGDA Outstanding Leadership Award

Pat Hefferan
“As Editor of the SIGDA Newsletter”

Call for SIGDA Newsletter Editor-in-Chief

ACM SIGDA announces the call for Editor-in-Chief for the SIGDA Newsletter, a monthly publication for news and event information in the design automation area. The Editor-in-Chief, along with the editorial board consisting of associate editors, is responsible for collecting and compiling information, as well as composing and disseminating the monthly newsletter to the SIGDA community. Please refer to the following URL for more information about the newsletter content:

Responsibility: The Editor-in-Chief role requires the initial formation of the editorial board, and the assigning of roles to and close co-ordination with several Associate Editors in charge of the different newsletter sections, including headlines, “What is” column, recent events and awards, technical activities, upcoming submission deadlines, and job positions. The Editor-in-Chief will be appointed with effect from 1 January 2024 for an initial period of two years.

Qualifications: The candidate must be an active and respected member of the SIGDA community, as evidenced by participation in recent conferences, journals, and events associated with SIGDA. It is important that the candidate be willing to devote the time required for consistent and punctual publication of the newsletter every month, over several years.

Application: Interested community members are requested to send an email to the ACM SIGDA Communications Chair, Preeti Ranjan Panda (panda@cse.iitd.ac.in), with a CV and 1-2 paragraphs indicating their interest in and willingness to devote the time required for successful operation of the newsletter activity. Applications are due by 15 November 2023.

HACK@DAC

HACK@DAC is a hardware security challenge contest, co-located with the Design and Automation Conference (DAC), for finding and exploiting security-critical vulnerabilities in hardware and firmware. In this competition, participants compete to identify the security vulnerabilities, implement the related exploits, propose mitigation techniques or patches, and report them. The participants are encouraged to use any tools and techniques with a focus on theory, tooling, and automation.

       The contest mimics real-world scenarios where security engineers have to find vulnerabilities in the given design. The vulnerabilities are diverse and range from data corruption to leaking sensitive information leading to compromise of the entire computing platform. The open-source SoC riddled with security vulnerabilities has been co-developed by Intel, the Technical University of Darmstadt, and Texas A&M University. HACK@DAC has been successfully running since 2018 with several hundred contestants from academia and industry.

       HACK@DAC is the first-of-its-kind security competition focusing on SoC Security first launched at the Design Automation Conference in 2018. Since then, it attracted massive interest and participations by teams across the globe and evolved into a franchise called HACK@EVENT with annual presence at top-tier industry and academic conferences.

       HACK@DAC 2023 consists of 2 phases, a Qualifying Round and a Live Round. The Qualifying Round was open to all participants and ended in mid-May. Top teams from the Qualifying Round were invited to participate in the Live Competition on July 9-10 at the Design Automation Conference (#DAC60) in San Francisco, CA. Teams had 48 hours to find and exploit as many security vulnerabilities in the “buggy SoC” provided by the competition organizers. Points were awarded to teams that correctly identified security vulnerabilities in the design. Bonus points were earned when teams demonstrated clever use of automation in vulnerability detection and/or exploitation. Teams with the highest scores win.

                     DAC Award Ceremony on July 13, 2023.

Not Your Typical Capture-the-Flag (CTF)

To deliver an immersive, hands-on pre-silicon hacking experience, we prepared a sophisticated “buggy SoC” that incorporated industry-scale security features along with common security vulnerabilities that were inspired by real-world product issues. In addition, we enabled each team with a set of powerful commercial-grade EDA tools in the cloud, while providing options for teams to deploy their own custom tools. Participants in all SoC hacking experience levels were covered.

Behind the Scenes

The HACK@DAC organizing team is made up of industry and academia experts. A big shout-out goes to the team for their flawless execution and outstanding collaboration, working tirelessly since the beginning of the year.

HACK@DAC 2023 Winners

1st: Team “Sycuricon”, Zhejiang University, China

  • Advisor: Yajin Zhoud
  • Members: Jinyan Xu, Yiyuan Liu, Xiaodi Zhao, Jiaxun Zhu

2nd: Team “Calgary ISH”, University of Calgary, Canada

3rd: Team “Bitwise Bandits”, University of Florida, United States and Indian Institute of Technology, Kharagpur, India

3rd: Team “NYU_bounty_hunters”, UNSW, Australia and New York University, United States

Congratulations to the winners of the 6th HACK@DAC Hardware Security Capture-the-Flag (CTF) Competition!

IEEE/ACM A. Richard Newton Technical Impact Award in Electronic Design Automation

The IEEE/ACM A. Richard Newton Technical Impact Award in Electronic Design Automation was conferred at DAC 2023 upon Moshe Vardi and Pierre Wolper for their research work “An Automata-Theoretic Approach to Automatic Program Verification”, published in the proceedings of the 1st Symposium on Logic in Computer Science,1986.

Highlights of CADAthlon Brazil 2023

The CADAthlon Brazil 2023 – 3rd Brazilian Programming Contest for Design Automation of Integrated Circuits (https://csbc.sbc.org.br/2023/cadathlon-brasil-en/) took place on August 8th in João Pessoa, Paraíba State, Brazil, as a co-located event of the 43rd Annual Congress of SBC (Brazilian Computer Society). It was organized by Federal University of Pelotas (UFPel), Sul-Rio-Grandense Federal Institute of Education, Science and Technology (IFSul), Federal Institute of Paraiba (IFPB), Federal University of Santa Catarina (UFSC) and Federal University of Paraiba (UFPB) and sponsored by ACM/SIGDA, IEEE CEDA (Council on Electronic Design Automation), SBC/CECCI (SBC Special Committee on Integrated Circuits and Systems Design) and SBMicro
(Brazilian Microelectronics Society). It was financially sponsored by Synopsys, Chipus Microelectronics, EnSilica, HCLTech, ACM/SIGDA, IEEE CEDA, IEEE Circuits and Systems Society (CASS) and SBC/CECCI.

As in the first edition, CADAthlon Brazil 2023 followed the same format as the ACM/SIGDA CADAthlon, which happens annually co-located with ICCAD (International Conference on Computer-Aided Design). During the whole day, 10 two-person teams of students coming from different regions of Brazil worked to solve 6 practical problems on classical EDA topics such as circuit design & analysis, physical design, logic synthesis, high-level synthesis, circuit verification, and application of AI to design automation. The problems were prepared by a team of researchers from industry and academia.

This year the first place was won by team “Flamengo”, from University of Brasília (UnB), formed by Enzo Yoshio Niho and Eduardo Quirino de Oliveira, and the second place was won by team “Rabisco UFSC”, from the Federal University of Santa Catarina (UFSC), formed by Arthur Joao Lourenço and Bernardo Borges Sandoval. The top 2 teams were invited to participate in CADAthlon@ICCAD – SIGDA, a competition that runs in conjunction with ICCAD (International Conference on Computer-Aided Design) and will be held in SanFrancisco/CA (USA), from October 29 to November 2, 2023.

The CADAthlon Brazil 2023 Organizing Committee greatly thank the Congress of SBC organizers for the logistics support, the problem preparation team and all sponsors, specially the financial support from Synopsys, Chipus Microelectronics, EnSilica, HCLTech, ACM/SIGDA, IEEE Circuits and Systems Society (CASS), SBC/CECCI (SBC Special Committee on Integrated Circuits and Systems Design) and IEEE CEDA (through the South Brazil Chapter), which made it possible to cover the travel expenses of the competitors, making the event a huge success. The next edition of CADAthlon Brazil will occur as a co-located event of the 44th Annual Congress of SBC, in July 2024, in Brasilia/DF, capital of Brazil.