Opportunities for RTL and Gate Level Simulation using GPUs (Invited Talk)
暂无分享,去创建一个
[1] Valeria Bertacco,et al. Gate-Level Simulation with GPU Computing , 2011, TODE.
[2] Yangdong Deng,et al. Massively Parallel Logic Simulation with GPUs , 2011, TODE.
[3] David Z. Pan,et al. DREAMPIace: Deep Learning Toolkit-Enabled GPU Acceleration for Modern VLSI Placement , 2019, 2019 56th ACM/IEEE Design Automation Conference (DAC).
[4] Hans-Joachim Wunderlich,et al. High-Throughput Logic Timing Simulation on GPGPUs , 2015, TODE.
[5] Max Welling,et al. Semi-Supervised Classification with Graph Convolutional Networks , 2016, ICLR.
[6] Yangdong Deng,et al. Accelerating RTL simulation with GPUs , 2011, 2011 IEEE/ACM International Conference on Computer-Aided Design (ICCAD).
[7] Kurt Keutzer,et al. Parallelizing CAD: A timely research agenda for EDA , 2008, 2008 45th ACM/IEEE Design Automation Conference.
[8] Alper Sen,et al. Parallel Cycle Based Logic Simulation Using Graphics Processing Units , 2010, 2010 Ninth International Symposium on Parallel and Distributed Computing.
[9] Franco Fummi,et al. SAGA: SystemC acceleration on GPU architectures , 2012, DAC Design Automation Conference 2012.
[10] Yanqing Zhang,et al. Problem C: GPU Accelerated Logic Re-simulation : (Invited Talk) , 2020, 2020 IEEE/ACM International Conference On Computer Aided Design (ICCAD).
[11] Jiaqi Gu,et al. DREAMPlace: Deep Learning Toolkit-Enabled GPU Acceleration for Modern VLSI Placement , 2020 .