FIST: A Feature-Importance Sampling and Tree-Based Method for Automatic Design Flow Parameter Tuning

Design flow parameters are of utmost importance to chip design quality and require a painfully long time to evaluate their effects. In reality, flow parameter tuning is usually performed manually based on designers’ experience in an ad hoc manner. In this work, we introduce a machine learning-based automatic parameter tuning methodology that aims to find the best design quality with a limited number of trials. Instead of merely plugging in machine learning engines, we develop clustering and approximate sampling techniques for improving tuning efficiency. The feature extraction in this method can reuse knowledge from prior designs. Furthermore, we leverage a state-of-the-art XGBoost model and propose a novel dynamic tree technique to overcome overfitting. Experimental results on benchmark circuits show that our approach achieves 25% improvement in design quality or 37% reduction in sampling cost compared to random forest method, which is the kernel of a highly cited previous work. Our approach is further validated on two industrial designs. By sampling less than 0.02% of possible parameter sets, it reduces area by 1.83% and 1.43% compared to the best solutions hand-tuned by experienced designers.

[1]  Luca P. Carloni,et al.  A synthesis-parameter tuning system for autonomous design-space exploration , 2016, 2016 Design, Automation & Test in Europe Conference & Exhibition (DATE).

[2]  James C. Hoe,et al.  Nautilus: Fast automated IP design space search using guided genetic algorithms , 2015, 2015 52nd ACM/EDAC/IEEE Design Automation Conference (DAC).

[3]  Luca P. Carloni,et al.  Scalable Auto-Tuning of Synthesis Parameters for Optimizing High-Performance Processors , 2016, ISLPED.

[4]  Hung-Yi Liu,et al.  Supervised Design-Space Exploration , 2015 .

[5]  Dorothea Heiss-Czedik,et al.  An Introduction to Genetic Algorithms. , 1997, Artificial Life.

[6]  Vittorio Zaccaria,et al.  OSCAR: An Optimization Methodology Exploiting Spatial Correlation in Multicore Design Spaces , 2012, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems.

[7]  Dong-Hyun Lee,et al.  Pseudo-Label : The Simple and Efficient Semi-Supervised Learning Method for Deep Neural Networks , 2013 .

[8]  Pingfan Meng,et al.  Adaptive Threshold Non-Pareto Elimination: Re-thinking machine learning for system level design space exploration on FPGAs , 2016, 2016 Design, Automation & Test in Europe Conference & Exhibition (DATE).

[9]  J. Friedman Greedy function approximation: A gradient boosting machine. , 2001 .

[10]  Vittorio Zaccaria,et al.  SPIRIT: Spectral-Aware Pareto Iterative Refinement Optimization for Supervised High-Level Synthesis , 2015, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems.

[11]  Vittorio Zaccaria,et al.  A meta-model assisted coprocessor synthesis framework for compiler/architecture parameters customization , 2013, 2013 Design, Automation & Test in Europe Conference & Exhibition (DATE).

[12]  Andreas Krause,et al.  Active Learning for Multi-Objective Optimization , 2013, ICML.

[13]  Tianqi Chen,et al.  XGBoost: A Scalable Tree Boosting System , 2016, KDD.

[14]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[15]  Luca P. Carloni,et al.  On learning-based methods for design-space exploration with High-Level Synthesis , 2013, 2013 50th ACM/EDAC/IEEE Design Automation Conference (DAC).

[16]  Andreas Krause,et al.  "Smart" design space sampling to predict Pareto-optimal solutions , 2012, LCTES '12.

[17]  William J. Dally,et al.  A 0.11 pJ/Op, 0.32-128 TOPS, Scalable Multi-Chip-Module-based Deep Neural Network Accelerator with Ground-Reference Signaling in 16nm , 2019, 2019 Symposium on VLSI Circuits.

[18]  Luca P. Carloni,et al.  A Learning-Based Recommender System for Autotuning Design FIows of Industrial High-Performance Processors , 2019, 2019 56th ACM/IEEE Design Automation Conference (DAC).