暂无分享,去创建一个
Yain-Whar Si | Cong Lin | Xueyuan Gong | Yongqi Tian | Xinyuan Zhang | Xiaoxiang Liu | Xinyuan Zhang | Yain-Whar Si | Cong Lin | Xiaoxiang Liu | Yongqi Tian | Xueyuan Gong
[1] Rohit J. Kate. Using dynamic time warping distances as features for improved time series classification , 2016, Data Mining and Knowledge Discovery.
[2] Seyed Iman Mirzadeh,et al. Improved Knowledge Distillation via Teacher Assistant , 2020, AAAI.
[3] Dumitru Erhan,et al. Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[4] Saeed Karimi-Bidhendi,et al. Scalable Classification of Univariate and Multivariate Time Series , 2018, 2018 IEEE International Conference on Big Data (Big Data).
[5] Wei Chen,et al. Multi-scale Attention Convolutional Neural Network for time series classification , 2021, Neural Networks.
[6] Simon Fong,et al. Fast multi-subsequence monitoring on streaming time-series based on Forward-propagation , 2018, Inf. Sci..
[7] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[8] Francesca Mangili,et al. Should We Really Use Post-Hoc Tests Based on Mean-Ranks? , 2015, J. Mach. Learn. Res..
[9] Germain Forestier,et al. Deep learning for time series classification: a review , 2018, Data Mining and Knowledge Discovery.
[10] Junmo Kim,et al. A Gift from Knowledge Distillation: Fast Optimization, Network Minimization and Transfer Learning , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[11] M. Friedman. A Comparison of Alternative Tests of Significance for the Problem of $m$ Rankings , 1940 .
[12] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[13] Geoffrey I. Webb,et al. InceptionTime: Finding AlexNet for time series classification , 2019, Data Mining and Knowledge Discovery.
[14] Tim Oates,et al. Time series classification from scratch with deep neural networks: A strong baseline , 2016, 2017 International Joint Conference on Neural Networks (IJCNN).
[15] Sergey Alyamkin,et al. MarginDistillation: distillation for margin-based softmax , 2020, ArXiv.
[16] Yixin Chen,et al. Multi-Scale Convolutional Neural Networks for Time Series Classification , 2016, ArXiv.
[17] Eamonn J. Keogh,et al. The great time series classification bake off: a review and experimental evaluation of recent algorithmic advances , 2016, Data Mining and Knowledge Discovery.
[18] Enhua Wu,et al. Squeeze-and-Excitation Networks , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[19] George C. Runger,et al. A Bag-of-Features Framework to Classify Time Series , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[20] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[21] Yuqing Wan,et al. A formal approach to chart patterns classification in financial time series , 2017, Inf. Sci..
[22] Jang Hyun Cho,et al. On the Efficacy of Knowledge Distillation , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[23] Yoshua Bengio,et al. FitNets: Hints for Thin Deep Nets , 2014, ICLR.
[24] Eamonn J. Keogh,et al. Searching and Mining Trillions of Time Series Subsequences under Dynamic Time Warping , 2012, KDD.
[25] Sergey Ioffe,et al. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning , 2016, AAAI.
[26] S. García,et al. An Extension on "Statistical Comparisons of Classifiers over Multiple Data Sets" for all Pairwise Comparisons , 2008 .
[27] Janez Demsar,et al. Statistical Comparisons of Classifiers over Multiple Data Sets , 2006, J. Mach. Learn. Res..
[28] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[29] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[30] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[31] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[32] Christos Faloutsos,et al. Stream Monitoring under the Time Warping Distance , 2007, 2007 IEEE 23rd International Conference on Data Engineering.
[33] Sergey Ioffe,et al. Rethinking the Inception Architecture for Computer Vision , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[34] Takio Kurita,et al. Triplet Loss for Knowledge Distillation , 2020, 2020 International Joint Conference on Neural Networks (IJCNN).
[35] Jason Lines,et al. Time-Series Classification with COTE: The Collective of Transformation-Based Ensembles , 2015, IEEE Transactions on Knowledge and Data Engineering.