Sparse Feature Extraction Model with Independent Subspace Analysis
暂无分享,去创建一个
[1] Quoc V. Le,et al. Learning hierarchical invariant spatio-temporal features for action recognition with independent subspace analysis , 2011, CVPR 2011.
[2] Aapo Hyvärinen,et al. Topographic Independent Component Analysis , 2001, Neural Computation.
[3] A. Hyvärinen,et al. Complex cell pooling and the statistics of natural images , 2007, Network.
[4] John D. Lafferty,et al. Learning image representations from the pixel level via hierarchical sparse coding , 2011, CVPR 2011.
[5] David J. Field,et al. Sparse coding with an overcomplete basis set: A strategy employed by V1? , 1997, Vision Research.
[6] Thomas Serre,et al. Realistic Modeling of Simple and Complex Cell Tuning in the HMAX Model, and Implications for Invariant Object Recognition in Cortex , 2004 .
[7] Thomas Serre,et al. Robust Object Recognition with Cortex-Like Mechanisms , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[8] Graham W. Taylor,et al. Adaptive deconvolutional networks for mid and high level feature learning , 2011, 2011 International Conference on Computer Vision.
[9] Pietro Perona,et al. Learning Generative Visual Models from Few Training Examples: An Incremental Bayesian Approach Tested on 101 Object Categories , 2004, 2004 Conference on Computer Vision and Pattern Recognition Workshop.
[10] Xiaolin Hu,et al. Sparsity-Regularized HMAX for Visual Recognition , 2014, PloS one.
[11] Edmund T. Rolls,et al. Invariant Visual Object and Face Recognition: Neural and Computational Bases, and a Model, VisNet , 2012, Front. Comput. Neurosci..
[12] Thomas Serre,et al. A feedforward architecture accounts for rapid categorization , 2007, Proceedings of the National Academy of Sciences.
[13] Matthieu Cord,et al. HMAX-S: Deep scale representation for biologically inspired image categorization , 2011, 2011 18th IEEE International Conference on Image Processing.
[14] Marc'Aurelio Ranzato,et al. Building high-level features using large scale unsupervised learning , 2011, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.
[15] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[16] Aapo Hyvärinen,et al. Emergence of Phase- and Shift-Invariant Features by Decomposition of Natural Images into Independent Feature Subspaces , 2000, Neural Computation.
[17] Thomas Serre,et al. Object recognition with features inspired by visual cortex , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).
[18] L. Abbott,et al. Responses of neurons in primary and inferior temporal visual cortices to natural scenes , 1997, Proceedings of the Royal Society of London. Series B: Biological Sciences.
[19] David G. Lowe,et al. University of British Columbia. , 1945, Canadian Medical Association journal.
[20] Edmund T. Rolls,et al. The neuronal encoding of information in the brain , 2011, Progress in Neurobiology.
[21] Honglak Lee,et al. Convolutional deep belief networks for scalable unsupervised learning of hierarchical representations , 2009, ICML '09.
[22] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[23] Thomas Serre,et al. Hierarchical Models of the Visual System , 2014, Encyclopedia of Computational Neuroscience.
[24] Aapo Hyvärinen,et al. Natural Image Statistics - A Probabilistic Approach to Early Computational Vision , 2009, Computational Imaging and Vision.
[25] Matthieu Cord,et al. Extended Coding and Pooling in the HMAX Model , 2013, IEEE Transactions on Image Processing.
[26] Jiaxing Zhang,et al. Scale-Invariant Convolutional Neural Networks , 2014, ArXiv.
[27] T. Poggio,et al. Hierarchical models of object recognition in cortex , 1999, Nature Neuroscience.
[28] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..