Non-Gaussian reflectivity, entropy, and deconvolution
暂无分享,去创建一个
[1] C. E. SHANNON,et al. A mathematical theory of communication , 1948, MOCO.
[2] Solomon Kullback,et al. Information Theory and Statistics , 1960 .
[3] Thomas S. Ferguson,et al. On the Rejection of Outliers , 1961 .
[4] E. Parzen. On Estimation of a Probability Density Function and Mode , 1962 .
[5] Jan Havrda,et al. Quantification method of classification processes. Concept of structural a-entropy , 1967, Kybernetika.
[6] C. L. Mallows,et al. Linear processes are nearly Gaussian , 1967 .
[7] J. Gower,et al. Multivariate data analysis , 1972 .
[8] R. Hogg. More Light on the Kurtosis and Related Statistics , 1972 .
[9] M. Kendall. Entropy, Probability and Information , 1973 .
[10] A. Hobson,et al. A comparison of the Shannon and Kullback information measures , 1973 .
[11] Mukhtar M. Ali. Stochastic Ordering and Kurtosis Measure , 1974 .
[12] Stephen J. Wernecke,et al. Maximum Entropy Image Reconstruction , 1977, IEEE Transactions on Computers.
[13] P. J. Huber. Robust Statistical Procedures , 1977 .
[14] R. Wiggins. Minimum entropy deconvolution , 1978 .
[15] T. Ulrych,et al. MINIMUM ENTROPY DECONVOLUTION WITH AN EXPONENTIAL TRANSFORMATION , 1979 .
[16] J. Skilling,et al. Deconvolution by maximum entropy, as illustrated by application to the jet of M87 , 1980 .
[17] D. Donoho. ON MINIMUM ENTROPY DECONVOLUTION , 1981 .
[18] Pedro A. Morettin,et al. The Levinson Algorithm and its Applications in Time Series Analysis , 1984 .