Bibliography

1
P. M. Baggenstoss, “Discriminative alignment of projected belief networks,” IEEE Signal Processing Letters, Sep 2021.

2
P. M. Baggenstoss, “Trainable compound activation functions for machine learning,” Submitted to EUSIPCO 2022, 2022.

3
P. M. Baggenstoss, “Maximum entropy PDF design using feature density constraints: Applications in signal processing,” IEEE Trans. Signal Processing, vol. 63, June 2015.

4
P. M. Baggenstoss, “A modified Baum-Welch algorithm for hidden Markov models with multiple observation spaces.,” IEEE Trans. Speech and Audio, pp. 411-416, May 2001.

5
P. M. Baggenstoss, “A theoretically optimum approach to classification using class-specific features.,” Proceedings of ICPR, Barcelona, 2000.

6
P. M. Baggenstoss, “The PDF projection theorem and the class-specific method,” IEEE Trans Signal Processing, pp. 672-685, March 2003.

7
N. Mitchell, M. Aanjaneya, R. Setaluri, and E. Sifakis, “Non-manifold level sets: A multivalued implicit surface representation with applications to self-collision processing,” ACM Transactions on Graphics, vol. 36, pp. 1-9, Oct. 2015.

8
J. M. Lee, Introduction to Smooth Manifolds.
Springer, 2002.

9
P. M. Baggenstoss, “Class-specific features in classification.,” IEEE Trans Signal Processing, pp. 3428-3432, December 1999.

10
S. Kay, “Sufficiency, classification, and the class-specific feature theorem,” IEEE Trans. Information Theory, vol. 46, pp. 1654-1658, July 2000.

11
T. Minka, P. Torr, and A. Zisserman, “Extending pictorial structures for object recognition,” BMVC 2004, September 2004.

12
T. Beierholm and P. M. Baggenstoss, “Speech music discrimination using class-specific features,” in Proc. ICPR 2004, 2004.

13
S. ichi Amari, “Integration of stochastic models by minimizing α-divergence,” Neural Computation, vol. 19, p. 2780–2796, 2007.

14
B. Tang, H. He, P. Baggenstoss, and S. Kay, “A bayesian classification approach using class-specific features for text categorization,” IEEE Transactions on Transactions on Knowledge and Data Engineering (accepted), 2016.

15
H. W. Sorensen, Parameter Estimation, Principles ans Problems.
New York: Marcel Dekker, 1980.

16
S. M. Kay, A. H. Nuttall, and P. M. Baggenstoss, “Multidimensional probability density function approximations for detection, classification, and model order selection,” IEEE Transactions on Signal Processing, vol. 49, pp. 2240-2252, Oct 2001.

17
D. R. Cox and D. V. Hinkley, Theoretical Statistics.
London: Chapman and Hall, 1974.

18
R. L. Strawderman, “Higher-order asymptotic approximation: Laplace, saddlepoint, and related methods,” Journal of the American Statistical Association, vol. 95, pp. 1358-1364, December 2000.

19
J. Durbin, “Approximations for densities of sufficient estimators,” Biometrika, vol. 67, no. 2, pp. 311-333, 1980.

20
P. M. Baggenstoss, “Optimal detection and classification of diverse short-duration signals,” in Proceedings of the International Conference on Cloud Engineering, (Boston, MA), pp. 534-539, 2014.

21
P. M. Baggenstoss and S. Kay, “Nonlinear dimension reduction by pdf estimation,” (Accepted in) IEEE Transactions on Signal Processing, 2022.

22
J.-P. Nadal and N. Parga, “Nonlinear neurons in the low-noise limit: a factorial code maximizes information transfer,” Network: Computation in Neural Systems, vol. 5, no. 4, pp. 565-581, 1994.

23
E. T. Jaynes, “On the rationale of maximum-entropy methods,” Proceedings of IEEE, vol. 70, no. 9, pp. 939-952, 1982.

24
P. M. Baggenstoss, “Uniform manifold sampling (UMS): Sampling the maximum entropy pdf,” IEEE Transactions on Signal Processing, vol. 65, pp. 2455-2470, May 2017.

25
J. W. Picone, “Signal modeling techniques in speech recognition,” Proceedings of the IEEE, vol. 81, no. 9, pp. 1215-1247, 1993.

26
L. Devroye, Non-Uniform Random Variate Generation.
Springer, 1986.

27
P. M. Baggenstoss, “On the duality between belief networks and feed-forward neural networks,” IEEE Transactions on Neural Networks and Learning Systems, pp. 1-11, 2018.

28
P. M. Baggenstoss, “A neural network based on first principles,” in ICASSP 2020, Barcelona (virtual), (Barcelona, Spain), Sep 2020.

29
P. M. Baggenstoss, “Applications of projected belief networks (PBN),” Proceedings of EUSIPCO, A Corunã, Spain, 2019.

30
S. Kay, Fundamentals of Statisticsl Signal Processing, Estimation Theory.
Prentice Hall, Upper Saddle River, New Jersey, USA, 1993.

31
S. Kay, Modern Spectral Estimation: Theory and Applications.
Prentice Hall, 1988.

32
P. Mermelstein, “Distance measures for speech recognition, psychological and instrumental,” Pattern Recognition and Artificial Intelligence, p. 374–388, 1976.

33
S. Young, G. Evermann, M. Gales, T. Hain, D. Kershaw, X. A. Liu, G. Moore, J. Odell, D. Ollason, D. Povey, V. Valtchev, and P. Woodland, The HTK Book, Version 3.4.
Cambridge University Engineering Department, 2006.

34
S. Geman and D. Geman, “Stochastic relaxation, Gibbs distributions, and the Bayesian restoration of images,” IEEE Transactions on Pattern Analysis and Machine Intelligence, no. 6, p. 721­741, 1984.

35
S. Kiatsupaibul, R. Smith, and Z. Zabinsky, “An analysis of a variation of hit-and-run for uniform sampling from general regions,” ACM Transactions on Modeling and Computer Simulation (TOMACS), vol. 21, no. 3, 2011.

36
R. L. Smith, “Efficient Monte Carlo procedures for generating points uniformly distributed over bounded regions,” Operations Research, vol. 32, pp. 1296-1308, 1984.

37
R. L. Smith, “The hit-and-run sampler: a globally reaching Markov chain sampler for generating arbitrary multivariate distributions,” Proceedings of the 1996 Winter Simulation Conference, 1996.

38
R. M. Neal, “Slice sampling,” The Annals of Statistics, vol. 31, no. 3, pp. 705-767, 2003.

39
J. N. Kapur, Maximum Entropy Models in Science and Engineering.
Wiley (Eastern), 1993.

40
N. A. Malik, “One and two dimensional maximum entropy spectral estimation,” MIT PhD Thesis, Nov 1981.

41
J. P. Burg, “The relationship between maximum entropy and maximum likelihood spectra,” Geophysics, vol. 37, no. 2, pp. 375-376, 1971.

42
S. J. Wernecke and L. R. D'Addario, “Maximum entropy image reconstruction,” IEEE Trans. Computers, vol. C-26, no. 4, pp. 351-364, 1977.

43
G. Wei and H. Zhen-Ya, “A new algorithm for maximum entropy image reconstruction,” in Proceedings of ICASSP-87, vol. 12, pp. 595-597, April 1987.

44
B. D. Ripley, Stochastic Simulation.
John Wiley & Sons, 1987.

45
B. Milner and X. Shao, “Speech reconstruction from mel-frequency cepstral coefficients using a source-filter model.,” in INTERSPEECH, Citeseer, 2002.

46
P. M. Baggenstoss, “Maximum entropy auto-encoding,” 2021.

47
P. M. Baggenstoss, “A neural network based on first principles,” 2020.

48
V. P. Singh, Entropy, Theory and its Applications in Environmental and Water Engineering.
John Wiley & Sons, 2013.

49
K. Conrad, “Probability distributions and maximum entropy,” Unpublished article, 2013.

50
A. H. Nuttall and P. M. Baggenstoss, “The joint distributions for two useful classes of statistics with applications to classification and hypothesis testing,” Submitted to IEEE Trans. Signal Processing, 2002.

51
A. H. Nuttall, “Joint probability density function of selected order statistics and the sum of the remaining random variables,” NUWC Technical Report 11345, October 2001.

52
A. H. Nuttall, “Joint probability density function of selected order statistics and the sum of the remainder as applied to arbitrary independent random variables,” NUWC Technical Report 11469, November 2003.

53
A. H. Nuttall, “Saddlepoint approximation for the combined probability and joint probability density function of selected order statistics and the sum of the remainder,” NUWC Technical Report 11XXX, February 2004.

54
A. H. Nuttall, “Detection performance of generalized likelihood ratio processors for random signals of unknown location, structure, extent, and strength,” NUWC Technical Report 10739, August 1994.

55
O. E. Barndorff-Nielsen and D. R. Cox, Asymptotic Techniques for Use in Statistics.
Chapman and Hall, 1989.

56
A. H. Nuttall, “Saddlepoint approximation and first-order correction term to the joint probability density function of M quadratic and linear forms in K Gaussian random variables with arbitrary means and covariances,” NUWC Technical Report 11262, December 2000.

57
P. M. Baggenstoss, “On the equivalence of hanning-weighted and overlapped analysis windows using different window sizes,” IEEE Signal Processing Letters, vol. 19, pp. 27-30, Jan 2012.

58
P. Baggenstoss, “Time-series segmentation,” United States Patent 6907367, June 2005.

59
E. Parzen, “On estimation of a probability density function and mode,” Annals of Mathematical Statistics, vol. 33, pp. 1065-1076, 1962.

60
D. M. Titterington, A. F. M. Smith, and U. E. Makov, Statistical Analysis Of Finite Mixture Distributions.
John Wiley & Sons, 1985.

61
R. A. Redner and H. F. Walker, “Mixture densities maximum likelihood, and the EM algorithm,” SIAM Review, vol. 26, April 1984.

62
N. Vlassis and A. Likas, “The kurtosis-EM algorithm for Gaussian mixture modelling,” IEEE Trans. SMC (submitted), 1999.

63
Anderson and Moore, Optimal Filtering.
PH, 1979.

64
M. Kendall and A. Stuart, The Advanced Theory of Statistics, Vol. 2.
London: Charles Griffin, 1979.

65
L. R. Rabiner, “A tutorial on hidden Markov models and selected applications in speech recognition,” Proceedings of the IEEE, vol. 77, pp. 257-286, February 1989.

66
B. H. Juang, “Maximum likelihood estimation for mixture multivariate stochastic observations of Markov chains,” AT&T Technical Journal, vol. 64, no. 6, pp. 1235-1249, 1985.

67
T. Amr, “Survey on time-series data classification,” in Submitted to: Time Series and Data Mining 2012 (unpublished), 2012.

68
E. Keogh, “UCR time series classification/clustering page,” 2015.

69
C. A. Ratanamahatana and E. Keogh, “Making time-series classification more accurate using learned constraints,” in SIAM International Conference on Data Mining (SDM2004), pp. 11-22, 2004.

70
Y.-S. Yun and Y.-H. Oh, “A segmental-feature hmm for continuous speech recognition based on a parametric trajectory model,” Speech Communication, vol. 38, pp. 115-130, September 2002.

71
P. M. Baggenstoss, “Class-specific features in classification.,” in IASTED International Conference on Signal and Image Processing, 1998.

72
P. M. Baggenstoss, “A multi-resolution hidden markov model using class-specific features,” Proceedings of EUSIPCO 2008, Lausanne, Switzerland, Aug 2008.

73
P. M. Baggenstoss, “A multi-resolution hidden markov model using class-specific features,” IEEE Transactions on Signal Processing, vol. 58, pp. 5165-5177, Oct 2010.

74
B. F. Harrison and P. M. Baggenstoss, “A multi-resolution hidden markov model for optimal detection, tracking, separation and classification of marine mammal vocalizations,” in Proc. Oceans 2008, Quebec City, September 2008.

75
M. Ostendorf, V. Digalakis, and O. A. Kimball, “From hmms to segment models: A unifed view of stochastic modeling for speech recognition,” IEEE Transactions on Speech and Audio Processing, 1996.

76
J. A. Bilmes, “Graphical models and automatic speech recognition,” Mathematical Foundations of Speech and Language Processing, 2003.

77
S. Furui, “Speaker-independent isolated word recognition using dynamic features of speech spectrum,” IEEE Trans on ASSP, vol. 34, no. 1, pp. 52-59, 1986.

78
K. Achan, S. Roweis, and B. Frey, “A segmental hmm for speech waveforms,” University of Toronto Technical Report, Jan 2004.

79
D. Gamerman, Markov Chain Monte Carlo.
Chapman and Hall, 1997.

80
J. S. Garofolo, “Timit acoustic-phonetic continuous speech corpus,” Linguistic Data Consortium, 1993.

81
P. Baggenstoss, “Office sounds database,” http://class-specific.com/os.
Accessed: 2022-02-28.

82
P. M. Baggenstoss, “The projected belief network classifier: both generative and discriminative,” Proceedings of EUSIPCO, Amsterdam, 2020.

83
M. Rosenblatt, “Remarks on a multivariate transformation,” The Annals of Mathematical Statistics, vol. 23, no. 3, pp. 470-472, 1952.

84
G. Deco and W. Brauer, “Nonlinear higher-order statistical decorrelation by volume-conserving neural architectures,” Neural Netw., vol. 8, p. 525–535, June 1995.

85
G. Deco and D. Obradovic, AN INFORMATION-THEORETIC APPROACH TO NEURAL COMPUTING.
Springer, 1996.

86
A. Hyvärinen, J. Karhunen, and E. Oja, Independent Component Analysis.
Wiley, 2001.

87
A. Hyvärinen and P. Pajunen, “Nonlinear independent component analysis: Existence and uniqueness results,” Neural Networks, vol. 12, no. 3, pp. 429-439, 1999.

88
P. M. Baggenstoss, “Applications of projected belief networks (pbn),” in Proceedings of EUSIPCO 2019, (La Coruña, Spain), Sep 2019.

89
M. Welling, M. Rosen-Zvi, and G. Hinton, “Exponential family harmoniums with an application to information retrieval,” Advances in neural information processing systems, 2004.

90
G. E. Hinton, S. Osindero, and Y.-W. Teh, “A fast learning algorithm for deep belief nets,” in Neural Computation 2006, 2006.

91
D. J. Rezende, S. Mohamed, and D. Wierstra, “Stochastic backpropagation and approximate inference in deep generative models,” in Proceedings of the 31st International Conference on Machine Learning (E. P. Xing and T. Jebara, eds.), vol. 32 of Proceedings of Machine Learning Research, (Bejing, China), pp. 1278-1286, PMLR, 22-24 Jun 2014.

92
I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial nets,” in Advances in Neural Information Processing Systems 27 (Z. Ghahramani, M. Welling, C. Cortes, N. D. Lawrence, and K. Q. Weinberger, eds.), pp. 2672-2680, Curran Associates, Inc., 2014.

93
V. Vapnik, The Nature of Statistical Learning.
Springer, 1999.

94
I. Goodfellow, Y. Bengio, A. Courville, and Y. Bengio, Deep Learning.
Cambridge, MA: MIT press, 2016.

95
H. L. Royden, Real Analysis.
Englewood Cliffs, New Jersey, USA: Prentice Hall, third ed., 1988.

96
T. Minka, “Exemplar-based likelihoods using the pdf projection theorem.,” Microsoft Research Ltd, technical report, March 2004.

97
D. J. Olive, “Applied robust statistics,” Unpublished online text, Chapter 4, http://lagrange.math.siu.edu/Olive/ol-bookp.htm, 2008.

98
D. J. Olive, Robust Multivariate Analysis (to appear).
New York: Springer, 2017.