- 1
-
P. M. Baggenstoss, “Discriminative alignment of projected belief networks,”
IEEE Signal Processing Letters, Sep 2021.
- 2
-
P. M. Baggenstoss, “Trainable compound activation functions for machine
learning,” Submitted to EUSIPCO 2022, 2022.
- 3
-
P. M. Baggenstoss, “Maximum entropy PDF design using feature density
constraints: Applications in signal processing,” IEEE Trans. Signal
Processing, vol. 63, June 2015.
- 4
-
P. M. Baggenstoss, “A modified Baum-Welch algorithm for hidden Markov
models with multiple observation spaces.,” IEEE Trans. Speech and
Audio, pp. 411-416, May 2001.
- 5
-
P. M. Baggenstoss, “A theoretically optimum approach to classification using
class-specific features.,” Proceedings of ICPR, Barcelona, 2000.
- 6
-
P. M. Baggenstoss, “The PDF projection theorem and the class-specific
method,” IEEE Trans Signal Processing, pp. 672-685, March 2003.
- 7
-
N. Mitchell, M. Aanjaneya, R. Setaluri, and E. Sifakis, “Non-manifold level
sets: A multivalued implicit surface representation with applications to
self-collision processing,” ACM Transactions on Graphics, vol. 36,
pp. 1-9, Oct. 2015.
- 8
-
J. M. Lee, Introduction to Smooth Manifolds.
Springer, 2002.
- 9
-
P. M. Baggenstoss, “Class-specific features in classification.,” IEEE
Trans Signal Processing, pp. 3428-3432, December 1999.
- 10
-
S. Kay, “Sufficiency, classification, and the class-specific feature
theorem,” IEEE Trans. Information Theory, vol. 46, pp. 1654-1658,
July 2000.
- 11
-
T. Minka, P. Torr, and A. Zisserman, “Extending pictorial structures for
object recognition,” BMVC 2004, September 2004.
- 12
-
T. Beierholm and P. M. Baggenstoss, “Speech music discrimination using
class-specific features,” in Proc. ICPR 2004, 2004.
- 13
-
S. ichi Amari, “Integration of stochastic models by minimizing
α-divergence,” Neural Computation, vol. 19, p. 2780–2796, 2007.
- 14
-
B. Tang, H. He, P. Baggenstoss, and S. Kay, “A bayesian classification
approach using class-specific features for text categorization,” IEEE
Transactions on Transactions on Knowledge and Data Engineering (accepted),
2016.
- 15
-
H. W. Sorensen, Parameter Estimation, Principles ans Problems.
New York: Marcel Dekker, 1980.
- 16
-
S. M. Kay, A. H. Nuttall, and P. M. Baggenstoss, “Multidimensional probability
density function approximations for detection, classification, and model
order selection,” IEEE Transactions on Signal Processing, vol. 49,
pp. 2240-2252, Oct 2001.
- 17
-
D. R. Cox and D. V. Hinkley, Theoretical Statistics.
London: Chapman and Hall, 1974.
- 18
-
R. L. Strawderman, “Higher-order asymptotic approximation: Laplace,
saddlepoint, and related methods,” Journal of the American Statistical
Association, vol. 95, pp. 1358-1364, December 2000.
- 19
-
J. Durbin, “Approximations for densities of sufficient estimators,” Biometrika, vol. 67, no. 2, pp. 311-333, 1980.
- 20
-
P. M. Baggenstoss, “Optimal detection and classification of diverse
short-duration signals,” in Proceedings of the International Conference
on Cloud Engineering, (Boston, MA), pp. 534-539, 2014.
- 21
-
P. M. Baggenstoss and S. Kay, “Nonlinear dimension reduction by pdf
estimation,” (Accepted in) IEEE Transactions on Signal Processing,
2022.
- 22
-
J.-P. Nadal and N. Parga, “Nonlinear neurons in the low-noise limit: a
factorial code maximizes information transfer,” Network: Computation in
Neural Systems, vol. 5, no. 4, pp. 565-581, 1994.
- 23
-
E. T. Jaynes, “On the rationale of maximum-entropy methods,” Proceedings
of IEEE, vol. 70, no. 9, pp. 939-952, 1982.
- 24
-
P. M. Baggenstoss, “Uniform manifold sampling (UMS): Sampling the maximum
entropy pdf,” IEEE Transactions on Signal Processing, vol. 65,
pp. 2455-2470, May 2017.
- 25
-
J. W. Picone, “Signal modeling techniques in speech recognition,” Proceedings of the IEEE, vol. 81, no. 9, pp. 1215-1247, 1993.
- 26
-
L. Devroye, Non-Uniform Random Variate Generation.
Springer, 1986.
- 27
-
P. M. Baggenstoss, “On the duality between belief networks and feed-forward
neural networks,” IEEE Transactions on Neural Networks and Learning
Systems, pp. 1-11, 2018.
- 28
-
P. M. Baggenstoss, “A neural network based on first principles,” in ICASSP 2020, Barcelona (virtual), (Barcelona, Spain), Sep 2020.
- 29
-
P. M. Baggenstoss, “Applications of projected belief networks (PBN),” Proceedings of EUSIPCO, A Corunã, Spain, 2019.
- 30
-
S. Kay, Fundamentals of Statisticsl Signal Processing, Estimation Theory.
Prentice Hall, Upper Saddle River, New Jersey, USA, 1993.
- 31
-
S. Kay, Modern Spectral Estimation: Theory and Applications.
Prentice Hall, 1988.
- 32
-
P. Mermelstein, “Distance measures for speech recognition, psychological and
instrumental,” Pattern Recognition and Artificial Intelligence,
p. 374–388, 1976.
- 33
-
S. Young, G. Evermann, M. Gales, T. Hain, D. Kershaw, X. A. Liu, G. Moore,
J. Odell, D. Ollason, D. Povey, V. Valtchev, and P. Woodland, The HTK
Book, Version 3.4.
Cambridge University Engineering Department, 2006.
- 34
-
S. Geman and D. Geman, “Stochastic relaxation, Gibbs distributions, and the
Bayesian restoration of images,” IEEE Transactions on Pattern
Analysis and Machine Intelligence, no. 6, p. 721741, 1984.
- 35
-
S. Kiatsupaibul, R. Smith, and Z. Zabinsky, “An analysis of a variation of
hit-and-run for uniform sampling from general regions,” ACM
Transactions on Modeling and Computer Simulation (TOMACS), vol. 21, no. 3,
2011.
- 36
-
R. L. Smith, “Efficient Monte Carlo procedures for generating points
uniformly distributed over bounded regions,” Operations Research,
vol. 32, pp. 1296-1308, 1984.
- 37
-
R. L. Smith, “The hit-and-run sampler: a globally reaching Markov chain
sampler for generating arbitrary multivariate distributions,” Proceedings of the 1996 Winter Simulation Conference, 1996.
- 38
-
R. M. Neal, “Slice sampling,” The Annals of Statistics, vol. 31, no. 3,
pp. 705-767, 2003.
- 39
-
J. N. Kapur, Maximum Entropy Models in Science and Engineering.
Wiley (Eastern), 1993.
- 40
-
N. A. Malik, “One and two dimensional maximum entropy spectral estimation,”
MIT PhD Thesis, Nov 1981.
- 41
-
J. P. Burg, “The relationship between maximum entropy and maximum likelihood
spectra,” Geophysics, vol. 37, no. 2, pp. 375-376, 1971.
- 42
-
S. J. Wernecke and L. R. D'Addario, “Maximum entropy image reconstruction,”
IEEE Trans. Computers, vol. C-26, no. 4, pp. 351-364, 1977.
- 43
-
G. Wei and H. Zhen-Ya, “A new algorithm for maximum entropy image
reconstruction,” in Proceedings of ICASSP-87, vol. 12, pp. 595-597,
April 1987.
- 44
-
B. D. Ripley, Stochastic Simulation.
John Wiley & Sons, 1987.
- 45
-
B. Milner and X. Shao, “Speech reconstruction from mel-frequency cepstral
coefficients using a source-filter model.,” in INTERSPEECH, Citeseer,
2002.
- 46
-
P. M. Baggenstoss, “Maximum entropy auto-encoding,” 2021.
- 47
-
P. M. Baggenstoss, “A neural network based on first principles,” 2020.
- 48
-
V. P. Singh, Entropy, Theory and its Applications in Environmental and
Water Engineering.
John Wiley & Sons, 2013.
- 49
-
K. Conrad, “Probability distributions and maximum entropy,” Unpublished
article, 2013.
- 50
-
A. H. Nuttall and P. M. Baggenstoss, “The joint distributions for two useful
classes of statistics with applications to classification and hypothesis
testing,” Submitted to IEEE Trans. Signal Processing, 2002.
- 51
-
A. H. Nuttall, “Joint probability density function of selected order
statistics and the sum of the remaining random variables,” NUWC
Technical Report 11345, October 2001.
- 52
-
A. H. Nuttall, “Joint probability density function of selected order
statistics and the sum of the remainder as applied to arbitrary independent
random variables,” NUWC Technical Report 11469, November 2003.
- 53
-
A. H. Nuttall, “Saddlepoint approximation for the combined probability and
joint probability density function of selected order statistics and the sum
of the remainder,” NUWC Technical Report 11XXX, February 2004.
- 54
-
A. H. Nuttall, “Detection performance of generalized likelihood ratio
processors for random signals of unknown location, structure, extent, and
strength,” NUWC Technical Report 10739, August 1994.
- 55
-
O. E. Barndorff-Nielsen and D. R. Cox, Asymptotic Techniques for Use in
Statistics.
Chapman and Hall, 1989.
- 56
-
A. H. Nuttall, “Saddlepoint approximation and first-order correction term to
the joint probability density function of M quadratic and linear forms in
K Gaussian random variables with arbitrary means and covariances,” NUWC Technical Report 11262, December 2000.
- 57
-
P. M. Baggenstoss, “On the equivalence of hanning-weighted and overlapped
analysis windows using different window sizes,” IEEE Signal Processing
Letters, vol. 19, pp. 27-30, Jan 2012.
- 58
-
P. Baggenstoss, “Time-series segmentation,” United States Patent
6907367, June 2005.
- 59
-
E. Parzen, “On estimation of a probability density function and mode,” Annals of Mathematical Statistics, vol. 33, pp. 1065-1076, 1962.
- 60
-
D. M. Titterington, A. F. M. Smith, and U. E. Makov, Statistical Analysis
Of Finite Mixture Distributions.
John Wiley & Sons, 1985.
- 61
-
R. A. Redner and H. F. Walker, “Mixture densities maximum likelihood, and the
EM algorithm,” SIAM Review, vol. 26, April 1984.
- 62
-
N. Vlassis and A. Likas, “The kurtosis-EM algorithm for Gaussian mixture
modelling,” IEEE Trans. SMC (submitted), 1999.
- 63
-
Anderson and Moore, Optimal Filtering.
PH, 1979.
- 64
-
M. Kendall and A. Stuart, The Advanced Theory of Statistics, Vol. 2.
London: Charles Griffin, 1979.
- 65
-
L. R. Rabiner, “A tutorial on hidden Markov models and selected applications
in speech recognition,” Proceedings of the IEEE, vol. 77,
pp. 257-286, February 1989.
- 66
-
B. H. Juang, “Maximum likelihood estimation for mixture multivariate
stochastic observations of Markov chains,” AT&T Technical Journal,
vol. 64, no. 6, pp. 1235-1249, 1985.
- 67
-
T. Amr, “Survey on time-series data classification,” in Submitted to:
Time Series and Data Mining 2012 (unpublished), 2012.
- 68
-
E. Keogh, “UCR time series classification/clustering page,” 2015.
- 69
-
C. A. Ratanamahatana and E. Keogh, “Making time-series classification more
accurate using learned constraints,” in SIAM International Conference
on Data Mining (SDM2004), pp. 11-22, 2004.
- 70
-
Y.-S. Yun and Y.-H. Oh, “A segmental-feature hmm for continuous speech
recognition based on a parametric trajectory model,” Speech
Communication, vol. 38, pp. 115-130, September 2002.
- 71
-
P. M. Baggenstoss, “Class-specific features in classification.,” in IASTED International Conference on Signal and Image Processing, 1998.
- 72
-
P. M. Baggenstoss, “A multi-resolution hidden markov model using
class-specific features,” Proceedings of EUSIPCO 2008, Lausanne,
Switzerland, Aug 2008.
- 73
-
P. M. Baggenstoss, “A multi-resolution hidden markov model using
class-specific features,” IEEE Transactions on Signal Processing,
vol. 58, pp. 5165-5177, Oct 2010.
- 74
-
B. F. Harrison and P. M. Baggenstoss, “A multi-resolution hidden markov model
for optimal detection, tracking, separation and classification of marine
mammal vocalizations,” in Proc. Oceans 2008, Quebec City, September
2008.
- 75
-
M. Ostendorf, V. Digalakis, and O. A. Kimball, “From hmms to segment models: A
unifed view of stochastic modeling for speech recognition,” IEEE
Transactions on Speech and Audio Processing, 1996.
- 76
-
J. A. Bilmes, “Graphical models and automatic speech recognition,” Mathematical Foundations of Speech and Language Processing, 2003.
- 77
-
S. Furui, “Speaker-independent isolated word recognition using dynamic
features of speech spectrum,” IEEE Trans on ASSP, vol. 34, no. 1,
pp. 52-59, 1986.
- 78
-
K. Achan, S. Roweis, and B. Frey, “A segmental hmm for speech waveforms,”
University of Toronto Technical Report, Jan 2004.
- 79
-
D. Gamerman, Markov Chain Monte Carlo.
Chapman and Hall, 1997.
- 80
-
J. S. Garofolo, “Timit acoustic-phonetic continuous speech corpus,” Linguistic Data Consortium, 1993.
- 81
-
P. Baggenstoss, “Office sounds database,” http://class-specific.com/os.
Accessed: 2022-02-28.
- 82
-
P. M. Baggenstoss, “The projected belief network classifier: both generative
and discriminative,” Proceedings of EUSIPCO, Amsterdam, 2020.
- 83
-
M. Rosenblatt, “Remarks on a multivariate transformation,” The Annals of
Mathematical Statistics, vol. 23, no. 3, pp. 470-472, 1952.
- 84
-
G. Deco and W. Brauer, “Nonlinear higher-order statistical decorrelation by
volume-conserving neural architectures,” Neural Netw., vol. 8,
p. 525–535, June 1995.
- 85
-
G. Deco and D. Obradovic, AN INFORMATION-THEORETIC APPROACH TO NEURAL
COMPUTING.
Springer, 1996.
- 86
-
A. Hyvärinen, J. Karhunen, and E. Oja, Independent Component
Analysis.
Wiley, 2001.
- 87
-
A. Hyvärinen and P. Pajunen, “Nonlinear independent component analysis:
Existence and uniqueness results,” Neural Networks, vol. 12, no. 3,
pp. 429-439, 1999.
- 88
-
P. M. Baggenstoss, “Applications of projected belief networks (pbn),” in Proceedings of EUSIPCO 2019, (La Coruña, Spain), Sep 2019.
- 89
-
M. Welling, M. Rosen-Zvi, and G. Hinton, “Exponential family harmoniums with
an application to information retrieval,” Advances in neural
information processing systems, 2004.
- 90
-
G. E. Hinton, S. Osindero, and Y.-W. Teh, “A fast learning algorithm for deep
belief nets,” in Neural Computation 2006, 2006.
- 91
-
D. J. Rezende, S. Mohamed, and D. Wierstra, “Stochastic backpropagation and
approximate inference in deep generative models,” in Proceedings of the
31st International Conference on Machine Learning (E. P. Xing and T. Jebara,
eds.), vol. 32 of Proceedings of Machine Learning Research, (Bejing,
China), pp. 1278-1286, PMLR, 22-24 Jun 2014.
- 92
-
I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair,
A. Courville, and Y. Bengio, “Generative adversarial nets,” in Advances in Neural Information Processing Systems 27 (Z. Ghahramani,
M. Welling, C. Cortes, N. D. Lawrence, and K. Q. Weinberger, eds.),
pp. 2672-2680, Curran Associates, Inc., 2014.
- 93
-
V. Vapnik, The Nature of Statistical Learning.
Springer, 1999.
- 94
-
I. Goodfellow, Y. Bengio, A. Courville, and Y. Bengio, Deep Learning.
Cambridge, MA: MIT press, 2016.
- 95
-
H. L. Royden, Real Analysis.
Englewood Cliffs, New Jersey, USA: Prentice Hall, third ed., 1988.
- 96
-
T. Minka, “Exemplar-based likelihoods using the pdf projection theorem.,”
Microsoft Research Ltd, technical report, March 2004.
- 97
-
D. J. Olive, “Applied robust statistics,” Unpublished online text,
Chapter 4, http://lagrange.math.siu.edu/Olive/ol-bookp.htm, 2008.
- 98
-
D. J. Olive, Robust Multivariate Analysis (to appear).
New York: Springer, 2017.