[1] S. Young, “A review of large-vocabulary continuous-speech recognition,” IEEE Signal Processing Magazine, vol. 13, no. 5, pp. 45–57, Sep. 1996.
[2] J. Holmes, Speech Synthesis and Recognition. Van Nostrand Reinhold (UK) Co. Ltd, 1988.
[3] K. Demuynck, J. Duchateau, D. Van Compernolle, and P. Wambacq, “Improved feature decorrelation for HMM-based speech recognition,” in Proc. International Conference on Spoken Language Processing, vol. VII, Sydney, Australia, Dec. 1998, pp. 2907–2910.
[4] J. Duchateau, K. Demuynck, D. Van Compernolle, and P. Wambacq, “Class definition in discriminant feature analysis,” in Proc. European Conference on Speech Communication and Technology, vol. III, Aalborg, Denmark, Sep. 2001, pp. 1621–1624.
[5] K. Demuynck, J. Duchateau, and D. Van Compernolle, “Optimal feature sub-space selection based on discriminant analysis,” in Proc. European Conference on Speech Communication and Technology, vol. III, Budapest, Hungary, Sep. 1999, pp. 1311–1314.
[6] D. Lee and H. Seung, “Learning the parts of objects by non-negative matrix factorization,” Nature, vol. 401, pp. 788–791, 1999.
[7] T. Landauer, P. Foltz, and D. Laham, “Introduction to latent semantic analysis,” Discourse Processes, vol. 25, pp. 259–284, 1998.
[8] T. Hofmann, “Probabilistic latent semantic analysis,” in Proc. of Uncertainty in Artificial Intelligence, Stockholm, 1999.
[9] T. Hofmann and J. Puzicha, “Unsupervised learning from dyadic data,” International Computer Science Insitute, Berkeley, CA, Tech. Rep. TR-98-042, 1998.
[10] A. Dempster, N. Laird, and D. Rubin, “Maximum likelihood from incomplete data via the em-algorithm,” Journal of the Royal Statistical Society, Series B, vol. 39, pp. 1–38, 1977.
[11] E. Gaussier and C. Goutte, “Relation between plsa and nmf and implications,” in Proceedings of the ACM SIGIR conference on research and development in information retrieval, Salvador, Brazil, 2005, pp. 601–602.
[12] D. Lee and H. Seung, “Algorithms for non-negative matrix factorization,” Advances in Neural Information Processing Systems, vol. 13, pp. 556–562, 2001.
[13] L. Lamel, R. Kassel, and S. Seneff, “Speech database development: Design and analysis of the acoustic-phonetic corpus,” in Proc. DARPA Speech Recognition Workshop, 1986, pp. 100–109.
[14] W. Fisher, V. Zue, J. Bernstein, and D. Pallett, “An acoustic-phonetic data base,” The Journal of the Acoustical Society of America, vol. 81, no. S1, pp. S92–S93, May 1987.
[15] Z. Hafed and M. Levine, “Face recognition using the discrete cosine transform,” International Journal of Computer Vision, vol. 43, no. 3, pp. 167–188, 2001.
[16] V. Eguiluz, M. Ospeck, Y. Choe, A. Hudspeth, and M. O. Magnasco, “Essential nonlinearities in hearing,” Physical Review Letters, vol. 84, no. 22, pp. 5232–5235, 2000.
[17] D. S. Pallett, “Benchmark tests for darpa resource management database performance,” in Proc. International Conference on Acoustics, Speech and Signal Processing, Glasgow, UK, May 1989, pp. 536–539.
[18] R. Patterson, K. Robinson, J. Holdsworth, D. McKeown, C. Zhang, and M. Allerhand, “Complex sounds and auditory images,” Auditory Physiology and Perception, Proc. 9th International Symposium on Hearing, 1992.
[19] K. Demuynck, “Extracting, modelling and combining information in speech recognition,” Ph.D. dissertation, K.U.Leuven, ESAT, Feb. 2001.
[20] P. O. Hoyer, “Non-negative matrix factorization with sparseness constraints,” ’Journal of Machine Learning Research’, vol. 5, pp. 1457–1469, 2004.