[1]
M. Liberman, et al., “Emotional Prosody Speech and Transcripts,” Philadelphia; Linguistic Data Consortium. 2002.
[2]
T. Huang and Y. Yang, “Learning Virtual HD Model for Bi-model Emotional Speaker Recognition,” ICPR, Istanbul, Turkey, 23-26 Aug. 2010, pp. 1614-1617.
[3]
D. G. Childers, J. J. Yea and E. L. Bocchieri, “Source/ Vocal-tract Interaction in Speech and Singing Synthesis,” Proc Stockholm Music Acoust Conf, 1983, pp. 125-141.
[4]
D. G. Childers and C. F. Wong, “Measuring and Modeling Vocal Source-Tract Interaction, Ieee Transactions on Biomedical Engineering, Vol. 41, No. 7, 1994, pp. 663-671.
doi:10.1109/10.301733
[5]
H. Akaike, “A New Look at the Statistical Model Identification,” Automatic Control, IEEE Transactions on, Vol. 19, No. 6, 1974, pp. 716-723.
doi:10.1109/TAC.1974.1100705
[6]
L. Rabiner, “On the Use of Autocorrelation Analysis for Pitch Detection, Acoustics,Speech and Signal Processing,” IEEE Transactions on, Vol. 25, No. 1, 1977, pp. 24-33.
doi:10.1109/TASSP.1977.1162905
[7]
L. H. Cai, D. Z. Huang and R. Cai, “Basis of Modern Speech Technology and Application,” Tsinghua University press, Beijing, China, 2003.