Publications
  2005   2006   2007   2008   2009   2010

Peer-reviewed Journal Papers

  1. Stefan R. Schweinberger, Christoph Casper, Nadine Hauthal, Juergen M. Kaufmann, Hideki Kawahara, Nadine Kloth, David M.C. Robertson, Adrian P. Simpson and Romi Zaeske, Auditory Adaptation in Voice Perception, Current Biology 18, 684-688, May 6, (2008).

Peer-reviewed Conference Papers

  1. Hiromasa Fujihara and Masataka Goto: Three Techniques for Improving Automatic Synchronization Between Music and Lyrics: Fricative Detection, Filler Model, and Novel Feature Vectors for Vocal Activity Detection, Proceedings of the 2008 IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2008), pp.69-72, April 2008.
  2. *Ken-ichi Miyamoto, Hirokazu Kameoka, Takuya Nishimoto, Nobutaka Ono, Shigeki Sagayama: Harmonic-Temporal-Timbral Clustering (HTTC) For the Analysis of Multi-instrument Polyphonic Music Signals, Proc. of ICASSP, April 2008.
  3. Hideki Kawahara, Masanori Morise, Toru Takahashi, Ryuichi Nisimura, Toshio Irino, Hideki Banno: "TANDEM-STRAIGHT: A Temporally Stable Power Spectral Representation for Periodic Signals and Applications to Interference-free Spectrum, F0, and Aperiodicity Estimation", Proc. ICASSP2008 (The 33nd International Conference on Acoustics, Speech, and Signal Processing), pp. 3933-3936, Las Vegas, Nevada, USA, March 30 - April 4, 2008.
  4. Hirokazu Kameoka, Nobutaka Ono, Shigeki Sagayama: Auxiliary Function Approach to Parameter Estimation of Constrained Sinusoidal Model, Proc. of 2008 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP2008), April 2008.
  5. Hideki Kawahara, Masanori Morise, Toru Takahashi, Ryuichi Nishimura, Hideki Banno, Toshio Irino: A temporally stable representation of power spectra of periodic signals and its application to F0 and periodicity estimation, Journal of Acoustic Society of America, Vol.123, No.5 (May 2008) Pt.2, pp.3074-3075.
  6. Hideki Kawahara, Masanori Morise, Toru Takahashi, Ryuichi Nisimura, Hideki Banno, Toshio Irino, A unified approach for F0 extraction and aperiodicity estimation based on a temporally stable power spectral representation, ISCA Tutorial and Research Workshop (ITRW) on "Speech Analysis and Processing for Knowledge Discovery", Aalborg, 4-6 June (2008)
  7. Hosokawa, T., Kazai, K., Yagi, A., & Katayose, H.. Simultaneous recording of fNIRS and SCR improves lie detection accuracy. The 14th Annual Meeting of the OHBM, #108 T-PM, Melbourne, Australia. 2008.6.
  8. Fujisawa, T. X., Hosokawa, T., Nagata, N. & Katayose, H.. Investigation of Brain Activity under Social Pressure using the Asch Paradigm: An fNIRS Study. The 14th Annual Meeting of the OHBM, #261 M-AM, Melbourne, Australia. 2008.6.
  9. Takahashi, R., Fujisawa, T. X., Nagata, N., Sugio, T., Inokuchi, S. (2008). Brain Activity in Colored-hearing Synesthetes When Listening to Tonal Music. Proc. 14th Annual Meeting of the Organization for Human Brain Mapping, 700T-PM.
  10. Jonathan Le Roux, Hirokazu Kameoka, Nobutaka Ono, Alain de Cheveigne, Shigeki Sagayama, "Music and speech signal processing using harmonic-temporal clustering," Acoustics'08, p.3805, Jul., 2008.
  11. Nobutaka Ono, Ken-ichi Miyamoto, Jonathan Le Roux, Hirokazu Kameoka, Shigeki Sagayama, "Separation of a Monaural Audio Signal into Harmonic/Percussive Components by Complementary Diffusion on Spectrogram," Proc. of EUSIPCO, Aug., 2008.
  12. Takeshi Mizumoto, Ryu Takeda, Kazuyoshi Yoshii, Kazunori Komatani, Tetsuya Ogata, Hiroshi G. Okuno: A Robot Listens to Music and Counts Its Beats Aloud by Separating Music from Counting Voice, Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS-2008), accepted, IEEE, RSJ, Nice, Sept. 2008. Award for Entertainment Robots and Systems (NTF Award) Nomination Finalist.
  13. Tomomi Sangenya, Takashi X. Fujisawa, & Noriko Nagata (accepted).Creation of a Sound-Image Scale - Quantification of the Images of Chord Progressions with Impression Evaluation Used -, 2008 IEEE International Conference on Systems, Man, and Cybernetics (SMC 2008).
  14. Murata, M., Harada, H., Fujisawa, T. X., Nagata, N. (2008). Facial expression analysis using motion capture and eye tracking. Proc. 14th Korea-Japan Joint Workshop on Frontiers of Computer Vision (FCV2008), 380-385.
  15. Kitahara, T., Katsura, M., Katayose, H., & Nagata, N. (accepted).Computational Model for Automatic Chord Voicing based on Bayesian Network. The 10th International Conference on Music Perception and Cognition.
  16. Hashida, M., Nakra, M., Katayose, H., Murao, T., Hirata, K., Suzuki, K., Kitahara, T.: Rencon: Performance Rendering Contest for Automated Music Systems, Proc. of International Conference on Music Perception and Cognition, PM1-R06(SY08), 2008.
  17. Tetsuro Kitahara, Makiko Katsura, Haruhiro Katayose, and Noriko Nagata: gComputational Model for Automatic Chord Voicing based on Bayesian Networkh, Proceedings of the 10th International Conference on Music Perception and Cognition (ICMPC 2008),August 2008.
  18. Furuya S, Aoki T, Nakahara H, Kinoshita H (2008) Kinematics and muscular activity of upper extremity movements in piano keystroke by professional pianists. Proceedings of 10th International Conference of Music Perception and Cognition, 293-296, CD-ROM
  19. Toshie Matsui, Koji Kazai, Minoru Tsuzaki and Haruhiro Katayose (2008).@"Investigation of the musician's brain activation during different music listening modes: A Near-Infrared Spectroscopy study,"Proc. of 10th International Conference on Music Perception and Cognition, Sapporo, August 25-29.
  20. Tetsuro Kitahara, Masahiro Nishiyama, and Hiroshi G. Okuno: gComputational Model of Congruency between Music and Videoh, Proceedings of the 10th International Conference on Music Perception and Cognition (ICMPC 2008), August 2008. (abstract only)
  21. Nakahara H, Furuya S, Masuko T, Obata S, Kinoshita H (2008) Emotion-related autonomic nerve activity with musical performance and perception. The 10th International Conference on Music Perception and Cognition, Hokkaido
  22. Takehiro Abe, Katsutoshi Itoyama, Kazuyoshi Yoshii, Kazunori Komatani, Tetsuya Ogata, Hiroshi G. Okuno: Synthesis Approach for Manipulating Pitch of a Musical Instrument Sound with Considering Timbral Characteristics, Proceeding of the 11th International Conference on Digital Audio Effects (DAFx-08), accepted Espoo, Finland, Sep.1-4. 2008.
  23. Hashida, M. & Katayose, H. A Directable Performance Rendering System: Itopul. Proceedings of New Instruments on Music Expression (NIME), pp.277-280, 2008
  24. Nobutaka Ono, Ken-ichi Miyamoto, Hirokazu Kameoka, Shigeki Sagayama, "A Real-time Equalizer of Harmonic and Percussive Components in Music Signals," Proc. of ISMIR, pp.139-144, Sep., 2008.
  25. Yusuke Tsuchihashi, Tetsuro Kitahara, and Haruhiro Katayose: gUsing Bass-line Features for Content-based MIRh, Proceedings of the 9thInternational Conference on Music Information Retrieval (ISMIR 2008), pp.620-625,September 2008.
  26. Katsutoshi Itoyama, Masataka Goto, Kazunori Komatani, Tetsuya Ogata, and Hiroshi G. Okuno: Instrument Equalizer for Query-by-Example Retrieval: Improving Sound Source Separation Based on Integrated Harmonic and Inharmonic Models, Proceedings of the 9th International Conference on Music Information Retrieval (ISMIR 2008), pp.133-138, September 2008.
  27. Kazuyoshi Yoshii and Masataka Goto: Music Thumbnailer: Visualizing Musical Pieces in Thumbnail Images Based on Acoustic Features, Proceedings of the 9th International Conference on Music Information Retrieval (ISMIR 2008), pp.211-216, September 2008.
  28. Hiromasa Fujihara, Masataka Goto, and Jun Ogata: Hyperlinking Lyrics: A Method for Creating Hyperlinks Between Phrases in Song Lyrics, Proceedings of the 9th International Conference on Music Information Retrieval (ISMIR 2008), pp.281-286, September 2008.
  29. Kohei Sumi, Kazuyoshi Yoshii, Kazunori Komatani, Tetsuya Ogata, Hiroshi G. Okuno: Automatic Chord Recognition Based on Probabilistic Integration of Chord Transition and Bass Pitch Estimation, Proceedings of 9th International Conference on Musical Information Retreival (ISMIR-2008), 39-44, Philadelphia, 15 Sep. 2008.
  30. Hashida, M., Matsui, T. & Katayose, H. A New Music Database Describing Deviation Information of Performance Expressions, International Conference of Music Information Retrieval (ISMIR), pp.489-494, 2008.
  31. Stanislaw Raczynski, Nobutaka Ono, Shigeki Sagayama, gUsing Dynamic Baysian Networks to Model Prior Distributions of Music Sequences,h International Conference of Music Information Retrieval (ISMIR), 2008. 137.
  32. Aoki T, Furuya S, Fukuoka Y, Kinoshita H (2008)Effect of long-term piano training on the motor function of individual fingers. Society for Neuroscience, Washington DC, USA
  33. Hideki Kawahara, Masanori Morise, Hideki Banno, Toru Takahashi, Ryuichi Nishimura, Toshio Irino: Spectral Envelope Recovery beyond the Nyquist Limit for High-Quality Manipulation of Speech Sounds, Proceedings of International Conference on Spoken Language Processing (Interspeech-2008), accepted, Brisbane, Sept. 2008.
  34. Tetsuro Kitahara, Yusuke Tsuchihashi, and Haruhiro Katayose: "Music Genre Classification and Similarity Calculation Using Bass-line Features", Proceesings of the 10th IEEE International Symposium on Multimedia, Workshop on Multimedia Audio and Speech Processing (ISM 2008 MASP Workshop), pp.574--579, December 2008.
  35. Nobuto Fujii, Mitsuyo Hashida, Haruhiro Katayose: Strategy-acquisition System for Video Trading Card Game, Intl. Conf. on Advances in Computer Entertainment Technology (ACE) 2008, T9-1, Dec. 2008.
  36. Masanori Morise, Hideki Kawahara, Haruhiro Katayose,``Fast and Reliable F0 Estimation Method Based on the PeriodExtraction of Vocal Fold Vibration of Singing Voice and Speech, ''AES 35th International Conference, 11-13 February 2009, London UK. (Accepted)
  37. Yuri Yoshida, Ryuichi Nisimura, Toshio Irino and Hideki Kawahara,``Vowel-Based Voice Conversion and its Application to Singing-Voice Manipulation, ''AES 35th International Conference, 11-13 February 2009, London UK. (Accepted)
  38. Nishimoto, M., Azuma, Y., Miyamoto, N., Fujisawa, T. X. & Nagata, N. (2008) Subjective Age Estimation System Using Speech Sounds: Comparison with Facial Images. The 2008 IEEE International Conference on Systems, Man, and Cybernetics (SMC 2008), pp. 1900-1904, Singapore.
  39. Sangenya, T., Fujisawa, T. X. & Nagata, N. (2008) The Feasibility of Establishing Sound Image Scale: Image measurement of chord progression using impression estimation method. The 2008 IEEE International Conference on Systems, Man, and Cybernetics (SMC 2008), pp. 1905-1909, Singapore.

Invited Talk

  1. Hiroshi G. Okuno, Kazuhiro Nakadai: COMPUTATIONAL AUDITORY SCENE ANALYSIS AND ITS APPLICATION TO ROBOT AUDITION, (invited talk), Proceedings of Hands-free Speech Communication and Microphone Arrays (HSCMA-2008), pp.123-127, May 7, 2008, Trento, Italy.
  2. Hideki Kawahra, "TANDEM-STRAIGHT, a research tool for L2 study enabling flexible manipulations of prosodic information", Speech Prosody 2008, Campinas Brazil, May 6-9, 2008.
  3. Masataka Goto: Keynote talk gActive Music Listening Interfaces Based on Music-Understanding Technologiesh, TELECOM ParisTech Workshop on Music Signal Processing, Paris, France, June 27, 2008.
  4. Masataka Goto and Hiromasa Fujihara, Modeling Vocal Sounds in Polyphonic Musical Audio Signals, The Journal of the Acoustical Society of America, Vol.123, No.5, Pt.2, p.3805, July 2008 (Invited Paper of the 155th Meeting of the Acoustical Society of America (Acoustics'08 Paris))
  5. Hideki Kawahara Revisiting power spectral representation of periodic signals, sampling theories and fundamental frequency estimation, ISCA SAPA Workshop, Brisbane, Australia, 21 September, 2008 26.
  6. Hiroshi G. Okuno: Computational Auditory Scene Analysis and Its Application to Robot Audition, (invited talk), Proceedings of 2nd International Symposium on Artificial Intelligence and Robots, Oct. 9, 2008, Univ. of Electro-Comm., Chofu, Japan.
  7. Masataka Goto: Invited talk gActive Music Listening Interfaces and Augmented Music-Understanding Interfacesh, Workshop on music processing and intelligent user interfaces to music, Tampere, Finland, December 11, 2008.
  8. Hideki Kawahara, Looking into the past: Power spectral representation of periodic signals, sampling theories and fundamental frequency estimation for remaking speech, The 6th ISCSLP 2008, Kunming, Chine, 16-19 December, 2008.
  9. Shigeki Sagayama, Hirokazu Kameoka and Haruto Takeda, gToward automatic music transcription from audio input,h The Journal of the Acoustical Society of America, Vol.123, No.6, Pt.2, p.26, Dec. 2008. (Invited Paper of the 156th Meeting of the Acoustical Society of America)
  10. Shigeki Sagayama, gWhat can speech researchers bring to music processing?h Proc. of 2008 6th International Symposium on Chinese Language Processing (ISCSLP2008), pp.19-20, 2008.