Zoom out Search Issue
IEEE SIGNAL PROCESSING MAGAZINE [113] MARCH 2015
AUTHORS
Kaushik Sunder (KAUSHIK1@e.ntu.edu.sg) received his B.Tech
degree in electrical and electronics engineering from the
National Institute of Technology Karnataka, Surathkal, India, in
2011. He is currently pursuing his Ph.D. degree in electrical and
electronics engineering at Nanyang Technological University,
Singapore. His research interest includes spatial audio, psycho-
acoustics, and music signal processing.
Jianjun He (JHE007@e.ntu.edu.sg) received his B.Eng.
degree in automation from Nanjing University of Posts and
Telecommunications, China, in 2011 and is currently pursuing
his Ph.D. degree in electrical and electronic engineering at
Nanyang Technological University, Singapore. His research
interests include audio and acoustic signal processing, three-
dimensional audio, psychoacoustics, active noise control, and
emerging audio and speech applications.
Ee-Leng Tan (ETanEL@ntu.edu.sg) received his B.Eng.
(first class honors) and Ph.D. degrees in electrical and electron-
ic engineering from Nanyang Technological University (NTU),
Singapore, in 2003 and 2012, respectively. Currently, he is with
NTU as a research fellow. His research interests include image/
audio processing and real-time digital signal processing.
Woon-Seng Gan (ewsgan@ntu.edu.sg) received his B.Eng.
(first class honors) and Ph.D. degrees in electrical and electronic
engineering from the University of Strathclyde, United Kingdom,
in 1989 and 1993, respectively. He is currently an associate pro-
fessor and the head of the Information Engineering Division,
School of Electrical and Electronic Engineering at Nanyang
Technological University. His research interests span a wide and
related areas of adaptive signal processing, active noise control,
and directional sound system.
REFERENCES
[1] D. R. Begault, 3-D Sound for Virtual Reality and Multimedia. Cambridge, MA:
AP Professional, 2000.
[2] S. Spors, H. Wierstorf, A. Raake, F. Melchior, M. Frank, and F. Zotter, “Spatial
sound with loudspeakers and its perception: A review of the current state,” Proc.
IEEE, vol. 101, no. 9, pp. 1920–1938, Sept. 2013.
[3] V. Pulkki, “Spatial sound reproduction with directional audio coding,” J. Audio
Eng. Soc., vol. 55, no. 6, pp. 503–516, June 2007.
[4] S. Olive, T. Welti, and E. McMullin, “Listener preferences for different head-
phone target response curves,” in Proc. 134th Audio Engineering Society Conven-
tion, Rome, Italy, May 2013, pp. 1–12.
[5] M. M. Goodwin and J. M. Jot, “Binaural 3-D audio rendering based on spatial
audio scene coding,” in Proc. 123rd Audio Engineering Society Convention,New
York,Oct.2007, pp. 1–12.
[6] J. Breebaart and E. Schuijers, “Phantom materialization: A novel method to
enhance stereo audio reproduction on headphones,” IEEE Trans. Audio, Speech,
Lang. Processing, vol. 16, no. 8, pp. 1503–1511, Nov. 2008.
[7] C. Avendano and J. M. Jot, “A frequency-domain approach to multichannel up-
mix,” J. Audio Eng. Soc., vol. 52, no. 7/8, pp. 740–749, July 2004.
[8] C. Faller, “Multiple-loudspeaker playback of stereo signals,” J. Audio Eng. Soc.,
vol. 54, no. 11, pp. 1051–1064, Nov. 2006.
[9] F. Menzer and C. Faller, “Stereo-to-binaural conversion using interaural coher-
ence matching,” in Proc. 128th Audio Engineering Society Convention,London,
UK, May 2010, pp. 1–14.
[10] D. R. Begault, E. M. Wenzel, and M. R. Anderson, “Direct comparison of the im-
pact of head tracking, reverberation, and individualized head-related transfer functions
on the spatial perception of a virtual speech source,” J. Audio Eng. Soc., vol. 49, no. 10,
pp. 904–916, Oct. 2001.
[11] C. Faller and F. Baumgarte, “Binaural cue coding—Part II: Schemes and appli-
cations,” IEEE Trans. Speech Audio, Speech, Lang. Processing, vol. 11, no. 6, pp.
520–531, Nov. 2003.
[12] V. R. Algazi and R. O. Duda, “Headphone-based spatial sound,” IEEE Signal Pro-
cessing Mag., vol. 28, no. 1, pp. 33–42, Jan. 2011.
[13] R. Nicol, Binaural Technology. New York: Audio Engineering Society, Inc., 2010.
[14] A. Hyvärinen, J. Karhunen, and E. Oja, Independent Component Analysis.
Hoboken, NJ: Wiley, 2004.
[15] M. D. Plumbley, T.
Blumensath, L. Daudet, R. Gribonval, and M. E. Davies,
“Sparse representations in audio and music: From coding to source separation,” Proc.
IEEE, vol. 98, no. 6, pp. 995–1005, June 2010.
[16] O. Yilmaz and S. Rickard, “Blind separation of speech mixtures via time-frequency
masking,” IEEE Trans. Signal Processing, vol. 52, no. 7, pp. 1830–1847, July 2004.
[17] T. Virtanen, “Sound source separation in monaural music signals,” Ph.D. thesis,
Tampere Univ. of Technology, 2006.
[18] E. Vincent, N. Bertin, R. Gribonval, and F. Bimbot, “From blind to guided audio
source separation,” IEEE Signal Processing Mag., vol. 31, no. 3, pp. 107–115, 2014.
[19] D. Wang and G. J. Brown, Computational Auditory Scene Analysis: Principles,
Algorithms, and Applications. Hoboken, NJ: Wiley-IEEE Press, 2006.
[20] J. Merimaa, M. M. Goodwin, and J. M. Jot, “Correlation-based ambience extrac-
tion from stereo recordings,” in Proc. 123rd Audio Engineering Society Convention,
New York, Oct. 2007, pp. 1–15.
[21] J. He, E. L. Tan, and W. S. Gan, “Linear estimation based primary-ambient extrac-
tion for stereo audio signals,” IEEE/ACM Trans. Audio, Speech, Lang. Processing, vol.
22, no. 2, pp. 505–517, 2014.
[22] J. He, E. L. Tan, and W. S. Gan, “Time-shifted principal component analysis based
cue extraction for stereo audio signals,” in Proc. IEEE Int. Conf. Acoustics, Speech
and Signal Processing (ICASSP), Canada, May 2013, pp. 266–270.
[23] J. He, E. L. Tan, and W. S. Gan, “A study on the frequency-domain primary-ambi-
ent extraction for stereo audio signals,” in Proc. IEEE Int. Conf. Acoustics, Speech and
Signal Processing (ICASSP)
, Florence, Italy, 2014, pp. 2892–2896.
[24] J. Thompson, B. Smith, A. Warner, and J. M. Jot, “Direct-diffuse decomposition
of multichannel signals using a system of pair-wise correlations,” in Proc. 133rd Audio
Engineering Society Convention, San Francisco, CA, 2012, pp. 1–15.
[25] H. Møller, M. F. Sørensen, D. Hammershøi, and C. B. Jensen, “Head-related
transfer functions of human subjects,” J. Audio Eng. Soc., vol. 43, no. 5, pp. 300–321,
May 1995.
[26] S. Xu, Z. Li, and G. Salvendy, “Individualization of head-related transfer function
for three-dimensional virtual auditory display: A review,” in Virtual Reality, R. Shu-
maker, Ed. New York: Springer, 2007, pp. 397–407.
[27] G. Enzner, “3D-continuous-azimuth acquisition of head-related impulse respons-
es using multi-channel adaptive filtering,” in Proc. IEEE Workshop on Applications
of Signal Processing to Audio and Acoustics (WASPAA), Oct. 2009, pp. 325–328.
[28] M. Rothbucher, M. Durkovic, H. Shen, and K. Diepold, “HRTF customization
using multiway array analysis,” in Proc. 18th European Signal Processing Conf. (EU-
SIPCO’10), Aalborg, Denmark, Aug. 2010, pp. 229–233.
[29] R. O. Duda,
V. R. Algazi, and D. M. Thompson, “The use of head-and-torso mod-
els for improved spatial sound synthesis,” in Proc. 113th Audio Engineering Society
Convention, Los Angeles, CA, Oct. 2002, pp. 1–18.
[30] D. N. Zotkin, J. Hwang, R. Duraiswaini, and L. S. Davis, “HRTF personaliza-
tion using anthropometric measurements,” in Proc. IEEE Workshop on Applications
of Signal Processing to Audio and Acoustics (WASPAA),New York,Oct.2003, pp.
157–160.
[31] J. C. Middlebrooks, “Individual differences in external-ear transfer functions re-
duced by scaling in frequency,” J. Acoust. Soc. Amer., vol. 106, no. 3, pp. 1480–1492,
Sept. 1999.
[32] K. J. Fink and L. Ray, “Tuning principal component weights to individualize
HRTFs,” in Proc. IEEE Int. Conf. Acoustics, Speech and Signal Processing (ICASSP),
Kyoto, Mar. 2012, pp. 389–392.
[33] K. Sunder, E. L. Tan, and W. S. Gan, “Individualization of binaural syn-
thesis using frontal projection headphones,” J. Audio Eng. Soc., vol. 61, no. 12,
pp. 989–1000,Dec.2013.
[34] R. Nicol, V. Lemaire, A. Bondu, and S. Busson, “Looking for a relevant similarity
criterion for HRTF clustering: A comparative study,” in Proc. 120th Audio Engineer-
ing Society Convention, Paris, France, May 2006, pp. 1–14.
[35] H. Møller, D. Hammershoi, C. B. Jensen, and M. F. Sorensen, “Transfer charac-
teristics of headphones measured on human ears,” J. Audio Eng. Soc., vol. 43, no. 4,
pp. 203–217, Apr. 1995.
[36] V. Larcher, J. M. Jot, and G. Vandernoot, “Equalization methods in binaural tech-
nology,” in Proc. 105th Audio Engineering Society Convention, San Francisco, CA,
Sept. 1998, pp. 1–29.
[37] A. Kulkarni and H. S. Colburn, “Variability in the characterization of the headphone
transfer-function,” J. Acoust. Soc. Amer., vol. 107, no. 2, pp. 1071–1074, Feb. 2000.
[38] H. Møller, C. B. Jensen, D. Hammershøi, and M. F. Sørensen, “Design criteria for
headphones,” J. Audio Eng. Soc., vol. 43, no. 4, pp. 218–232, Apr. 1995.
[39] W. S. Gan and E. L. Tan, “Listening device and accompanying signal processing
method,” U.S. Patent 2014/0153765 A1, 2014.
[SP]
Previous Page | Contents | Zoom in | Zoom out | Front Cover | Search Issue | Next Page
q
q
M
M
q
q
M
M
q
M
THE WORLD’S NEWSSTAND
®
Previous Page | Contents | Zoom in | Zoom out | Front Cover | Search Issue | Next Page
q
q
M
M
q
q
M
M
q
M
THE WORLD’S NEWSSTAND
®
_______________
______________
_____________
____________