Publications

Google Scholar
Public Profile (Mendeley)
dblp: Engin Erzin

JOURNAL PUBLICATIONS

  1. Z. Bucinca, Y. Yemez, E. Erzin, M. Sezgin, “AffectON: Incorporating Affect into Dialog Generation,” in IEEE Transactions on Affective Computing, vol. 14, no. 1, pp. 823-835, 1 Jan.-March 2023, doi: 10.1109/TAFFC.2020.3043067.
  2. E. Kesim, T. Numanoglu, O. Bayramoglu, B.B. Turker, N. Hussain, M. Sezgin, Y. Yemez, and E. Erzin, “The eHRI database: A multimodal database of engagement in human-robot interactions,” Language Resources & Evaluation, vol. 57, no. 3, pp. 985–1009, 2023.
  3. B. Köprü and E. Erzin, “Use of Affective Visual Information for Summarization of Human-Centric Videos,” in IEEE Transactions on Affective Computing, 2022, doi: 10.1109/TAFFC.2022.3222882.
  4. N. Hussain, E. Erzin, T. M. Sezgin, and Y. Yemez, “Training Socially Engaging Robots: Modeling Backchannel Behaviors with Batch Reinforcement Learning,” in IEEE Transactions on Affective Computing, vol. 13, no. 4, pp. 1840-1853, 1 Oct.-Dec. 2022, doi: 10.1109/TAFFC.2022.3190233.
  5. R. Sadiq, E. Erzin, “Emotion Dependent Domain Adaptation for Speech Driven Affective Facial Feature Synthesis,” in IEEE Transactions on Affective Computing, vol. 13, no. 3, pp. 1501-1513, 1 July-Sept. 2022.
  6. S.N. Fatima, E. Erzin, “Use of affect context in dyadic interactions for continuous emotion recognition,” Speech Communication, Vol. 132, pp. 70-82, 2021.
  7. M.A.T. Turan, E. Erzin, “Improving phoneme recognition of throat microphone speech recordings using transfer learning,”  Speech Communication, Volume 129, pp. 25-32, 2021.
  8. M.A.T. Turan, E. Erzin, “Domain Adaptation for Food Intake Classification with Teacher/Student Learning,” IEEE Transactions on Multimedia, vol. 23, pp. 4220-4231, 2021.
  9. S. Asadiabadi, E. Erzin, “Vocal Tract Contour Tracking in rtMRI Using Deep Temporal Regression Network,” IEEE Transactions on Audio, Speech and Language Processing, Vol. 28, pp. 3053-3064, November 2020.
  10. S. Mojtahedi, E. Erzin, P. Ungan, “A Diversity Combination Model Incorporating an Inward Bias for Interaural Time-Level Difference Cue Integration in Sound Lateralization,” Applied Sciences. 2020; 10(18):6356.
  11. E. Bozkurt, Y. Yemez, E. Erzin ”Affective Synthesis and Animation of Arm Gestures from Speech Prosody,” Speech Communication, Vol. 119, pp. 1-11, May 2020.
  12. A. Isabekov, E. Erzin, ”On the Importance of Hidden Bias and Hidden Entropy in Representational Efficiency of the Gaussian-Bipolar Restricted Boltzmann Machines,” Neural Networks, Vol. 105, pp. 405-418, September 2018.
  13. B.B. Turker, Y. Yemez, M. Sezgin, E. Erzin, “Audio-Facial Laughter Detection in Naturalistic Dyadic Conversations,” IEEE Transactions on Affective Computing, Volume: 8, Issue: 4, pp. 534 – 545,  2017.
  14. E. Bozkurt, H. Khaki, S. Kececi, B.B. Turker, Y. Yemez, E. Erzin ”The JESTKOD Database: An affective multimodal database of dyadic interactions,” Language Resources and Evaluation, Volume 51, Issue 3, pp. 857–872, September 2017.
  15. E. Bozkurt, Y. Yemez, E. Erzin, ”Multimodal Analysis of Speech and Arm Motion for Prosody-Driven Synthesis of Beat Gestures,” Speech Communication, Volume 85, pp. 29–42, December 2016.
  16. M.A.T. Turan, E. Erzin, ”Source and Filter Estimation for Throat-Microphone Speech Enhancement,” IEEE/ACM Trans. on Audio, Speech and Language Processing, Volume 24, Issue 2, pp. 265-275, 2016.
  17. C. Yagli, M.A.T. Turan, E. Erzin, ”Artificial Bandwidth Extension of Spectral Envelope along a Viterbi Path,” Speech Communication, Volume 55, Issue 1, pp. 111-118, Jan. 2013.
  18. F. Ofli, E. Erzin, Y. Yemez, A.M. Tekalp, ”Learn2Dance: Learning Statistical Music-to-Dance Mappings for Choreography Synthesis,” IEEE Transactions on Multimedia, Volume 14, Issue 3, pp. 747-759, 2012.
  19. E. Bozkurt, C. Eroglu Erdem, T. Erdem and E. Erzin ”Formant Position based Weighted Spectral Features for Emotion Recognition,” Speech Communication, Volume 53, pp. 1186-1197, Nov-Dec 2011.
  20. E. Erzin, ”Improving Throat Microphone Speech Recognition by Joint Analysis of Throat and Acoustic Microphone Recordings,” IEEE Transactions on Audio, Speech and Language Processing, vol. 17, no. 7, pp. 1316-1324, September 2009.
  21. F. Ofli, Y. Demir, Y. Yemez, E. Erzin, A.M. Tekalp, K. Balci, I. Kizoglu, L. Akarun, C. Canton-Ferrer, J. Tilmanne, ve E. Bozkurt, “An audio-driven dancing avatar,” Journal on Multimodal User Interfaces, vol. 2, no. 2, pp. 93-103, September 2008.
  22. M. E. Sargın, Y. Yemez, E. Erzin, and A. M. Tekalp, ”Analysis of Head Gesture and Prosody Patterns for Prosody-Driven Head-Gesture Animation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, Volume 30, Issue 8, pp. 1330-1345, August 2008.
  23. M. E. Sargın, Y. Yemez, E. Erzin, and A. M. Tekalp, ”Audio-Visual Synchronization and Fusion using Canonical Correlation Analysis,” IEEE Transactions on Multimedia, Vol. 9, No. 7, pp. 1396-1403, November 2007.
  24. U. Bagcı, E. Erzin ”Automatic Classification of Musical Genres Using Inter-Genre Similarity,” IEEE Signal Processing Letters, Vol. 14, No. 8, pp. 521-524, August 2007.
  25. H.E. Cetingul, E. Erzin, Y. Yemez, A. M. Tekalp, “Multimodal Speaker/Speech Recognition using Lip Motion, Lip Texture and Audio,” Signal Processing (Elsevier), Vol.86, Issue 12, pp. 3549-3558, December 2006.
  26. Osman N. Yogurtcu, Engin Erzin, Attila Gursoy ”Extracting gene regulation information from microarray time-series data using hidden Markov models,” 21st International Symposium on Computer and Information Sciences (ISCIS 2006), Lecture Notes in Computer Science, LNCS 4263, pp. 144-153, Springer Berlin Heidelberg 2006.
  27. H.E. Cetingul, Y. Yemez, E. Erzin, A. M. Tekalp, “Discriminative Analysis of Lip Motion Features for Speaker Identification and Speech-Reading,” IEEE Transactions on Image Processing, Vol: 15, Issue: 10, pp: 2879 – 2891, Oct. 2006.
  28. E. Erzin, Y. Yemez, A. M. Tekalp, A. Ercil, H. Erdogan and H. Abut “Multimodal Person Recognition for Human-Vehicle Interaction,” IEEE MultiMedia, Vol. 13, No. 2, pp.18-31, April-June 2006.
  29. U. Bagcı, E. Erzin ”Boosting Classifiers for Music Genre Classification,” 20th International Symposium on Computer and Information Sciences (ISCIS 2005), Lecture Notes in Computer Science, LNCS 3733, pp. 575-584, Springer-Verlag Berlin Heidelberg 2005.
  30. E. Erzin, Y. Yemez, A. M. Tekalp, “Multimodal Speaker Identification Using an Adaptive Classifier Cascade based on Modality Reliability,” IEEE Transactions on Multimedia, Vol. 7, No. 5, pp. 840- 852, October 2005.
  31. F. Jabloun, A. E. Cetin, and E. Erzin, “Teager Energy Based Feature Parameters for Speech Recognition in Car Noise,” IEEE Signal Processing Letters, Vol. 6, No. 10, pp. 259-261, October 1999.
  32. E. Erzin, A. E. Cetin, “Line Spectral Frequency Representation of Subbands for Speech Recognition,” Signal Processing, Vol. 44, No. 1, pp. 117-119, June 1995.
  33. O. Arıkan, A.E. Cetin, E. Erzin, “Adaptive Filtering for Non-Gaussian Processes,” IEEE Signal Processing Letters, vol. 1, no. 11, pp. 163-165, Nov. 1994. Also presented in part at ”Twenty-eighth Annual Conference on Information Sciences and Systems”, Princeton, NJ, March 1994.
  34. E. Erzin, A.E. Cetin, “Interframe Differential Coding of Line Spectrum Frequencies,” IEEE Transactions on Speech and Audio Processing, vol. 2, no. 2, pp. 350-352, April 1994. Also presented in part at Twenty-sixth Annual Conference on Information Sciences and Systems, Princeton, NJ. March 1992.

CONFERENCE PUBLICATIONS

  1. Ibrahim Shoer, Berkay Köprü, Engin Erzin, “Role of Audio in Audio-Visual Video Summarization,” 2023 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops – IWCIM 2023: The Eleventh International Workshop on Computational Intelligence for Multimedia Understanding, Rhodes, Greece.
  2. B. Köprü, E. Erzin, “Affective Burst Detection from Speech using Kernel-fusion Dilated Convolutional Neural Networks,” 30th European Signal Processing Conference (EUSIPCO), 2022, pp. 105-109, Belgrade, Serbia.
  3. Ö.Z. Bayramoglu, E. Erzin, T.M. Sezgin, Y. Yemez, “Engagement Rewarded Actor-Critic with Conservative Q-Learning for Speech-Driven Laughter Backchannel Generation,” in ACM International Conference on Multimodal Interaction (ICMI’21), October 18–22, 2021, Montréal, QC, Canada.
  4. E. Kesim, E. Erzin, “Investigating Contributions of Speech and Facial Landmarks for Talking Head Generation,” INTERSPEECH: Annual Conference of the International Speech Communication Association, Brno, Czechia, 2021.
  5. R. Sadiq, S. Asadiabadi, E. Erzin, “Emotion Dependent Facial Animation from Affective Speech,” in IEEE 22nd International Workshop on Multimedia Signal Processing (MMSP), 2020.
  6. S. Asadiabadi, E. Erzin “Automatic Vocal Tract Landmark Tracking in rtMRI using Fully Convolutional Networks and Kalman Filter,” IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Barcelona, Spain, May 2020.
  7. N. Hussain, E. Erzin, T. M. Sezgin, and Y. Yemez, “Speech Driven Backchannel Generation using Deep Q-Network for Enhancing Engagement in Human-Robot Interaction,” INTERSPEECH: Annual Conference of the International Speech Communication Association, Graz, Austria, 2019.
  8. N. Hussain, E. Erzin, T. M. Sezgin, and Y. Yemez, “Batch Recurrent Q-Learning for Backchannel Generation Towards Engaging Agents,” 8th International Conference on Affective Computing and Intelligent Interaction (ACII), 2019.
  9. S. Asadiabadi, E. Erzin, “A Deep Learning Approach for Data Driven Vocal Tract Area Function Estimation,” in the IEEE Workshop on Spoken Language Technology (SLT), Athens, Greece, December 18-21, 2018.
  10. S. Asadiabadi, R. Sadiq, E. Erzin, “Multimodal Speech Driven Facial Shape Animation Using Deep Neural Networks,” in Proceedings – 10th Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA, Honolulu, Hawaii, 2018.
  11. M. A. T. Turan and E. Erzin, “Monitoring Infant’s Emotional Cry in Domestic Environments using the Capsule Network Architecture,” in INTERSPEECH: Annual Conference of the International Speech Communication Association, Hyderabad, India, 2018.
  12. B. B. Turker, E. Erzin, Y. Yemez, and M. Sezgin, “Audio-Visual Prediction of Head-Nod and Turn-Taking Events in Dyadic Interactions,” in INTERSPEECH: Annual Conference of the International Speech Communication Association, Hyderabad, India, 2018.
  13. M. A. T. Turan and E. Erzin, “Detection of Food Intake Events from Throat Microphone Recordings using Convolutional Neural Networks,” in the ICME workshop on Multimedia Services and Technologies for Smart-Health, San Diego, July 23-27, 2018.
  14. L. Devillers, S. Rosset, G. Dubuisson Duplessis, L. Bechade, Y. Yemez, B.B. Turker, M. Sezgin, E. Erzin, K. El Haddad, S. Dupont, P. Deleglise, Y. Esteve, C. Lailler, E. Gilmartin, N. Campbell, “Multifaceted engagement in social interaction with a machine: The JOKER Project,” in Proceedings of the 13th IEEE International Conference on Automatic Face and Gesture Recognition, FG 2018, China, June 2018.
  15. M. A. T. Turan and E. Erzin, “Empirical Mode Decomposition of Throat Microphone Recordings for Intake Classification,” in the Second International Workshop on Multimedia for Personal Health and Health Care, Mountain View, CA, USA, October 23, 2017.
  16. S. N. Fatima and E. Erzin, ”Acoustic Feature Selection for Natural Dialog Identification,” in the 3rd International Workshop on Affective Social Multimedia Computing, Stockholm, Sweden, 2017.
  17. S. A. Abadi and E. Erzin, ”Vocal Tract Airway Tissue Boundary Tracking for rtMRI using Shape and Appearance Priors,” in INTERSPEECH: Annual Conference of the International Speech Communication Association, Stockholm, Sweden, 2017.
  18. B. B. Turker, Z. Bucinca, E. Erzin, Y. Yemez, and M. Sezgin, ”Analysis of Engagement and User Experience with a Laughter Responsive Social Robot,” in INTERSPEECH: Annual Conference of the International Speech Communication Association, Stockholm, Sweden, 2017.S.
  19. N. Fatima and E. Erzin, ”Cross-Subject Continuous Emotion Recognition using Speech and Body Motion in Dyadic Interactions,” in INTERSPEECH: Annual Conference of the International Speech Communication Association, Stockholm, Sweden, 2017.
  20. H. Ramezani, H. Khaki, E. Erzin, O.B. Akan, ”Speech Features for Telemonitoring of Parkinson’s Disease Symptoms,” in the 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Jeju Island, Korea, 2017.
  21. H. Khaki and E. Erzin, “Use of Affect Based Interaction Classification for Continuous Emotion Tracking,” in IEEE International Conference on Acoustics, Speech and Signal Processing, New Orleans, USA, 2017.
  22. R. Sadiq and E. Erzin, “Affect Recognition from Lip Articulations,” in IEEE International Conference on Acoustics, Speech and Signal Processing, New Orleans, USA, 2017.
  23. H. Khaki, E. Erzin, ”Use of Agreement/Disagreement Classification in Dyadic Interactions for Continuous Emotion Recognition,” in INTERSPEECH: Annual Conference of the International Speech Communication Association, San Francisco, USA, 2016.
  24. H. Khaki, E. Bozkurt, E. Erzin, ”Agreement and disagreement classification of dyadic interactions using vocal and gestural cues,” in IEEE International Conference on Acoustics, Speech and Signal Processing, Shanghai, China, 2016.
  25. J. Abel, M. Kaniewska, C. Guillaume, W. Tirry, H. Pulakka, V. Myllyl, J. Sj, P. Alku, I. Katsir, D. Malah, I. Cohen, M.A.T. Turan, E. Erzin, T. Schlien, P. Vary, A. Nour-Eldin, P. Kabal, T. Fingscheidt, ”A subjective listening test of six different artificial bandwidth extension approaches in English, Chinese, German, and Korean,” in IEEE International Conference on Acoustics, Speech and Signal Processing, Shanghai, China, 2016.
  26. M. A. T. Turan and E. Erzin, Synchronous Overlap and Add of Spectra for Enhancement of Excitation in Artificial Bandwidth Extension of Speech, in INTERSPEECH: Annual Conference of the International Speech Communication Association, Dresden, Germany, 2015.
  27. M. A. T. Turan and E. Erzin, Synchronous Overlap and Add of Spectra for Enhancement of Excitation in Artificial Bandwidth Extension of Speech, in INTERSPEECH: Annual Conference of the International Speech Communication Association, Dresden, Germany, 2015.
  28. H. Khaki and E. Erzin, Continuous Emotion Tracking using Total Variability Space, in INTERSPEECH: Annual Conference of the International Speech Communication Association, Dresden, Germany, 2015.
  29. E. Bozkurt, E. Erzin, Y. Yemez, Affect-Expressive Hand Gestures Synthesis and Animation, in IEEE International Conference on Multimedia and Expo (ICME), Torino, Italy, 2015.
  30. Z. Yang, A. Metallinou, E. Erzin, S. Narayanan, ”Analysis of interaction attitudes using data-driven hand gesture phrases,” in IEEE International Conference on Acoustics, Speech and Signal Processing, Florence, Italy, 2014.
  31. M. A. T. Turan and E. Erzin, ”A New Statistical Excitation Mapping for Enhancement of Throat Microphone Recordings,” in Interspeech’2013 Conference, Lyon, France.
  32. E. Bozkurt, S. Asta, S. Ozkul, Y. Yemez, and E. Erzin, ”Multimodal Analysis of Speech Prosody and Upper Body Gestures using Hidden Semi-Markov Models,” in IEEE International Conference on Acoustics, Speech and Signal Processing, (Vancouver), 2013.
  33. M. A. T. Turan and E. Erzin, ”Enhancement of Throat Microphone Recordings by Learning Phone- Dependent Mappings of Speech Spectra,” in IEEE International Conference on Acoustics, Speech and Signal Processing, (Vancouver), 2013.
  34. S. Ozkul, E. Bozkurt, S. Asta, Y. Yemez and E. Erzin, ”Multimodal Analysis of Upper-Body Gestures, Facial Expressions, and Speech,” 4th International Workshop on Corpora for Research on Emotion Sentiment & Social Signals, May 2012, Istanbul, Turkey.
  35. E. Bozkurt, E. Erzin, C. Eroglu Erdem and T. Erdem, ”RANSAC-based Training Data Selection for Speaker State Recognition,” Interspeech’2011, Florence, Italy.
  36. C. Yagli and E. Erzin, ”Artificial Bandwidth Extension of Spectral Envelope with Temporal Clustering,” ICASSP’11 Prague, Czech Republic.
  37. F. Ofli, E. Erzin, Y. Yemez, A.M. Tekalp, ”Multi-modal analysis of dance performances for music- driven choreography synthesis,” ICASSP’10 Dallas, USA.
  38. E. Bozkurt, E. Erzin, C. Eroglu Erdem and T. Erdem, ”Use of Line Spectral Frequencies for Emotion Recognition from Speech,” ICPR’2010, Istanbul, Turkey.
  39. C. Eroglu Erdem, E. Bozkurt, E. Erzin, and T. Erdem, ”RANSAC-based Training Data Selection for Emotion Recognition from Spontaneous Speech,” AFFINE’10, Firenze, Italy.
  40. E. Bozkurt, E. Erzin, C. Eroglu Erdem and T. Erdem, ”Improving Automatic Emotion Recognition from Speech Signals,” INTERSPEECH 2009 Emotion Challenge, Sept. 2009.
  41. E. Ozturk and E. Erzin, “Driving Status Identification Under Different Distraction Conditions from Driving Behaviour Signals,” 4th Biennial Workshop on DSP for In-Vehicle Systems and Safety, 25-27 June 2009, UTD, TX, USA.
  42. H. Abut, H. Erdogan, E. Erzin, and E. Cokelek “Sample Statistics from Collaborative Naturalistic Vehicular Corpus,” 4th Biennial Workshop on DSP for In-Vehicle Systems and Safety, 25-27 June 2009, UTD, TX, USA.
  43. F. Ofli, E. Erzin, Y. Yemez, A. M. Tekalp, A. T. Erdem, C. Erdem, T. Abaci, M. Ozkan ”Unsupervised Dance Figure Analysis from Video for Dancing Avatar Animation,” ICIP’08 San Diego, USA.
  44. Y. Demir, E. Erzin, Y. Yemez, and A. M. Tekalp, Evaluation of Audio Features for Audio-Visual Analysis of Dance Figures, EUSIPCO’08, Lausanne, Switzerland.
  45. E. Bozkurt, C. Eroglu Erdem, E. Erzin, T. Erdem, M. Ozkan and A.M. Tekalp, ”Speech-Driven Automatic Facial Expression Synthesis,” 3DTV Conference, Istanbul, 28-30 May 2008.
  46. F. Ofli, C. Canton-Ferrer, J. Tilmanne, Y. Demir, E. Bozkurt, Y. Yemez, E. Erzin, A.M. Tekalp, ”Audio-driven human body motion analysis and synthesis,” ICASSP’08 Las Vegas, USA.
  47. F. Ofli, E. Erzin, Y. Yemez, A.M. Tekalp, ”Estimation and Analysis of Facial Animation Parameter Patterns,” IEEE Int. Conf. on Image Processing, ICIP-2007.
  48. F. Ofli, Y. Demir, E. Erzin, Y. Yemez, and A. M. Tekalp, ”Multicamera Audio-Visual Analysis of Dance Figures,” IEEE Int. Conf. on Multimedia Expo, ICME-2007.
  49. E. Bozkurt, C. Eroglu Erdem, E. Erzin, T. Erdem and M. Ozkan, ”Comparison of Phoneme and Viseme Based Acoustic Units for Speech Driven Realistic Lip Animation,” 3DTV-Conference, 7-9 May 2007, Kos Island, Greece.
  50. M. E. Sargin, E. Erzin, Y. Yemez, A. M. Tekalp, A. T. Erdem, C. Erdem, M. Ozkan ”Prosody- Driven Head-Gesture Animation,” ICASSP-2007.
  51. M.E. Sargin, O. Aran, A. Karpov, F. Ofli, Y. Yasinnik, S. Wilson, E. Erzin, Y. Yemez and A.M. Tekalp ”Combined Gesture-Speech Analysis and Speech Driven Gesture Synthesis,” IEEE Int. Conf. on Multimedia Expo, Toronto, July 2006.
  52. M.E. Sargin, E. Erzin, Y. Yemez, and A.M. Tekalp, ”Multimodal Speaker Identification using Canonical Correlation Analysis, ICASSP-2006, vol. I, pp. 613-616, May, Toulouse, France.
  53. U. Bagci, E. Erzin ”Boosting Classifiers for Music Genre Classification,” 20th International Symposium on Computer and Information Sciences (ISCIS 2005), Istanbul, Turkey. Also to appear in Lecture Notes in Computer Science (LNCS) by Springer-Verlag.
  54. M.E. Sargin, E. Erzin, Y. Yemez, A.M. Tekalp ”Lip Feature Extraction based on Audio-Visual Correlation,” EUSIPCO 2005, Antalya, Turkey.
  55. H.E. Cetingul, E. Erzin, Y. Yemez, A. M. Tekalp ”Use of Lip Information for Robust Speaker Identification and Speech Recognition,” Biennial on DSP for In-Vehicle and Mobile Systems, Sesimbra, Portugal, September 2-3, 2005.
  56. H.E. Cetingul, Y. Yemez, E. Erzin, A. M. Tekalp ”Robust Lip-Motion Features For Speaker Identification,” IEEE Int. Conf. on Acoustic, Speech and Signal Processing, Philadelphia, March 2005.
  57. E. Erzin, Y. Yemez, A. M. Tekalp “Adaptive Classifier Cascade for Multimodal Speaker Identification,” INTERSPEECH 2004 – ICSLP, Korea, 2004.
  58. H.E. Cetingul, E. Erzin, Y. Yemez, A. M. Tekalp “On Optimal Selection of Lip-Motion Features for Speaker Identification,” IEEE Multimedia Signal Processing Workshop, Siena, Italy, 2004.
  59. H.E. Cetingul, Y. Yemez, E. Erzin, A.M. Tekalp “Discriminative Lip-Motion Features for Biometric Speaker Identification,” IEEE Int. Conf. on Image Processing, Singapore, 2004.
  60. E. Erzin, Y. Yemez, A. M. Tekalp, “Multimodal Audio-Visual Speaker Identification,” Special Workshop in MAUI (SWIM): Lectures by Masters in Speech Processing, Maui, Hawaii, January 12-14, 2004.
  61. A. Kanak, E. Erzin, Y. Yemez, and A. M. Tekalp, ”Joint Audio-Video Processing for Robust Biometric Speaker Identification in Car,” Workshop on DSP in Mobile and Vehicular Systems, Nagoya, Japan, April 3-4, 2003.
  62. A. Kanak, E. Erzin, Y. Yemez, and A. M. Tekalp, ”Speaker Identification Using Multimodal Audio-Video Processing,” IEEE Int. Conf. on Image Processing, Barcelona, Spain, September 2003.
  63. A. Kanak, E. Erzin, Y. Yemez, and A. M. Tekalp, ”Joint Audio-Video Processing for Biometric Speaker Identification,” IEEE Int. Conf. on Acoustic, Speech and Signal Processing, Hong Kong, China, April 2003.
  64. E. Erzin, “Shaped Fixed Codebook Search for CELP Coding at Low Bit Rates,” Int. Conf. on Acoustic, Speech, and Signal Processing 2000 (ICASSP 2000), June 2000.
  65. E. Erzin, A. Kumar, A. Gersho, “Natural Quality Variable-Rate Spectral Speech Coding below 3.0 kbps,” Int. Conf. on Acoustic, Speech, and Signal Processing 1997 (ICASSP ’97), Vol. II, pp. 1579-82, April 1997.
  66. O. Arıkan, M. Belge, A.E. C ̧etin, E. Erzin, ”Adaptive Filtering Approaches for Non-Gaussian Processes,” Int. Conf. on Acoustic, Speech, and Signal Processing 1995 (ICASSP ’95), May 1995.
  67. E. Erzin, A. E. C ̧etin, Y. Yardımcı, “Subband Analysis for Robust Speech Recognition in the Presence of Car Noise,” ”Int. Conf. on Acoustic, Speech, and Signal Processing 1995 (ICASSP ’95), May 1995.
  68. E. Erzin and A.E. C ̧etin, “Interframe Differential Vector Coding of Line Spectrum Frequencies,” Proc. of the Int. Conf. on Acoustic, Speech, and Signal Processing 1993 (ICASSP ’93), Vol. II, pp. 25-28, April 1993.
  69. E. Erzin and A.E. C ̧etin, “On the Use of Interframe Information of Line Spectral Frequencies in Speech Coding,” NATO-ASI, New Advances, and Trends in Speech Recognition and Coding, Bubion (Granada), June-July 1993.
  70. B. Tuzun, E. Erzin, M. Demirekler, T. Memisoglu, S. Ugur, and A.E. Çetin, “A Speaker Independent Isolated Word Recognition System for Turkish,” NATO-ASI, New Advances and Trends in Speech Recognition and Coding, Bubion (Granada), June-July 1993.
  71. E. Gunduzhan, E. Erzin, and A.E. Cetin, “Zero-Crossing Based Speech Vocoder at Low Bit Rates,” was presented in the 4th International Conference on Advances in Communication and Control, Rhodes, Greece, June 1993.
  72. M. Karaman, C. Aydın, E. Kolagasıoglu, M. Toygar, A. Baktır, E. Erzin, R. Tahboub, F. Kılıc, M. Asyalı, and A. Atalar, ”Design of an Integrated Circuit for Ultrasonic Imaging”, presented in BILKON’91, Ankara, TURKEY, 1991.

BOOK CHAPTERS

  1. Elif Bozkurt, Engin Erzin, Cigdem Erdem Eroglu, A. Tanju Erdem, ”RANSAC-Based Training Data Selection on Spectral Features for Emotion Recognition from Spontaneous Speech,” in Analysis of Verbal and Nonverbal Communication and Enactment: The Processing Issues, A. Esposito, A. Vinciarelli, K. Vicsi, C. Pelachaud, A. Nijholt (Editors), Springer Verlag, LNCS: 6800, pp. 36-47, 2011.
  2. H. Ertan Cetingul, E. Erzin, Y. Yemez, and A. M. Tekalp, ”Multimodal Speaker Identification using Discriminative Lip Motion Features,” in Visual Speech Recognition: Lip Segmentation and Mapping, A. Liew, S. Wang (Editors), ISBN: 978-1-60566-186-5, IGI Global, 2009.
  3. U.C. Akargun, E. Erzin, ”Estimation of Acoustic Microphone Vocal Track Parameters from Throat Microphone Recordings,” in In-Vehicle Corpus and Signal Processing for Driver Behavior, K. Takeda, H. Erdogan, J.H.L. Hansen, and H. Abut (Editors), Springer Verlag, 2008.
  4. E. Erzin, Y. Yemez, A. M. Tekalp, ”DSP for In-Vehicle and Mobile Systems, Chapter: 16 Joint Audio-Video Processing for Robust Biometric Speaker Identification in Car,” Kluwer Academic Publishers, 2005.
  5. E. Erzin, A. E. Cetin, ”Speech Recognition and Coding, Chapter: On the Use of Interframe Information of Line Spectral Frequencies in Speech Coding,” Volume 147 of the series NATO ASI Series, pp. 431-434, Springer Berlin Heidelberg, 1995.

CONFERENCE PUBLICATIONS IN TURKISH

  1. A. Emirdagi, F. Tokmak, N. Köprücü, K. Akar, E. Erzin and A. Vural, “Detection of Stride Time and Stance Phase Ratio from Accelerometer Data for Gait Analysis/ Yürüme Analizi için İvmeölçer Verilerinden Uzun Adım Süresi ve Basma Fazı Oranı Tespiti,” 30. Sinyal İşleme ve İletişim Uygulamaları Kurultayı (SIU), 2022.
  2. E. Kesim, T. Numanoğlu, B. B. Türker,  E. Erzin, Y. Yemez, M. T. Sezgin, “İnsan-Robot Etkileşiminde Gerçek Zamanlı İlgi Seviyesi Ölçümü,” 28. Sinyal İşleme ve İletişim Uygulamaları Kurultayı (SIU), 2020.
  3. T. Numanoğlu, M. T. Sezgin, Y. Yemez, E. Erzin, “İkili İletişimde Kafa Sallama Tespiti,” 27. Sinyal İşleme ve İletişim Uygulamaları Kurultayı (SIU), Sivas, 2019.
  4. U. Fidan, D. Tomar, P. G. Özdil, E. Erzin, “Konuşmadan Duygu Durum Kestirimi ve Etiketlenmesi için Yeni Bir Arayüz,”  27. Sinyal İşleme ve İletişim Uygulamaları Kurultayı (SIU), Sivas, 2019.
  5. M. A. T. Turan, E. Erzin, “Otokodlayıcı Tabanlı Derin Sinir Ağları Kullanarak Gıda Tüketiminin Tespit Edilmesi,” 26. Sinyal İşleme ve İletişim Uygulamaları Kurultayı (SIU), Çesme, İzmir, 2018.
  6. B. B. Turker, M. T. Sezgin, Y. Yemez, E. Erzin, “İkili İletişimde Olası Kafa Sallama Anlarının Çok Kipli Kestirimi,” 26. Sinyal İşleme ve İletişim Uygulamaları Kurultayı (SIU), Çesme, İzmir, 2018.
  7. B. B. Turker, Z. Buçinca, M. T. Sezgin, Y. Yemez, E. Erzin, “Çok Kipli ve Gerçek Zamanlı Gülme Sezimi,” 25. Sinyal İşleme ve İletişim Uygulamaları Kurultayı (SIU), Antalya, 2017.
  8. M. A. T. Turan, E. Erzin, “Yeme-İçme Seslerinin Hilbert-Huang Dönüşümü ile Sınıflandırılması,” 25. Sinyal İşleme ve İletişim Uygulamaları Kurultayı (SIU), Antalya, 2017.
  9. M. A. T. Turan and E. Erzin, Gırtlak Mikrofonu ile Gıda Alımlarının Sınıflandırılması, SIU: Sinyal İşleme ve İletişim Uygulamaları Kurultayı, Zonguldak, 2016.
  10. E. Bozkurt, H. Khaki, S. Keçeci, B. B. Türker, Y. Yemez, and E. Erzin, JESTKOD Veritabanının Duygu Durum Etiketlemerini Kullanarak Analizi, SIU: Sinyal İşleme ve İletişim Uygulamaları Kurultayı, Zonguldak, 2016.
  11. K. Kaşarcı, E. Bozkurt, Y. Yemez, and E. Erzin, Gerçek Zamanlı Konuşma Sürümlü Jest Animasyonu, SIU: Sinyal İşleme ve İletişim Uygulamaları Kurultayı, Zonguldak, 2016.
  12. M. A. T. Turan and E. Erzin, Konuşma Kaynağı için Yapay Bant Genişletme, SIU: Sinyal İşleme ve İletişim Uygulamaları Kurultayı, Malatya, 2015.
  13. E. Bozkurt, H. Khaki, S. Keçeci, B. B. Türker, Y. Yemez, and E. Erzin, JESTKOD Veritabanı: İkili İletişim Analizi, SIU: Sinyal İşleme ve İletişim Uygulamaları Kurultayı, Malatya, 2015.
  14. B. B. Türker, S. Marzban, M. T. Sezgin, Y. Yemez, and E. Erzin, Çok-Kipli Modelleme ile Duygusal Patlama Sezimi, SIU: Sinyal İşleme ve İletişim Uygulamaları Kurultayı, Malatya, 2015.
  15. M. A. T. Turan, E. Erzin, ”Fonetik Sınıflandırma ile Gırtlak Mikrofonu İyileştirmesi,” SIU’2014, Trabzon, Turkey.
  16. B.B. Türker, S. Marzban, E. Erzin, Y. Yemez, T.M. Sezgin, ”Çok-kipli ipucları kullanarak duygusal patlama tanıma,” SIU’2014, Trabzon, Turkey.
  17. E. Bozkurt, E. Erzin, Y. Yemez, ”Konuşma Ritmi Sürümlü̈ Jest Animasyonu,” SIU’2013, Girne, KKTC.
  18. E. Bozkurt, E. Erzin, ”Konuşmadan Duygu Tanıma Üzerine Değerlendirmeler,” SIU’2012, Fethiye, Turkey.
  19. F. Ofli, E. Erzin, Y. Yemez, A.M. Tekalp, ”Çok Kipli Koreografi Modeli (Multimodal Dance Choreography Model),” SIU’11 Antalya, Turkey.
  20. E. Bozkurt, E. Erzin, C. Eroglu Erdem, T. Erdem, ”INTERSPEECH 2009 Emotion Recognition Challenge Evaluation,” SIU’2010, Diyarbakır, Turkey.
  21. Elif Bozkurt, Engin Erzin, Cigdem Eroglu Erdem, A. Tanju Erdem, “Yüz İfadesi Canlandırma için Konuşma Sinyalinden Otomatik Duygu Tanıma (Automatic Emotion Recognition for Facial Expression Animation from Speech) SIU’09 Antalya, Turkey.
  22. F. Ofli, Y. Demir, C. Canton-Ferrer, J. Tilmanne, K. Balci, E. Bozkurt, I. Kiziloglu, Y. Yemez, E. Erzin, A.M. Tekalp, L. Akarun, A.T. Erdem, ”Çok Bakışlı İşitsel-Görsel Dans Verilerinin Analizi ve Sentezi (Analysis and Synthesis of Multiview Audio-Visual Dance Figures),” SIU’08 Didim, Turkey.
  23. A.O. Abalı, E. Erzin, A. Gürsoy, ”Saklı Markov Modelleri Aracılığı ile Gen Düzenlenmelerinin Mikrodizi Verilerinden Öğrenilmesi,” SIU 2007, Eskişehir.
  24. U.C.Akargun,E.Erzin,”Akustik Mikrofon Ses Yolu Parametrelerinin Gırtlak Mikrofon Kayıtlarından Kestirimi,” SIU 2007, Eskişehir.
  25. E. Bozkurt, Ç. Eroglu Erdem, E. Erzin, T. Erdem, M. Ozkan, ”Gerçekci Dudak Animasyonu için Fonem ve Vizeme Dayalı Akustik Birimlerin Karşılaştırılması,” SIU 2007, Eskişehir.
  26. M.E. Sargın, E. Erzin, Y. Yemez, A.M. Tekalp, A.T. Erdem ”Konuşma ile Sürülen Kafa Jesti Analizi ve Sentezi,” SIU 2006, Antalya.
  27. U. Bagcı, E. Erzin ”Mü̈zik Türlerinin Sınıflanmasında Sınıflandırıcıların Yükseltilmesi,” SIU 2006, Antalya
  28. Bagcı, E. Erzin ”Müzik Türlerinin Sınıflanmasında Benzer Kesişim Bilgileri Uygulamaları,” SIU 2006, Antalya
  29. O.N. Yogurtçu, E. Erzin ”Nicemlenmiş Yansıma Katsayılarıyla Müzikal Enstrüman Tanıma,” SIU 2005, Kayseri.
  30. H.E. Cetingul, Y. Yemez, E. Erzin, A.M. Tekalp ”Konuşmacı Tanıma için Karşılaştırmalı Dudak Devinim Analizi,” SIU 2005, Kayseri.
  31. M.E. Sargın, Y. Yemez, E. Erzin, A.M. Tekalp, ”Görsel-İşitsel İlintiye Dayalı Dudak Öznitelik Çıkarımı,” SIU 2005, Kayseri.
  32. H. E. Cetingül, Y. Yemez, E. Erzin, A. M. Tekalp, “Biyometrik Konuşmacı Tanıma için Dudak Devinimi Kullanımı,” SIU 2004, Kuşadası.
  33. E. Akyol, E. Erzin, A. M. Tekalp, “Uyarlanabilir Gürültü̈ Temizleme ile Dayanıklı Ses Tanıma,”  SIU 2004, Kuşadası.
  34. E. Erzin, “Konuşma ve Ses Sıkıştırma,” Türkiye Bilişim Ansiklopedisi, davetli yazı.
  35. A. Kanak, E. Erzin, Y. Yemez ve A. M. Tekalp, ”Biyometrik konuşmacı tanıma için birleşik ses-görüntü işleme,” SIU 2003, Koç Üniversitesi, İstanbul.
  36. E.Erzin,“Öngörülü dönüşüm ile geniş bantlı konuşmanın 13kb/s hızında kodlanması,” SIU2002, Pamukkale, Denizli.
  37. E. Erzin, A. E. Cetin ve Y. Yardımcı, “Araç içi gürbüz ses tanıma için alt bant analizi,” SIU 1995, Kapadokya, Nevşehir.
  38. M. Demirekler, A. E. Cetin, B. Nakiboglu, E. Erzin, D. Cetin ve F. Yıldırım, “Ayrık sözcük tanıma,” Konuşma İşleme Ç̧alıştayı, sf. 48-52, METU, Ankara, 1995.
  39. M. Demirekler, A. E. Cetin, B. Nakiboglu, E. Erzin, D. Cetin ve F. Yıldırım, “Birleşik rakam tanıma,” Konuşma İşleme Çalıştayı, sf. 53-58, METU, Ankara, 1995.
  40. M. Demirekler, A. E. Çetin, B. Nakiboglu, E. Erzin, D. Cetin ve F. Yıldırım, “Konuşma Analiz Sistemi (KASIS),” Konuşma İşleme Çalıştayı, sf. 59-62, METU, Ankara, 1995.
  41. M. Demirekler, A. E. Çetin, B. Nakiboglu, E. Erzin, D. Cetin ve F. Yıldırım, “Metinden bağımsız konuşmacı tanıma,” Konuşma İşleme Çalıştayı, sf. 63-65, METU, Ankara, 1995.
  42. E. Erzin ve A. E. Cetin, “Çizgisel spektrum frekanslarına dayalı sözcük sınırları belirleme yöntemi,” SIU 1994, sf. 288-292, Gökova, Muğla.
  43. A. E. Cetin, O. Arıkan ve E. Erzin, “Gauss olmayan kararlı süreçler iç̧in uyarlanır süzgeçleme,” SIU 1994, sf. 370-372, Gökova, Muğla.
  44. E. Erzin ve A. E. Cetin, “Çizgisel spektrum frekansları için çerçeveler arası fark vektörü kodlama yöntemi,” SIU 1993, sf. 25-29, İstanbul.

PATENTS

  1. E. Erzin and M. C. Recchione, “Shaped Fixed Codebook Search for CELP Speech Coding”, US Patent No. 6,449,313, issued September 10, 2002.