Department of Electrical and Computer Engineering

MENUMENU
  • Department
      • Profile
      • Faculty
      • Evaluation
      • Administration
      • Staff
  • Studies
    • Subject Areas
    • Undergraduate Studies
    • Postgraduate Studies
      • MSc Studies in “Science and Technology of ECE”
      • MSc Studies in “Smart Grid Energy Systems”
      • MSc Studies in “Applied Informatics”
    • PhD Studies
    • Course List
      • Undergraduate Courses
      • Postgraduate Courses
        • Science and Technology of ECE
        • Smart Grid Energy Systems
        • Applied Informatics
      • Erasmus
    • ECTS
    • Career Opportunities
    • Practise Training
  • Research
    • Labs
    • Research Projects
    • Postdoc Research
    • Ph.D. Candidates
    • Theses – Technical Reports
    • Research Projects Form
    • Active Research Projects

      MORCIC: Model Order Reduction of Electromagnetic Models for Large Integrated Circuits

      Scientific Responsible

      Nestor EvmorfopoulosNestor Evmorfopoulos, Associate Professor
      E-mail: nestevmo@e-ce.uth.gr

      Title MORCIC: Model Order Reduction of Electromagnetic Models for Large Integrated Circuits
      Duration 2021 – 2023

      Read More

      SL-ReDu: Sign Language Recognition in Education

      Scientific Responsible

      Gerasimos PotamianosGerasimos Potamianos, Associate Professor
      E-mail: gpotamianos@e-ce.uth.gr

      Title SL-ReDu: Sign Language Recognition in Education
      Duration 2020 – 2023
      Site https://sl-redu.e-ce.uth.gr/

      Read More

      Characterisation of LN2 UUI

      Scientific Responsible

      Christos SotiriouChristos Sotiriou, Associate Professor
      E-mail: chsotiriou@e-ce.uth.gr

      Title Characterisation of LN2 UUI
      Duration 2019 – 2023
      Site https://caslab.e-ce.uth.gr/

      Read More

      Qualcomm Faculty R&D Award 2019

      Scientific Responsible

      Christos SotiriouChristos Sotiriou, Associate Professor
      E-mail: chsotiriou@e-ce.uth.gr

      Title Qualcomm Faculty R&D Award 2019
      Duration 2019 – 2023
      Site https://caslab.e-ce.uth.gr/

      Read More

  • Alumni
    • Ph.D. Graduates
  • Service Offices
    • Secretariat
    • Technical support
  • Announcements
    • General Announcements
    • Academic News
  • Contact
    • Department of Electrical and Computer Engineering
      • Sekeri – Cheiden Str
        Pedion Areos, ECE Building
        383 34 Volos – Greece
      Tel. +30 24210 74967
      e-mail gece ΑΤ e-ce.uth.gr
      PGS Tel. +30 24210 74934
      PGS e-mail pgsec ΑΤ e-ce.uth.gr
      URL https://www.e-ce.uth.gr/contact-info/?lang=en
  • Login

Gerasimos Potamianos

Home » Department » Faculty » Gerasimos Potamianos
Potamianos Gerasimos
Gerasimos Potamianos, Associate Professor
DiplomaECE, National Technical University of Athens (1988)
Ph.D.ECE, The Johns Hopkins University (1994)
Office420
Office HoursWednesday 13:00 - 14:00 (or by appointment)
Tel.+30 24210 74928
Emailgpotamianos@e-ce.uth.gr
Web Page   CV   Google Scholar

  • Publications
  • Research
  • Courses
  • PhD Students

[Last update: 09/05/2021]

2021

  • P. Giannoulis, G. Potamianos, and P. Maragos, “Overlapped sound event classification via multi-channel sound separation network,” [To Appear in:] Proc. Europ. Conf. Signal Process. (EUSIPCO), Dublin, Ireland, 2021.
  • A. Koumparoulis, G. Potamianos, S. Thomas, and E. da Silva Morais, “Resource-efficient TDNN architectures for audio-visual speech recognition,” [To Appear in:] Proc. Europ. Conf. Signal Process. (EUSIPCO), Dublin, Ireland, 2021.
  • P. P. Filntisis, N. Efthymiou, G. Potamianos, and P. Maragos, “An audiovisual child emotion recognition system for child-robot interaction applications,” [To Appear in:] Proc. Europ. Conf. Signal Process. (EUSIPCO), Dublin, Ireland, 2021.
  • N. Efthymiou, P. P. Filntisis, G. Potamianos, and P. Maragos, “A robotic edutainment framework for designing child-robot interaction scenarios,” [To Appear in:] Proc. ACM Int. Conf. PErvarsive Technologies Related to Assistive Environments (PETRA), Corfu, Greece, 2021.
  • K. Papadimitriou, M. Parelli, G. Sapountzaki, G. Pavlakos, P. Maragos, and G. Potamianos, “Multimodal fusion and sequence learning for cued speech recognition from videos,” [To Appear in:] Proc. Int. Conf. Universal Access in Human-Computer Interaction (UAHCI), Washington, D.C., 2021.
  • E. Efthimiou, S.-E. Fotinea, C. Flouda, T. Goulas, G. Ametoglou, G. Sapountzaki, K. Papadimitriou, and G. Potamianos, “The SL-ReDu environment for self-monitoring and objective learner assessment in Greek sign language,” [To Appear in:] Proc. Int. Conf. Universal Access in Human-Computer Interaction (UAHCI), Washington, D.C., 2021.
  • G. Sapountzaki, E. Efthimiou, S.-E. Fotinea, K. Papadimitriou, and G. Potamianos, “Educational material organization in a platform for Greek sign language self monitoring and assessment,” [To Appear in:] Proc. Ann. Int. Conf. Education and New Learning Technologies (EDULEARN), Palma, Spain, 2021.

2020

  • S. Thermos, G. T. Papadopoulos, P. Darras, and G. Potamianos, “Deep sensorimotor learning for RGB-D object recognition,” Computer Vision and Image Understanding, vol. 190, 2020.
  • A. Koumparoulis, G. Potamianos, S. Thomas, and E. da Silva Morais, “Resource-adaptive deep learning for visual speech recognition,” Proc. Conf. Int. Speech Comm. Assoc. (Interspeech), pp. 3510–3514, Shanghai, China, 2020.
  • K. Papadimitriou and G. Potamianos, “Multimodal sign language recognition via temporal deformable convolutional sequence learning,” Proc. Conf. Int. Speech Comm. Assoc. (Interspeech), pp. 2752–2756, Shanghai, China, 2020.
  • K. Papadimitriou and G. Potamianos, “A fully convolutional sequence learning approach for cued speech recognition from videos,” Proc. Europ. Conf. Signal Process. (EUSIPCO), pp. 326–330, Amsterdam, The Netherlands, 2020.
  • M. Parelli, K. Papadimitriou, G. Potamianos, G. Pavlakos, and P. Maragos, “Exploiting 3D hand pose estimation in deep learning-based sign language recognition from RGB videos,” Computer Vision – ECCV 2020 Workshops Proceedings, Part II, A. Bartoli and A. Fusiello (Eds.), pp. 249–263, LNCS/LNIP vol. 12536, 2020.
  • P. P. Filntisis, N. Efthymiou, G. Potamianos, and P. Maragos, “Emotion understanding in videos through body, context, and visual-semantic embedding loss,” Computer Vision – ECCV 2020 Workshops Proceedings, Part I, A. Bartoli and A. Fusiello (Eds.), pp. 747–755, LNCS/LNIP vol. 12535, Springer, 2020.
  • G. Potamianos, K. Papadimitriou, E. Efthimiou, S.-E. Fotinea, G. Sapountzaki, and P. Maragos, “SL-ReDu: Greek sign language recognition for educational applications. Project description and early results,” Proc. ACM Int. Conf. PErvarsive Technologies Related to Assistive Environments (PETRA), article no. 59, pp. 1–6, Corfu, Greece, 2020.
  • A. Koumparoulis, G. Potamianos, S. Thomas, and E. Morais, “Audio-assisted image inpainting for talking faces,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 7664–7668, Barcelona, Spain, 2020.
  • S. Thermos, P. Darras, and G. Potamianos, “A deep learning approach to object affordance segmentation,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 2358–2362, Barcelona, Spain, 2020.

2019

  • S. Thermos, G. T. Papadopoulos, P. Darras, and G. Potamianos, “Deep sensorimotor learning for RGB-D object recognition,” Computer Vision and Image Understanding, vol. 190, 2020.
  • P. Giannoulis, G. Potamianos, and P. Maragos, “Room-localized speech activity detection in multi-microphone smart homes,” EURASIP Journal on Audio, Speech, and Music Processing, vol. 2019, no. 15, pp. 1–23, 2019.
  • P. P. Filntisis, N. Efthymiou, P. Koutras, G. Potamianos, and P. Maragos, “Fusing body posture with facial expressions for joint recognition of affect in child-robot interaction,” IEEE Robotics and Automation Letters, vol. 4, no. 4, pp. 4011–4018, 2019.
  • S. Oviatt, B. Schuller, P. Cohen, D. Sonntag, G. Potamianos, and A. Kruger (Eds.), The Handbook of Multimodal-Multisensor Interfaces, Volume 3: Language Processing, Software and Commercialization, Emerging Directions, ACM Books / Morgan-Claypool Publishers, San Rafael, CA, 2019.
  • S. P. Chytas and G. Potamianos, “Hierarchical detection of sound events and their localization using convolutional neural networks with adaptive thresholds,” Proc. Detection and Classification of Acoustic Scenes and Events Workshop (DCASE), pp. 50–54, New York, NY, 2019.
  • A. Koumparoulis and G. Potamianos, “MobiLipNet: Resource-efficient deep learning based lipreading,” Proc. Conf. Int. Speech Comm. Assoc. (Interspeech), pp. 2763–2767, Graz, Austria, 2019.
  • K. Papadimitriou and G. Potamianos, “End-to-end convolutional sequence learning for ASL fingerspelling recognition,” Proc. Conf. Int. Speech Comm. Assoc. (Interspeech), pp. 2315–2319, Graz, Austria, 2019.
  • K. Papadimitriou and G. Potamianos, “Fingerspelled alphabet sign recognition in upper-body videos,” Proc. Europ. Conf. Signal Process. (EUSIPCO), La Coruna, Spain, 2019.

2018

  • S. Oviatt, B. Schuller, P. Cohen, D. Sonntag, G. Potamianos, and A. Kruger (Eds.), The Handbook of Multimodal-Multisensor Interfaces, Volume 2: Signal Processing, Architectures, and Detection of Emotion and Cognition, ACM Books / Morgan-Claypool Publishers, San Rafael, CA, 2018.
  • A. Koumparoulis and G. Potamianos, “Deep view2view mapping for view-invariant lipreading,” Proc. IEEE Spoken Language Technology Works. (SLT), pp. 588–594, Athens, Greece, 2018.
  • K. Papadimitriou and G. Potamianos, “A hybrid approach to hand detection and type classification in upper-body videos,” Proc. Europ. Work. Visual Information Process. (EUVIP), Tampere, Finland, 2018.
  • J. Hadfield, P. Koutras, N. Efthymiou, G. Potamianos, C. S. Tzafestas, and P. Maragos, “Object assembly guidance in child-robot interaction using RGB-D based 3D tracking,” Proc Int. Conf. Intell. Robots Systems (IROS), pp. 347–354, Madrid, Spain, 2018.
  • P. Giannoulis, G. Potamianos, and P. Maragos, “Multi-channel non-negative matrix factorization for overlapped acoustic event detection,” Proc. Europ. Conf. Signal Process. (EUSIPCO), pp. 857–861, Rome, Italy, 2018.
  • N. Efthymiou, P. Koutras, P. P. Filntisis, G. Potamianos, and P. Maragos, “Multi-view fusion for action recognition in child-robot interaction,” Proc. Int. Conf. Image Process. (ICIP), pp. 455–459, Athens, Greece, 2018.
  • S. Thermos, G. T. Papadopoulos, P. Darras, and G. Potamianos, “Attention-enhanced sensorimotor object recognition,” Proc. Int. Conf. Image Process. (ICIP), pp. 336–340, Athens, Greece, 2018.
  • A. Tsiami, P. Koutras, N. Efthymiou, P. P. Filntisis, G. Potamianos, and P. Maragos, “Multi3: Multi-sensory perception system for multi-modal child interaction with multiple robots,” Proc. IEEE Int. Conf. on Robotics and Automation (ICRA), Brisbane, Australia, 2018.
  • A. Tsiami, P. P. Filntisis, N. Efthymiou, P. Koutras, G. Potamianos, and P. Maragos, “Far-field audio-visual scene perception of multi-party human-robot interaction for children and adults,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 6568–6572, Calgary, Canada, 2018.

2017

  • I. Rodomagoulakis, A. Katsamanis, G. Potamianos, P. Giannoulis, A. Tsiami, and P. Maragos, “Room-localized spoken command recognition in multi-room, multi-microphone environments,” Computer Speech and Language, vol. 46, pp. 419–443, 2017.
  • G. Potamianos, E. Marcheret, Y. Mroueh, V. Goel, A. Koumbaroulis, A. Vartholomaios, and S. Thermos, “Audio and visual modality combination in speech processing applications”, The Handbook of Multimodal-Multisensor Interfaces, Volume 1: Foundations, User Modeling, and Common Modality Combinations, S. Oviatt, B. Schuller, P. Cohen, D. Sonntag, G. Potamianos, and A. Kruger (Eds.), Ch. 12, pp. 489–543, ACM Books / Morgan-Claypool Publishers, San Rafael, CA, 2017.
  • S. Oviatt, B. Schuller, P. Cohen, D. Sonntag, G. Potamianos, and A. Kruger (Eds.), The Handbook of Multimodal-Multisensor Interfaces, Volume 1: Foundations, User Modeling, and Common Modality Combinations, ACM Books / Morgan-Claypool Publishers, San Rafael, CA, 2017.
  • A. Koumparoulis, G. Potamianos, Y. Mroueh, and S. J. Rennie, “Exploring ROI size in deep learning based lipreading,” Proc. Int. Conf. on Auditory-Visual Speech Process. (AVSP), pp. 64–69, Stockholm, Sweden, 2017.
  • P. Giannoulis, G. Potamianos, and P. Maragos, “On the joint use of NMF and classification for overlapping acoustic event detection,” Proc. Int. Works. on Computational Intelligence for Multimedia Understanding (IWCIM), Held in Conjunction with Europ. Conf. Signal Process. (EUSIPCO), Kos Island, Greece, 2017.
  • S. Thermos, G. T. Papadopoulos, P. Darras, and G. Potamianos, “Deep affordance-grounded sensorimotor object recognition,” Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), pp. 49–57, Honolulu, HI, 2017.

2016

  • S. Thermos and G. Potamianos, “Audio-visual speech activity detection in a two-speaker scenario incorporating depth information from a profile or frontal view,” Proc. IEEE Spoken Language Technology Works. (SLT), pp. 579–584, San Diego, CA, 2016.
  • P. Giannoulis, G. Potamianos, P. Maragos, and A. Katsamanis, “Improved dictionary selection and detection schemes in sparse-CNMF-based overlapping acoustic event detection,” Proc. Detection and Classification of Acoustic Scenes and Events Works. (DCASE), pp. 25–29, Budapest, Hungary, 2016.
  • A. Tsiami, A. Katsamanis, I. Rodomagoulakis, G. Potamianos, and P. Maragos, “Sweet home listen: A distant speech recognition system for home automation control,” “Show and Tell” Session Demonstration, Int. Conf. Acoust. Speech Signal Process. (ICASSP), Shanghai, China, 2016.

2015

  • E. Marcheret, G. Potamianos, J. Vopicka, and V. Goel, “Scattering vs. discrete cosine transform features in visual speech processing,” Proc. Int. Joint Conf. on Facial Analysis, Animation and Auditory-Visual Speech Process. (FAAVSP), pp. 175–180, Vienna, Austria, 2015.
  • E. Marcheret, G. Potamianos, J. Vopicka, and V. Goel, “Detecting audio-visual synchrony using deep neural networks,” Proc. Conf. Int. Speech Comm. Assoc. (Interspeech), pp. 548–552, Dresden, Germany, 2015.
  • P. Giannoulis, A. Brutti, M. Matassoni, A. Abad, A. Katsamanis, M. Matos, G. Potamianos, and P. Maragos, “Multiroom speech activity detection using a distributed microphone network in domestic environments,” Proc. Europ. Conf. Signal Process. (EUSIPCO), pp. 1271–1275, Nice, France, 2015.
  • Z. I. Skordilis, A. Tsiami, P. Maragos, G. Potamianos, L. Spelgatti, and R. Sannino, “Multichannel speech enhancement using MEMS microphones,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 2729–2733, Brisbane, Australia, 2015.

2014

  • G. Floros, K. Kyritsis, and G. Potamianos, “Database and baseline system for detecting degraded traffic signs in urban environments,” Proc. Europ. Work. Visual Information Process. (EUVIP), Paris, France, 2014.
  • A. Tsiami, I. Rodomagoulakis, P. Giannoulis, A. Katsamanis, G. Potamianos, and P. Maragos, “ATHENA: A Greek multisensory database for home automation control,” Proc. Conf. Int. Speech Comm. Assoc. (Interspeech), pp. 1608–1612, Singapore, 2014.
  • P. Giannoulis, G. Potamianos, A. Katsamanis, and P. Maragos, “Multi-microphone fusion for detection of speech and acoustic events in smart spaces,” Proc. Europ. Conf. Signal Process. (EUSIPCO), pp. 2375–2379, Lisbon, Portugal, 2014.
  • A. Tsiami, A. Katsamanis, P. Maragos, and G. Potamianos, “Experiments in acoustic source localization using sparse arrays in adverse indoors environments, Proc. Europ. Conf. Signal Process. (EUSIPCO), pp. 2390–2394, Lisbon, Portugal, 2014.
  • P. Giannoulis, A. Tsiami, I. Rodomagoulakis, A. Katsamanis, G. Potamianos, and P. Maragos, “The Athena-RC system for speech activity detection and speaker localization in the DIRHA smart home,” Proc. Joint Work. Hands-Free Speech Communication and Microphone Arrays (HSCMA), pp. 167–171, Nancy, France, 2014.
  • A. Katsamanis, I. Rodomagoulakis, G. Potamianos, P. Maragos, and A. Tsiami, “Robust far-field spoken command recognition for home automation combining adaptation and multichannel processing,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 5547–5551, Florence, Italy, 2014.

2013

  • I. Rodomagoulakis, G. Potamianos, and P. Maragos, “Advances in large vocabulary continuous speech recognition in Greek: Modeling and nonlinear features,” Proc. Europ. Conf. Signal Process. (EUSIPCO), Marrakech, Morocco, 2013.
  • I. Rodomagoulakis, P. Giannoulis, Z.-I. Skordilis, P. Maragos, and G. Potamianos, “Experiments on far-field multichannel speech processing in smart homes,” Proc. Int. Conf. Digital Signal Process. (DSP), pp. 1–6, Santorini, Greece, 2013.
  • G. Galatas, G. Potamianos, and F. Makedon, “Robust multi-modal speech recognition in two languages utilizing video and distance information from the Kinect,” Proc. Int. Conf. Human-Computer Interaction (HCII), Las Vegas, NV, 2013.

2012

  • G. Potamianos, C. Neti, J. Luettin, and I. Matthews, “Audio-visual automatic speech recognition: An overview,” Audio-Visual Speech Processing, E. Vatikiotis-Bateson, G. Bailly, and P. Perrier (Eds.), Ch. 9, Cambridge University Press, 2012.
  • G. Galatas, G. Potamianos, and F. Makedon, “Audio-visual speech recognition incorporating facial depth information captured by the Kinect,” Proc. Europ. Conf. Signal Process. (EUSIPCO), pp. 2714–2717, Bucharest, Romania, 2012.
  • G. Galatas, G. Potamianos, and F. Makedon, “Audio-visual speech recognition using depth information from the Kinect in noisy video conditions,” Proc. Int. Conf. Pervasive Technologies Related to Assistive Environments (PETRA), Crete, Greece, 2012.
  • P. Giannoulis and G. Potamianos, “A hierarchical approach with feature selection for emotion recognition from speech,” Proc. Int. Conf. Language Resources and Evaluation (LREC), Istanbul, Turkey, 2012.

2011

  • K. Kumar, G. Potamianos, J. Navratil, E. Marcheret, and V. Libal, “Audio-visual speech synchrony detection by a family of bimodal linear prediction models,” Multibiometrics for Human Identification, B. Bhanu and V. Govindaraju (Eds.), Ch. 2, pp. 31–50, Cambridge University Press, 2011.
  • S.-H. G. Chan, J. Li, P. Frossard, and G. Potamianos, “Special section on interactive multimedia,” IEEE Transactions on Multimedia, vol. 13, no. 5, pp. 841–843, 2011.
  • G. Galatas, G. Potamianos, D. Kosmopoulos, C. McMurrough, and F. Makedon, “Bilingual corpus for AVASR using multiple sensors and depth information,” Proc. Int. Conf. Auditory-Visual Speech Process. (AVSP), pp. 103–106, Volterra, Italy, 2011.
  • N. Sarris, G. Potamianos, J.-M. Renders, C. Grover, E. Karstens, L. Kallipolitis, V. Tountopoulos, G. Petasis, A. Krithara, M. Galle, G. Jacquet, B. Alex, R. Tobin, and L. Bounegru, “A system for synergistically structuring news content from traditional media and the blogosphere,” Proc. E-Challenges Conference, Florence, Italy, 2011.
  • G. Galatas, G. Potamianos, A. Papangelis, and F. Makedon, “Audio visual speech recognition in noisy visual environments,” Proc. Int. Conf. Pervasive Technologies Related to Assistive Environments (PETRA), Crete, Greece, 2011.

2010

  • A. Waibel, R. Stiefelhagen, R. Carlson, J. Casas, J. Kleindienst, L. Lamel, O. Lanz, D. Mostefa, M. Omologo, F. Pianesi, L. Polymenakos, G. Potamianos, J. Soldatos, G. Sutschet, and J. Terken, “Computers in the human interaction loop,” Handbook of Ambient Intelligence and Smart Environments, H. Nakashima, H. Aghajan, and J.C. Augusto (Eds.), Part IX, pp. 1071–1116, Springer, 2010.
  • L.-H. Kim, M. Hasegawa-Johnson, G. Potamianos, and V. Libal, “Joint estimation of DOA and speech based on EM beamforming,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 121–124, Dallas, TX, 2010.

2009

  • P. Lucey, G. Potamianos, and S. Sridharan, “Visual speech recognition across multiple views,” Visual Speech Recognition: Lip Segmentation and Mapping, A. Wee-Chung Liew and S. Wang (Eds.), ch. X, pp. 294–325, Information Science Reference (IGI), 2009.
  • G. Potamianos, L. Lamel, M. Wolfel, J. Huang, E. Marcheret, C. Barras, X. Zhu, J. McDonough, J. Hernando, D. Macho, and C. Nadeu, “Automatic speech recognition,” Computers in the Human Interaction Loop, A. Waibel and R. Stiefelhagen (Eds.), ch. 6, pp. 43–59, Springer, 2009.
  • K. Bernardin, R. Stiefelhagen, A. Pnevmatikakis, O. Lanz, A. Brutti, J.R. Casas, and G. Potamianos, “Person tracking,” Computers in the Human Interaction Loop, A. Waibel and R. Stiefelhagen (Eds.), ch. 3, pp. 11–22, Springer, 2009.
  • H. Meng, S. Oviatt, G. Potamianos, and G. Rigoll, “Introduction to the special issue on multimodal processing in speech-based interactions,” IEEE Transactions on Audio, Speech, and Language Processing, vol. 17, no. 3, pp. 409–410, 2009.
  • K. Kumar, J. Navratil, E. Marcheret, V. Libal, and G. Potamianos, “Robust audio-visual speech synchrony detection by generalized bimodal linear prediction,” Proc. Conf. Int. Speech Comm. Assoc. (Interspeech), pp. 2251–2254, Brighton, United Kingdom, 2009.
  • K. Kumar, J. Navratil, E. Marcheret, V. Libal, G. Ramaswamy, and G. Potamianos, “Audio-visual speech synchronization detection using a bimodal linear prediction model,” Proc. IEEE Comp. Soc. Works. Biometrics, Held in Conjunction with CVPR, Miami Beach, FL, 2009.
  • V. Libal, B. Ramabhadran, N. Mana, F. Pianesi, P. Chippendale, O. Lanz, and G. Potamianos, “Multimodal classification of activities of daily living inside smart homes,” Int. Works. Ambient Assisted Living (IWAAL), LNCS vol. 5518, Part II, pp. 687–694, Salamanca, Spain, 2009.
  • X. Zhuang, J. Huang, G. Potamianos, and M. Hasegawa-Johnson, “Acoustic fall detection using Gaussian mixture models and GMM supervectors,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 69–72, Taipei, Taiwan, 2009.
  • J. Huang, X. Zhuang, V. Libal, and G. Potamianos, “Long-time span acoustic activity analysis from far-field sensors in smart homes,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 4173–4176, Taipei, Taiwan, 2009.
  • S.M. Chu, V. Goel, E. Marcheret, and G. Potamianos, Method for Likelihood Computation in Multi-Stream HMM Based Speech Recognition, Patent No.: US007480617B2, Jan. 20, 2009.

2008

  • G. Potamianos, J. Huang, E. Marcheret, V. Libal, R. Balchandran, M. Epstein, L. Seredi, M. Labsky, L. Ures, M. Black, and P. Lucey, “Far-field multimodal speech processing and conversational interaction in smart spaces,” Proc. Joint Work. Hands-Free Speech Communication and Microphone Arrays (HSCMA), Trento, Italy, 2008.
  • A Leone, G. Diraco, C. Distante, P. Siciliano, M. Grassi, A. Lombardi, G. Rescio, P. Malcovati, M. Malfatti, L. Gonzo, V. Libal, J. Huang, and G. Potamianos, “A multi-sensor approach for people fall detection in home environment,” Proc. 10th Europ. Conf. Computer Vision (ECCV), Marseille, France, 2008.
  • P. Lucey, G. Potamianos, and S. Sridharan, “Patch-based analysis of visual speech from multiple views,” Proc. Int. Conf. Auditory-Visual Speech Process. (AVSP), pp. 69–73, Tangalooma, Australia, 2008.
  • R. Balchandran, M. Epstein, G. Potamianos, and L. Seredi, “A multi-modal spoken dialog system for interactive TV,” Proc. Int. Conf. Multimodal Interfaces (ICMI) – Demo Papers, pp. 191–192, Chania, Greece, 2008.
  • M. Grassi, A. Lombardi, G. Rescio, P. Malcovati, A Leone, G. Diraco, C. Distante, P. Siciliano, M. Malfatti, L. Gonzo, V. Libal, J. Huang, and G. Potamianos, “A hardware-software framework for high-reliability people fall detection,” Proc. 7th IEEE Conf. on Sensors (SENSORS), pp. 1328–1331, Lecce, Italy, 2008.
  • A. Tyagi, J.W. Davis, and G. Potamianos, “Steepest descent for efficient covariance tracking,” Proc. IEEE Work. Motion and Video Computing (WMVC), Copper Mountain, Colorado, 2008.
  • J. Huang, E. Marcheret, K. Visweswariah, and G. Potamianos, “The IBM RT07 evaluation systems for speaker diarization on lecture meetings,” in Multimodal Technologies for Perception of Humans: International Evaluation Workshops CLEAR 2007 and RT 2007, Baltimore, Maryland, May 2007, LNCS vol. 4625, pp. 497–508, Springer, Berlin, 2008.
  • J. Huang, E. Marcheret, K. Visweswariah, V. Libal, and G. Potamianos, “The IBM Rich Transcription Spring 2007
    speech-to-text systems for lecture meetings,” in Multimodal Technologies for Perception of Humans: International Evaluation Workshops CLEAR 2007 and RT 2007, Baltimore, Maryland, May 2007, LNCS vol. 4625, pp. 429–441, Springer, Berlin, 2008.
  • S. Deligne, C.V. Neti, and G. Potamianos, Audio-Visual Codebook Dependent Cepstral Normalization, Patent No.: US007319955B2, Jan. 15, 2008.

2007

  • D. Mostefa, N. Moreau, K. Choukri, G. Potamianos, S.M. Chu, A. Tyagi, J.R. Casas, J. Turmo, L. Christoforetti, F. Tobia, A. Pnevmatikakis, V. Mylonakis, F. Talantzis, S. Burger, R. Stiefelhagen, K. Bernardin, and C. Rochet, “The CHIL audiovisual corpus for lecture and meeting analysis inside smart rooms,” Journal of Language Resources and Evaluation, vol. 41, pp. 389–407, 2007.
  • Z. Zhang, G. Potamianos, A.W. Senior, and T.S. Huang, “Joint face and head tracking inside multi-camera smart rooms,” Signal, Image and Video Processing, vol. 1, pp. 163–178, 2007.
  • V. Libal, J. Connell, G. Potamianos, and E. Marcheret, “An embedded system for in-vehicle visual speech activity detection,” Proc. Int. Work. Multimedia Signal Process. (MMSP), pp. 255–258, Chania, Greece, 2007.
  • P. Lucey, G. Potamianos, and S. Sridharan, “A unified approach to multi-pose audio-visual ASR,” Proc. Conf. Int. Speech Comm. Assoc. (Interspeech), pp. 650–653, Antwerp, Belgium, 2007.
  • J. Huang, E. Marcheret, K. Visweswariah, V. Libal, and G. Potamianos, “Detection, diarization, and transcription of far-field lecture speech,” Proc. Conf. Int. Speech Comm. Assoc. (Interspeech), pp. 2161–2164, Antwerp, Belgium, 2007.
  • P. Lucey, G. Potamianos, and S. Sridharan, “Pose-invariant audio-visual automatic speech recognition,” Proc. Work. Audio-Visual Speech Process. (AVSP), pp. 176–180, Hilvarenbeek, The Netherlands, 2007.
  • A. Tyagi, M. Keck, J.W. Davis, and G. Potamianos, “Kernel-based 3D tracking,” Proc. IEEE Int. Work. Visual Surveillance (VS/CVPR), Minneapolis, Minnesota, 2007.
  • A. Tyagi, G. Potamianos, J.W. Davis, and S.M. Chu, “Fusion of multiple camera views for kernel-based 3D tracking,” Proc. IEEE Work. Motion and Video Computing (WMVC), Austin, Texas, 2007.
  • E. Marcheret, V. Libal, and G. Potamianos, “Dynamic stream weight modeling for audio-visual speech recognition,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), vol. 4, pp. 945–948, Honolulu, Hawaii, 2007.
  • J.H. Connell, N. Haas, E. Marcheret, C.V. Neti, and G. Potamianos, Audio-Only Backoff in Audio-Visual Speech Recognition System, Patent No.: US007251603B2, July 31, 2007.
  • U.V. Chaudhari, C. Neti, G. Potamianos, and G.N. Ramaswamy, Automated Decision Making Using Time-Varying Stream Reliability Prediction, Patent No.: US007228279B2, June 5, 2007.

2006

  • G. Potamianos, “Audio-visual speech recognition,” Short Article, Encyclopedia of Language and Linguistics, Second Edition, (Speech Technology Section – Computer Understanding of Speech), K. Brown (Ed. In Chief), Elsevier, Oxford, United Kingdom, ISBN: 0-08-044299-4, vol. 11, pp. 800–805, 2006.
  • G. Potamianos and P. Lucey, “Audio-visual ASR from multiple views inside smart rooms,” Proc. Int. Conf. Multisensor Fusion and Integration for Intelligent Systems (MFI), pp. 35–40, Heidelberg, Germany, 2006.
  • Z. Zhang, G. Potamianos, S.M. Chu, J. Tu, and T.S. Huang, “Person tracking in smart rooms using dynamic programming and adaptive subspace learning,” Proc. Int. Conf. Multimedia Expo. (ICME), pp. 2061–2064, Toronto, Canada, 2006.
  • P. Lucey and G. Potamianos, “Lipreading using profile versus frontal views,” Proc. IEEE Work. Multimedia Signal Process. (MMSP), pp. 24–28, Victoria, Canada, 2006.
  • A.W. Senior, G. Potamianos, S. Chu, Z. Zhang, and A. Hampapur, “A comparison of multicamera person-tracking algorithms,” Proc. IEEE Int. Work. Visual Surveillance (VS/ECCV), Graz, Austria, 2006.
  • G. Potamianos and Z. Zhang, “A joint system for single-person 2D-face and 3D-head tracking in CHIL seminars,” Multimodal Technologies for Perception of Humans: First Int. Eval. Work. on Classification of Events, Activities and Relationships, CLEAR 2006, R. Stiefelhagen and J. Garofolo (Eds.), LNCS vol. 4122, pp. 105–118, Southampton, United Kingdom, 2006.
  • Z. Zhang, G. Potamianos, M. Liu, and T. Huang, “Robust multi-view multi-camera face detection inside smart rooms using spatio-temporal dynamic programming,” Proc. Int. Conf. Automatic Face and Gesture Recog. (FGR), Southampton, United Kingdom, 2006.
  • E. Marcheret, G. Potamianos, K. Visweswariah, and J. Huang, “The IBM RT06s evaluation system for speech activity detection in CHIL seminars,” Proc. RT06s Evaluation Work. – held with Joint Work. on Multimodal Interaction and Related Machine Learning Algorithms (MLMI), S. Renals, S. Bengio, and J.G. Fiscus (Eds.), LNCS vol. 4299, pp. 323–335, Washington DC, 2006.
  • J. Huang, M. Westphal, S. Chen, O. Siohan, D. Povey, V. Libal, A. Soneiro, H. Schulz, T. Ross, and G. Potamianos, “The IBM Rich Transcription Spring 2006 speech-to-text system for lecture meetings,” Proc. RT06s Evaluation Work.– held with Joint Work. on Multimodal Interaction and Related Machine Learning Algorithms (MLMI), S. Renals, S. Bengio, and J.G. Fiscus (Eds.), LNCS vol. 4299, pp. 432–443, Washington DC, 2006.

2005

  • P.S. Aleksic, G. Potamianos, and A.K. Katsaggelos, “Exploiting visual information in automatic speech processing,” Handbook of Image and Video Processing, Second Edition, Al. Bovic (Ed.), ch. 10.8, pp. 1263–1289, Elsevier Academic Press, Burlington, MA, ISBN: 0-12-119792-1, 2005.
  • G. Potamianos and P. Scanlon, “Exploiting lower face symmetry in appearance-based automatic speechreading,” Proc. Work. Audio-Visual Speech Process. (AVSP), pp. 79–84, Vancouver Island, Canada, 2005.
  • S.M. Chu, E. Marcheret, and G. Potamianos, “Automatic speech recognition and speech activity detection in the CHIL smart room,” Proc. Joint Work. on Multimodal Interaction and Related Machine Learning Algorithms (MLMI), LNCS vol. 3869, pp. 332–343, Edinburgh, United Kingdom, 2005.
  • Z. Zhang, G. Potamianos, A. Senior, S. Chu, and T. Huang, “A joint system for person tracking and face detection,” Proc. Int. Work. Human-Computer Interaction (ICCV 2005 Work. on HCI), pp. 47–59, Beijing, China, 2005.
  • E. Marcheret, K. Visweswariah, and G. Potamianos, “Speech activity detection fusing acoustic phonetic and energy features,” Proc. Europ. Conf. Speech Comm. Technol. (Interspeech), pp. 241–244, Lisbon, Portugal, 2005.
  • J. Jiang, G. Potamianos, and G. Iyengar, “Improved face finding in visually challenging environments,” Proc. Int. Conf. Multimedia Expo. (ICME), Amsterdam, The Netherlands, 2005.
  • D. Macho, J. Padrell, A. Abad, C. Nadeu, J. Hernando, J. McDonough, M. Wolfel, U. Klee, M. Omologo, A. Brutti, P. Svaizer, G. Potamianos, and S.M. Chu, “Automatic speech activity detection, source localization, and speech recognition on the CHIL seminar corpus,” Proc. Int. Conf. Multimedia Expo. (ICME), Amsterdam, The Netherlands, 2005.

2004

  • J. Huang, G. Potamianos, J. Connell, and C. Neti. “Audio-visual speech recognition using an infrared headset,” Speech Communication, vol. 44, no. 4, pp. 83–96, 2004.
  • P. Scanlon, G. Potamianos, V. Libal, and S.M. Chu, “Mutual information based visual feature selection for lipreading,” Proc. Int. Conf. Spoken Lang. Process. (ICSLP), Jeju Island, Korea, 2004.
  • E. Marcheret, S.M. Chu, V. Goel, and G. Potamianos, “Efficient likelihood computation in multi-stream HMM based audio-visual speech recognition,” Proc. Int. Conf. Spoken Lang. Process. (ICSLP), Jeju Island, Korea, 2004.
  • G. Potamianos, C. Neti, J. Huang, J.H. Connell, S. Chu, V. Libal, E. Marcheret, N. Haas, and J. Jiang, “Towards practical deployment of audio-visual speech recognition,” Invited, Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), vol. 3, pp. 777–780, Montreal, Canada, 2004.
  • J. Jiang, G. Potamianos, H. Nock, G. Iyengar, and C. Neti, “Improved face and feature finding for audio-visual speech recognition in visually challenging environments,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), vol. 5, pp. 873–876, Montreal, Canada, 2004.
  • S.M. Chu, V. Libal, E. Marcheret, C. Neti, and G. Potamianos, “Multistage information fusion for audio-visual speech recognition,” Proc. Int. Conf. Multimedia Expo. (ICME), Taipei, Taiwan, 2004.
  • P. de Cuetos, G.R. Iyengar, C.V. Neti, and G. Potamianos, System and Method for Microphone Activation Using Visual Speech Cues, Patent No.: US006754373B1, June 22, 2004.

2003

  • G. Potamianos, C. Neti, G. Gravier, A. Garg, and A.W. Senior, “Recent advances in the automatic recognition of audio-visual speech,” Invited, Proceedings of the IEEE, vol. 91, no. 9, pp. 1306–1326, 2003.
  • G. Potamianos, C. Neti, and S. Deligne, “Joint audio-visual speech processing for recognition and enhancement,” Proc. Work. Audio-Visual Speech Process. (AVSP), pp. 95–104, St. Jorioz, France, 2003.
  • J. Huang, G. Potamianos, and C. Neti, “Improving audio-visual speech recognition with an infrared headset,” Proc. Work. Audio-Visual Speech Process. (AVSP), pp. 175–178, St. Jorioz, France, 2003.
  • G. Potamianos and C. Neti, “Audio-visual speech recognition in challenging environments,” Proc. Europ. Conf. Speech Comm. Technol. (Eurospeech), pp. 1293–1296, Geneva, Switzerland, 2003.
  • J.H. Connell, N. Haas, E. Marcheret, C. Neti, G. Potamianos, and S. Velipasalar, “A real-time prototype for small-vocabulary audio-visual ASR,” Proc. Int. Conf. Multimedia Expo. (ICME), vol. II, pp. 469-472, Baltimore, MD, 2003.
  • U.V. Chaudhari, G.N. Ramaswamy, G. Potamianos, and C. Neti, “Information fusion and decision cascading for audio-visual speaker recognition based on time varying stream reliability prediction,” Proc. Int. Conf. Multimedia Expo. (ICME), pp. 9–12, Baltimore, MD, 2003.
  • A. Garg, G. Potamianos, C. Neti, and T.S. Huang, “Frame-dependent multi-stream reliability indicators for audio-visual speech recognition,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), vol. I, pp. 24–27, Hong Kong, China, 2003.
  • U.V. Chaudhari, G.N. Ramaswamy, G. Potamianos, and C. Neti, “Audio-visual speaker recognition using time-varying stream reliability prediction,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), vol. V, pp. 712–715, Hong Kong, China, 2003.
  • E. Cosatto, H.P. Graf, G. Potamianos, and J. Schroeter, Audio-Visual Selection Process for the Synthesis of Photo-Realistic Talking-Head Animations, Patent No.: US006654018B1, Nov. 25, 2003.

2002

  • C. Neti, G. Potamianos, J. Luettin, and E. Vatikiotis-Bateson, “Editorial of the special issue on joint audio-visual speech processing,” EURASIP Journal on Applied Signal Processing, vol. 2002, no. 11, 2002.
  • S. Deligne, G. Potamianos, and C. Neti, “Audio-visual speech enhancement with AVCDCN (audio-visual codebook dependent cepstral normalization)”, Proc. Int. Conf. Spoken Lang. Process. (ICSLP), vol. 3, pp. 1449–1452, Denver, CO, 2002.
  • R. Goecke, G. Potamianos, and C. Neti, “Noisy audio feature enhancement using audio-visual speech data,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 2025–2028, Orlando, FL, 2002.
  • G. Gravier, S. Axelrod, G. Potamianos, and C. Neti, “Maximum entropy and MCE based HMM stream weight estimation for audio-visual ASR,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 853–856, Orlando, FL, 2002.
  • G. Gravier, G. Potamianos, and C. Neti, “Asynchrony modeling for audio-visual speech recognition,” Proc. Human Lang. Techn. Conf. (HLT), pp. 1–6, San Diego, CA, 2002.

2001

  • G. Potamianos, C. Neti, G. Iyengar, A.W. Senior, and A. Verma, “A cascade visual front end for speaker independent automatic speechreading,” International Journal of Speech Technology, Special Issue on Multimedia, vol. 4, pp. 193–208, 2001.
  • G. Potamianos, C. Neti, G. Iyengar, and E. Helmuth, “Large-vocabulary audio-visual speech recognition by machines and humans,” Proc. Europ. Conf. Speech Comm. Technol. (Eurospeech), pp. 1027–1030, Aalborg, Denmark, 2001.
  • G. Potamianos and C. Neti, “Automatic speechreading of impaired speech,” Proc. Work. Audio-Visual Speech Process. (AVSP), pp. 177–182, Aalborg, Denmark, 2001.
  • G. Potamianos and C. Neti, “Improved ROI and within frame discriminant features for lipreading,” Proc. Int. Conf. Image Process. (ICIP), pp. 250–253, Thessaloniki, Greece, 2001.
  • C. Neti, G. Potamianos, J. Luettin, I. Matthews, H. Glotin, and D. Vergyri, “Large-vocabulary audio-visual speech recognition: A summary of the Johns Hopkins Summer 2000 Workshop,” Proc. IEEE Work. Multimedia Signal Process. (MMSP), pp. 619–624, Cannes, France, 2001.
  • G. Iyengar, G. Potamianos, C. Neti, T. Faruquie, and A. Verma, “Robust detection of visual ROI for automatic
    speechreading,” Proc. IEEE Work. Multimedia Signal Process. (MMSP), pp. 79–84, Cannes, France, 2001.
  • I. Matthews, G. Potamianos, C. Neti, and J. Luettin, “A comparison of model and transform-based visual features for audio-visual LVCSR,” Proc. Int. Conf. Multimedia Expo. (ICME), Tokyo, Japan, 2001.
  • G. Potamianos, J. Luettin, and C. Neti, “Hierarchical discriminant features for audio-visual LVCSR,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), vol. 1, pp. 165–168, Salt Lake City, UT, 2001.
  • J. Luettin, G. Potamianos, and C. Neti, “Asynchronous stream modeling for large-vocabulary audio-visual speech recognition,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), vol. 1, pp. 169–172, Salt Lake City, UT, 2001.
  • H. Glotin, D. Vergyri, C. Neti, G. Potamianos, and J. Luettin, “Weighting schemes for audio-visual fusion in speech recognition,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), vol. 1, pp. 173–176, Salt Lake City, UT, 2001.

2000

  • G. Potamianos and C. Neti, “Stream confidence estimation for audio-visual speech recognition,” Proc. Int. Conf. Spoken Language Process. (ICSLP), vol. III, pp. 746–749, Beijing, China, 2000.
  • C. Neti, G. Iyengar, G. Potamianos, A. Senior, and B. Maison, “Perceptual interfaces for information interaction: Joint processing of audio and visual information for human-computer interaction”, Proc. Int. Conf. Spoken Language Process. (ICSLP), vol III, pp. 11-14, Beijing, China, 2000.
  • G. Potamianos, A. Verma, C. Neti, G. Iyengar, and S. Basu, “A cascade image transform for speaker independent automatic speechreading,” Proc. IEEE Int. Conf. Multimedia Expo. (ICME), vol. II, pp. 1097–1100, New York, NY, 2000.
  • E. Cosatto, G. Potamianos, and H.P. Graf, “Audio-visual unit selection for the synthesis of photo-realistic talking-heads,” Proc. IEEE Int. Conf. Multimedia Expo. (ICME), vol. II, pp. 619–622, New York, NY, 2000.
  • E. Cosatto, H.P. Graf, and G. Potamianos, Robust Multi-Modal Method for Recognizing Objects, Patent No.: US006118887A, Sep. 12, 2000.

Past Millennium

  • G. Potamianos and A. Potamianos, “Speaker adaptation for audio-visual automatic speech recognition,” Proc. Europ. Speech Comm. Technol. (Eurospeech), Budapest, Hungary, vol. 3, pp. 1291–1294, 1999.
  • G. Potamianos and F. Jelinek, “A study of n-gram and decision tree letter language modeling methods,” Speech Communication, vol. 24, no. 3, pp. 171–192, 1998.
  • G. Potamianos and H.P. Graf, “Linear discriminant analysis for speechreading,” Proc. IEEE Work. Multimedia Signal Process. (MMSP), Los Angeles, CA, pp. 221–226, 1998.
  • G. Potamianos, H.P. Graf, and E. Cosatto, “An image transform approach for HMM based automatic lipreading,” Proc. Int. Conf. Image Process. (ICIP), Chicago, IL, pp. 173–177, 1998.
  • G. Potamianos and H.P. Graf, “Discriminative training of HMM stream exponents for audio-visual speech recognition,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), Seattle, WA, vol. 6, pp. 3733–3736, 1998.
  • H.P. Graf, E. Cosatto, and G. Potamianos, “Machine vision of faces and facial features,” Proc. R.I.E.C. Int. Symp. Design Archit. Inform. Process. Systems Based Brain Inform. Princ., Sendai, Japan, pp. 48–53, 1998.
  • G. Potamianos and J. Goutsias, “Stochastic approximation algorithms for partition function estimation of Gibbs random fields,” IEEE Transactions on Information Theory, vol. 43, no. 6, pp. 1948–1965, 1997.
  • G. Potamianos, E. Cosatto, H.P. Graf, and D.B. Roe, “Speaker independent audio-visual database for bimodal ASR,” Proc. Europ. Tutorial Research Work. Audio-Visual Speech Process. (AVSP), Rhodes, Greece, pp. 65–68, 1997.
  • H.P. Graf, E. Cosatto, and G. Potamianos, “Robust recognition of faces and facial features with a multi-modal system,” Proc. Int. Conf. Systems Man Cybern. (ICSMC), Orlando, FL, pp. 2034–2039, 1997.
  • G. Potamianos, “Efficient Monte Carlo estimation of partition function ratios of Markov random field images,” Proc. Conf. Inform. Sci. Systems (CISS), Princeton, NJ, vol. II, pp. 1212–1215, 1996.
  • G. Potamianos and J. Goutsias, “A unified approach to Monte Carlo likelihood estimation of Gibbs random field images,” Proc. Conf. Inform. Sci. Systems (CISS), Princeton, NJ, vol. I, pp. 84–90, 1994.
  • G.G. Potamianos and J. Goutsias, “Partition function estimation of Gibbs random field images using Monte Carlo simulations,” IEEE Transactions on Information Theory, vol. 39, no. 4, pp. 1322–1332, 1993.
  • G. Potamianos and J. Goutsias, “An analysis of Monte Carlo methods for likelihood estimation of Gibbsian images,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), Minneapolis, MN, vol. V, pp. 519–522, 1993.
  • G. Potamianos and J. Goutsias, “On computing the likelihood function of partially observed Markov random field images using Monte Carlo simulations,” Proc. Conf. Inform. Sci. Systems (CISS), Princeton, vol. I, pp. 357–362, 1992.
  • G. Potamianos and J. Goutsias, “A novel method for computing the partition function of Markov random field images using Monte Carlo simulations,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), Toronto, Canada, vol. 4, pp. 2325–2328, 1991.
  • G. Potamianos and J. Diamessis, “Frequency sampling design of 2-D IIR filters using continued fractions,” Proc. Int. Symp. Circuits Systems (ISCAS), New Orleans, LA, pp. 2454–2457, 1990.
  • J. Diamessis and G. Potamianos, “A novel method for designing IIR filters with nonuniform samples,” Proc. Conf. Inform. Sci. Systems (CISS), Princeton, NJ, vol. 1, pp. 192–195, 1990.
  • J. Diamessis and G. Potamianos, “Modeling unequally spaced 2-D discrete signals by rational functions,” Proc. Int. Symp. Circuits Systems (ISCAS), Portland, OR, pp. 1508–1511, 1989.

[Last update: 09/05/2021]

Research Interests: Multimodal speech processing with applications to human-computer interaction and ambient intelligence. Particular emphasis on audio-visual speech processing, far-field automatic speech recognition, multichannel acoustic scene analysis, multimedia signal processing and fusion, sign language recognition, computer vision for human tracking and activity recognition, incorporating deep learning methods.

Undergraduate Courses:

ECE218 Signals and Systems

ECE325 Digital Signal Processing

ECE334 Pattern Recognition

ECE352 Image Processing

ECE432 Audio and Video Conding

ECE443 Speech and Audio Processing

ECE457 Introduction to Computer Vision

PGS671 Advanced Topics in Speech and Language Processing

Postgraduate Courses of PGS in "Science and Technology of ECE":

PGS613 Computer Vision

PGS671 Advanced Topics in Speech and Language Processing

Ph.D. Graduates:

Spyridon Thermos

Supervisor Gerasimos Potamianos, Associate Professor
Phd Thesis TitleA Deep Learning Approach to 2D/3D Object Affordance Understanding
Year 2020
Ph.D. Candidates:

Alexandros Koumparoulis

Supervisor Gerasimos Potamianos, Associate Professor
Phd Thesis TitleAudio Visual Speech Processing in the Wild
Email alkoumpa@e-ce.uth.gr

Catherine Papadimitriou

Supervisor Gerasimos Potamianos, Associate Professor
Phd Thesis TitleSign Language Recognition
Email aipapadimitriou@e-ce.uth.gr

Nikolaos Tsikrikonis

Supervisor Gerasimos Potamianos, Associate Professor
Phd Thesis TitleConformal Microstrip Antenna with Beam Steering

e-Yπηρεσίες

Contact Info

  • Sekeri – Cheiden Str, Pedion Areos, Volos
  • Phone: +30 24210 74967
  • Email: gece@e-ce.uth.gr
  • PGS Tel.: +30 24210 74934

Announcements

  • Academic News

Find us

  • Facebook
  • Twitter
  • Youtube
  • Linkedin
© Copyright 2023 Department of Electrical and Computer Engineering