Gerasimos Potamianos, Associate Professor | |
---|---|
Diploma | ECE, National Technical University of Athens (1988) |
Ph.D. | ECE, The Johns Hopkins University (1994) |
Office | 420 |
Office Hours | Wednesday 14:00 - 16:00 (or by appointment) |
Tel. | +30 24210 74928 |
gpotamianos@e-ce.uth.gr | |
Web Page CV Google Scholar |
[Last update: 04/04/2024]
2023
- K. Papadimitriou, G. Potamianos, G. Sapountzaki, T. Goulas, E. Efthimiou, S.-E. Fotinea, and P. Maragos, “Greek sign language recognition for an education platform,” Universal Access in the Information Society, 2023.
- K. Papadimitriou and G. Potamianos, “Multimodal locally enhanced Transformer for continuous sign language recognition,” Proc. Conf. Int. Speech Comm. Assoc. (Interspeech), pp. 1513–1517, Dublin, Ireland, 2023.
- K. Papadimitriou and G. Potamianos, “Sign language recognition via deformable 3D convolutions and modulated graph convolutional networks,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 1–5, Rhodes, Greece, 2023.
- K. Papadimitriou, G. Sapountzaki, K. Vasilaki, E. Efthimiou, S.-E. Fotinea, and G. Potamianos, “SL-REDU GSL: A large Greek sign language recognition corpus,” Proc. Int. Conf. Acoust. Speech Signal Process. Works. (ICASSPW) – Int. Works. Sign Language Translation and Avatar Technology (SLTAT), pp. 1–5, Rhodes, Greece, 2023.
- G. Sapountzaki, E. Efthimiou, S.-E. Fotinea, K. Papadimitriou, and G. Potamianos, “Remote learning and assessment of Greek sign language in the undergraduate curriculum in COVID time,” Proc. Int. Conf. Education and New Learning Technologies (EDULEARN), pp. 5452–5459, Palma, Spain, 2023.
2022
- N. Efthymiou, P. P. Filntisis, P. Koutras, A. Tsiami, J. Hadfield, G. Potamianos, and P. Maragos, “ChildBot: Multi-robot perception and interaction with children,” Robotics and Autonomous Systems, vol. 150 (103975), 2022.
- M. Parelli, K. Papadimitriou, G. Potamianos, G. Pavlakos, and P. Maragos, “Spatio-temporal graph convolutional networks for continuous sign language recognition,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 8457–8461, Singapore, 2022.
- K. Papadimitriou, G. Potamianos, G. Sapountzaki, T. Goulas, E. Efthimiou, S.-E. Fotinea, and P. Maragos, “Greek sign language recognition for the SL-ReDu learning platform,” Proc. Int. Work. Sign Language Translation and Avatar Technology (SLTAT) – Satellite to LREC, pp. 79–84, Marseille, France, 2022.
- G. Sapountzaki, E. Efthimiou, S.-E. Fotinea, K. Papadimitriou, and G. Potamianos, “3D Greek sign language classifiers as a learning object in the SL-ReDu online education platform,” Proc. Int. Conf. Education and New Learning Technologies (EDULEARN), pp. 6146–6153, Palma, Spain, 2022.
2021
- S. Thermos, G. Potamianos, and P. Darras, “Joint object affordance reasoning and segmentation in RGB-D videos,” IEEE Access, vol. 9, pp. 89699–89713, 2021.
- N. Efthymiou, P. P. Filntisis, G. Potamianos, and P. Maragos, “Visual robotic perception system with incremental learning for child-robot interaction scenarios,” Technologies, vol. 9(4), no. 86, 2021.
- P. Giannoulis, G. Potamianos, and P. Maragos, “Overlapped sound event classification via multi-channel sound separation network,” Proc. Europ. Conf. Signal Process. (EUSIPCO), pp. 571–575, Dublin, Ireland, 2021.
- A. Koumparoulis, G. Potamianos, S. Thomas, and E. da Silva Morais, “Resource-efficient TDNN architectures for audio-visual speech recognition,” Proc. Europ. Conf. Signal Process. (EUSIPCO), pp. 506–510, Dublin, Ireland, 2021.
- P. P. Filntisis, N. Efthymiou, G. Potamianos, and P. Maragos, “An audiovisual child emotion recognition system for child-robot interaction applications,” Proc. Europ. Conf. Signal Process. (EUSIPCO), pp. 791–795, Dublin, Ireland, 2021.
- N. Efthymiou, P. P. Filntisis, G. Potamianos, and P. Maragos, “A robotic edutainment framework for designing childrobot interaction scenarios,” Proc. ACM Int. Conf. PErvarsive Technologies Related to Assistive Environments (PETRA), pp. 160–166, Corfu, Greece, 2021.
- K. Papadimitriou, M. Parelli, G. Sapountzaki, G. Pavlakos, P. Maragos, and G. Potamianos, “Multimodal fusion and sequence learning for cued speech recognition from videos,” Universal Access in Human-Computer Interaction. Access to Media, Learning and Assistive Environments. UAHCI 2021 / HCII 2021, Part II, M. Antona and C. Stephanidis (Eds.), pp. 277–290, LNCS vol. 12769, Springer, Cham, 2021.
- E. Efthimiou, S.-E. Fotinea, C. Flouda, T. Goulas, G. Ametoglou, G. Sapountzaki, K. Papadimitriou, and G. Potamianos, “The SL-ReDu environment for self-monitoring and objective learner assessment in Greek sign language,” Universal Access in Human-Computer Interaction. Access to Media, Learning and Assistive Environments. UAHCI 2021 / HCII 2021, Part II, M. Antona and C. Stephanidis (Eds.), pp. 72–81, LNCS vol. 12769, Springer, Cham, 2021.
- G. Sapountzaki, E. Efthimiou, S.-E. Fotinea, K. Papadimitriou, and G. Potamianos, “Educational material organization in a platform for Greek sign language self monitoring and assessment,” Proc. Int. Conf. Education and New Learning Technologies (EDULEARN), pp. 3322–3331, Palma, Spain, 2021.
2020
- S. Thermos, G. T. Papadopoulos, P. Darras, and G. Potamianos, “Deep sensorimotor learning for RGB-D object recognition,” Computer Vision and Image Understanding, vol. 190, 2020.
- A. Koumparoulis, G. Potamianos, S. Thomas, and E. da Silva Morais, “Resource-adaptive deep learning for visual speech recognition,” Proc. Conf. Int. Speech Comm. Assoc. (Interspeech), pp. 3510–3514, Shanghai, China, 2020.
- K. Papadimitriou and G. Potamianos, “Multimodal sign language recognition via temporal deformable convolutional sequence learning,” Proc. Conf. Int. Speech Comm. Assoc. (Interspeech), pp. 2752–2756, Shanghai, China, 2020.
- K. Papadimitriou and G. Potamianos, “A fully convolutional sequence learning approach for cued speech recognition from videos,” Proc. Europ. Conf. Signal Process. (EUSIPCO), pp. 326–330, Amsterdam, The Netherlands, 2020.
- M. Parelli, K. Papadimitriou, G. Potamianos, G. Pavlakos, and P. Maragos, “Exploiting 3D hand pose estimation in deep learning-based sign language recognition from RGB videos,” Computer Vision – ECCV 2020 Workshops Proceedings, Part II, A. Bartoli and A. Fusiello (Eds.), pp. 249–263, LNCS/LNIP vol. 12536, 2020.
- P. P. Filntisis, N. Efthymiou, G. Potamianos, and P. Maragos, “Emotion understanding in videos through body, context, and visual-semantic embedding loss,” Computer Vision – ECCV 2020 Workshops Proceedings, Part I, A. Bartoli and A. Fusiello (Eds.), pp. 747–755, LNCS/LNIP vol. 12535, Springer, 2020.
- G. Potamianos, K. Papadimitriou, E. Efthimiou, S.-E. Fotinea, G. Sapountzaki, and P. Maragos, “SL-ReDu: Greek sign language recognition for educational applications. Project description and early results,” Proc. ACM Int. Conf. PErvarsive Technologies Related to Assistive Environments (PETRA), article no. 59, pp. 1–6, Corfu, Greece, 2020.
- A. Koumparoulis, G. Potamianos, S. Thomas, and E. Morais, “Audio-assisted image inpainting for talking faces,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 7664–7668, Barcelona, Spain, 2020.
- S. Thermos, P. Darras, and G. Potamianos, “A deep learning approach to object affordance segmentation,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 2358–2362, Barcelona, Spain, 2020.
2019
- S. Thermos, G. T. Papadopoulos, P. Darras, and G. Potamianos, “Deep sensorimotor learning for RGB-D object recognition,” Computer Vision and Image Understanding, vol. 190, 2020.
- P. Giannoulis, G. Potamianos, and P. Maragos, “Room-localized speech activity detection in multi-microphone smart homes,” EURASIP Journal on Audio, Speech, and Music Processing, vol. 2019, no. 15, pp. 1–23, 2019.
- P. P. Filntisis, N. Efthymiou, P. Koutras, G. Potamianos, and P. Maragos, “Fusing body posture with facial expressions for joint recognition of affect in child-robot interaction,” IEEE Robotics and Automation Letters, vol. 4, no. 4, pp. 4011–4018, 2019.
- S. Oviatt, B. Schuller, P. Cohen, D. Sonntag, G. Potamianos, and A. Kruger (Eds.), The Handbook of Multimodal-Multisensor Interfaces, Volume 3: Language Processing, Software and Commercialization, Emerging Directions, ACM Books / Morgan-Claypool Publishers, San Rafael, CA, 2019.
- S. P. Chytas and G. Potamianos, “Hierarchical detection of sound events and their localization using convolutional neural networks with adaptive thresholds,” Proc. Detection and Classification of Acoustic Scenes and Events Workshop (DCASE), pp. 50–54, New York, NY, 2019.
- A. Koumparoulis and G. Potamianos, “MobiLipNet: Resource-efficient deep learning based lipreading,” Proc. Conf. Int. Speech Comm. Assoc. (Interspeech), pp. 2763–2767, Graz, Austria, 2019.
- K. Papadimitriou and G. Potamianos, “End-to-end convolutional sequence learning for ASL fingerspelling recognition,” Proc. Conf. Int. Speech Comm. Assoc. (Interspeech), pp. 2315–2319, Graz, Austria, 2019.
- K. Papadimitriou and G. Potamianos, “Fingerspelled alphabet sign recognition in upper-body videos,” Proc. Europ. Conf. Signal Process. (EUSIPCO), La Coruna, Spain, 2019.
2018
- S. Oviatt, B. Schuller, P. Cohen, D. Sonntag, G. Potamianos, and A. Kruger (Eds.), The Handbook of Multimodal-Multisensor Interfaces, Volume 2: Signal Processing, Architectures, and Detection of Emotion and Cognition, ACM Books / Morgan-Claypool Publishers, San Rafael, CA, 2018.
- A. Koumparoulis and G. Potamianos, “Deep view2view mapping for view-invariant lipreading,” Proc. IEEE Spoken Language Technology Works. (SLT), pp. 588–594, Athens, Greece, 2018.
- K. Papadimitriou and G. Potamianos, “A hybrid approach to hand detection and type classification in upper-body videos,” Proc. Europ. Work. Visual Information Process. (EUVIP), Tampere, Finland, 2018.
- J. Hadfield, P. Koutras, N. Efthymiou, G. Potamianos, C. S. Tzafestas, and P. Maragos, “Object assembly guidance in child-robot interaction using RGB-D based 3D tracking,” Proc Int. Conf. Intell. Robots Systems (IROS), pp. 347–354, Madrid, Spain, 2018.
- P. Giannoulis, G. Potamianos, and P. Maragos, “Multi-channel non-negative matrix factorization for overlapped acoustic event detection,” Proc. Europ. Conf. Signal Process. (EUSIPCO), pp. 857–861, Rome, Italy, 2018.
- N. Efthymiou, P. Koutras, P. P. Filntisis, G. Potamianos, and P. Maragos, “Multi-view fusion for action recognition in child-robot interaction,” Proc. Int. Conf. Image Process. (ICIP), pp. 455–459, Athens, Greece, 2018.
- S. Thermos, G. T. Papadopoulos, P. Darras, and G. Potamianos, “Attention-enhanced sensorimotor object recognition,” Proc. Int. Conf. Image Process. (ICIP), pp. 336–340, Athens, Greece, 2018.
- A. Tsiami, P. Koutras, N. Efthymiou, P. P. Filntisis, G. Potamianos, and P. Maragos, “Multi3: Multi-sensory perception system for multi-modal child interaction with multiple robots,” Proc. IEEE Int. Conf. on Robotics and Automation (ICRA), Brisbane, Australia, 2018.
- A. Tsiami, P. P. Filntisis, N. Efthymiou, P. Koutras, G. Potamianos, and P. Maragos, “Far-field audio-visual scene perception of multi-party human-robot interaction for children and adults,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 6568–6572, Calgary, Canada, 2018.
2017
- I. Rodomagoulakis, A. Katsamanis, G. Potamianos, P. Giannoulis, A. Tsiami, and P. Maragos, “Room-localized spoken command recognition in multi-room, multi-microphone environments,” Computer Speech and Language, vol. 46, pp. 419–443, 2017.
- G. Potamianos, E. Marcheret, Y. Mroueh, V. Goel, A. Koumbaroulis, A. Vartholomaios, and S. Thermos, “Audio and visual modality combination in speech processing applications”, The Handbook of Multimodal-Multisensor Interfaces, Volume 1: Foundations, User Modeling, and Common Modality Combinations, S. Oviatt, B. Schuller, P. Cohen, D. Sonntag, G. Potamianos, and A. Kruger (Eds.), Ch. 12, pp. 489–543, ACM Books / Morgan-Claypool Publishers, San Rafael, CA, 2017.
- S. Oviatt, B. Schuller, P. Cohen, D. Sonntag, G. Potamianos, and A. Kruger (Eds.), The Handbook of Multimodal-Multisensor Interfaces, Volume 1: Foundations, User Modeling, and Common Modality Combinations, ACM Books / Morgan-Claypool Publishers, San Rafael, CA, 2017.
- A. Koumparoulis, G. Potamianos, Y. Mroueh, and S. J. Rennie, “Exploring ROI size in deep learning based lipreading,” Proc. Int. Conf. on Auditory-Visual Speech Process. (AVSP), pp. 64–69, Stockholm, Sweden, 2017.
- P. Giannoulis, G. Potamianos, and P. Maragos, “On the joint use of NMF and classification for overlapping acoustic event detection,” Proc. Int. Works. on Computational Intelligence for Multimedia Understanding (IWCIM), Held in Conjunction with Europ. Conf. Signal Process. (EUSIPCO), Kos Island, Greece, 2017.
- S. Thermos, G. T. Papadopoulos, P. Darras, and G. Potamianos, “Deep affordance-grounded sensorimotor object recognition,” Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), pp. 49–57, Honolulu, HI, 2017.
2016
- S. Thermos and G. Potamianos, “Audio-visual speech activity detection in a two-speaker scenario incorporating depth information from a profile or frontal view,” Proc. IEEE Spoken Language Technology Works. (SLT), pp. 579–584, San Diego, CA, 2016.
- P. Giannoulis, G. Potamianos, P. Maragos, and A. Katsamanis, “Improved dictionary selection and detection schemes in sparse-CNMF-based overlapping acoustic event detection,” Proc. Detection and Classification of Acoustic Scenes and Events Works. (DCASE), pp. 25–29, Budapest, Hungary, 2016.
- A. Tsiami, A. Katsamanis, I. Rodomagoulakis, G. Potamianos, and P. Maragos, “Sweet home listen: A distant speech recognition system for home automation control,” “Show and Tell” Session Demonstration, Int. Conf. Acoust. Speech Signal Process. (ICASSP), Shanghai, China, 2016.
2015
- E. Marcheret, G. Potamianos, J. Vopicka, and V. Goel, “Scattering vs. discrete cosine transform features in visual speech processing,” Proc. Int. Joint Conf. on Facial Analysis, Animation and Auditory-Visual Speech Process. (FAAVSP), pp. 175–180, Vienna, Austria, 2015.
- E. Marcheret, G. Potamianos, J. Vopicka, and V. Goel, “Detecting audio-visual synchrony using deep neural networks,” Proc. Conf. Int. Speech Comm. Assoc. (Interspeech), pp. 548–552, Dresden, Germany, 2015.
- P. Giannoulis, A. Brutti, M. Matassoni, A. Abad, A. Katsamanis, M. Matos, G. Potamianos, and P. Maragos, “Multiroom speech activity detection using a distributed microphone network in domestic environments,” Proc. Europ. Conf. Signal Process. (EUSIPCO), pp. 1271–1275, Nice, France, 2015.
- Z. I. Skordilis, A. Tsiami, P. Maragos, G. Potamianos, L. Spelgatti, and R. Sannino, “Multichannel speech enhancement using MEMS microphones,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 2729–2733, Brisbane, Australia, 2015.
2014
- G. Floros, K. Kyritsis, and G. Potamianos, “Database and baseline system for detecting degraded traffic signs in urban environments,” Proc. Europ. Work. Visual Information Process. (EUVIP), Paris, France, 2014.
- A. Tsiami, I. Rodomagoulakis, P. Giannoulis, A. Katsamanis, G. Potamianos, and P. Maragos, “ATHENA: A Greek multisensory database for home automation control,” Proc. Conf. Int. Speech Comm. Assoc. (Interspeech), pp. 1608–1612, Singapore, 2014.
- P. Giannoulis, G. Potamianos, A. Katsamanis, and P. Maragos, “Multi-microphone fusion for detection of speech and acoustic events in smart spaces,” Proc. Europ. Conf. Signal Process. (EUSIPCO), pp. 2375–2379, Lisbon, Portugal, 2014.
- A. Tsiami, A. Katsamanis, P. Maragos, and G. Potamianos, “Experiments in acoustic source localization using sparse arrays in adverse indoors environments, Proc. Europ. Conf. Signal Process. (EUSIPCO), pp. 2390–2394, Lisbon, Portugal, 2014.
- P. Giannoulis, A. Tsiami, I. Rodomagoulakis, A. Katsamanis, G. Potamianos, and P. Maragos, “The Athena-RC system for speech activity detection and speaker localization in the DIRHA smart home,” Proc. Joint Work. Hands-Free Speech Communication and Microphone Arrays (HSCMA), pp. 167–171, Nancy, France, 2014.
- A. Katsamanis, I. Rodomagoulakis, G. Potamianos, P. Maragos, and A. Tsiami, “Robust far-field spoken command recognition for home automation combining adaptation and multichannel processing,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 5547–5551, Florence, Italy, 2014.
2013
- I. Rodomagoulakis, G. Potamianos, and P. Maragos, “Advances in large vocabulary continuous speech recognition in Greek: Modeling and nonlinear features,” Proc. Europ. Conf. Signal Process. (EUSIPCO), Marrakech, Morocco, 2013.
- I. Rodomagoulakis, P. Giannoulis, Z.-I. Skordilis, P. Maragos, and G. Potamianos, “Experiments on far-field multichannel speech processing in smart homes,” Proc. Int. Conf. Digital Signal Process. (DSP), pp. 1–6, Santorini, Greece, 2013.
- G. Galatas, G. Potamianos, and F. Makedon, “Robust multi-modal speech recognition in two languages utilizing video and distance information from the Kinect,” Proc. Int. Conf. Human-Computer Interaction (HCII), Las Vegas, NV, 2013.
2012
- G. Potamianos, C. Neti, J. Luettin, and I. Matthews, “Audio-visual automatic speech recognition: An overview,” Audio-Visual Speech Processing, E. Vatikiotis-Bateson, G. Bailly, and P. Perrier (Eds.), Ch. 9, Cambridge University Press, 2012.
- G. Galatas, G. Potamianos, and F. Makedon, “Audio-visual speech recognition incorporating facial depth information captured by the Kinect,” Proc. Europ. Conf. Signal Process. (EUSIPCO), pp. 2714–2717, Bucharest, Romania, 2012.
- G. Galatas, G. Potamianos, and F. Makedon, “Audio-visual speech recognition using depth information from the Kinect in noisy video conditions,” Proc. Int. Conf. Pervasive Technologies Related to Assistive Environments (PETRA), Crete, Greece, 2012.
- P. Giannoulis and G. Potamianos, “A hierarchical approach with feature selection for emotion recognition from speech,” Proc. Int. Conf. Language Resources and Evaluation (LREC), Istanbul, Turkey, 2012.
2011
- K. Kumar, G. Potamianos, J. Navratil, E. Marcheret, and V. Libal, “Audio-visual speech synchrony detection by a family of bimodal linear prediction models,” Multibiometrics for Human Identification, B. Bhanu and V. Govindaraju (Eds.), Ch. 2, pp. 31–50, Cambridge University Press, 2011.
- S.-H. G. Chan, J. Li, P. Frossard, and G. Potamianos, “Special section on interactive multimedia,” IEEE Transactions on Multimedia, vol. 13, no. 5, pp. 841–843, 2011.
- G. Galatas, G. Potamianos, D. Kosmopoulos, C. McMurrough, and F. Makedon, “Bilingual corpus for AVASR using multiple sensors and depth information,” Proc. Int. Conf. Auditory-Visual Speech Process. (AVSP), pp. 103–106, Volterra, Italy, 2011.
- N. Sarris, G. Potamianos, J.-M. Renders, C. Grover, E. Karstens, L. Kallipolitis, V. Tountopoulos, G. Petasis, A. Krithara, M. Galle, G. Jacquet, B. Alex, R. Tobin, and L. Bounegru, “A system for synergistically structuring news content from traditional media and the blogosphere,” Proc. E-Challenges Conference, Florence, Italy, 2011.
- G. Galatas, G. Potamianos, A. Papangelis, and F. Makedon, “Audio visual speech recognition in noisy visual environments,” Proc. Int. Conf. Pervasive Technologies Related to Assistive Environments (PETRA), Crete, Greece, 2011.
2010
- A. Waibel, R. Stiefelhagen, R. Carlson, J. Casas, J. Kleindienst, L. Lamel, O. Lanz, D. Mostefa, M. Omologo, F. Pianesi, L. Polymenakos, G. Potamianos, J. Soldatos, G. Sutschet, and J. Terken, “Computers in the human interaction loop,” Handbook of Ambient Intelligence and Smart Environments, H. Nakashima, H. Aghajan, and J.C. Augusto (Eds.), Part IX, pp. 1071–1116, Springer, 2010.
- L.-H. Kim, M. Hasegawa-Johnson, G. Potamianos, and V. Libal, “Joint estimation of DOA and speech based on EM beamforming,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 121–124, Dallas, TX, 2010.
2009
- P. Lucey, G. Potamianos, and S. Sridharan, “Visual speech recognition across multiple views,” Visual Speech Recognition: Lip Segmentation and Mapping, A. Wee-Chung Liew and S. Wang (Eds.), ch. X, pp. 294–325, Information Science Reference (IGI), 2009.
- G. Potamianos, L. Lamel, M. Wolfel, J. Huang, E. Marcheret, C. Barras, X. Zhu, J. McDonough, J. Hernando, D. Macho, and C. Nadeu, “Automatic speech recognition,” Computers in the Human Interaction Loop, A. Waibel and R. Stiefelhagen (Eds.), ch. 6, pp. 43–59, Springer, 2009.
- K. Bernardin, R. Stiefelhagen, A. Pnevmatikakis, O. Lanz, A. Brutti, J.R. Casas, and G. Potamianos, “Person tracking,” Computers in the Human Interaction Loop, A. Waibel and R. Stiefelhagen (Eds.), ch. 3, pp. 11–22, Springer, 2009.
- H. Meng, S. Oviatt, G. Potamianos, and G. Rigoll, “Introduction to the special issue on multimodal processing in speech-based interactions,” IEEE Transactions on Audio, Speech, and Language Processing, vol. 17, no. 3, pp. 409–410, 2009.
- K. Kumar, J. Navratil, E. Marcheret, V. Libal, and G. Potamianos, “Robust audio-visual speech synchrony detection by generalized bimodal linear prediction,” Proc. Conf. Int. Speech Comm. Assoc. (Interspeech), pp. 2251–2254, Brighton, United Kingdom, 2009.
- K. Kumar, J. Navratil, E. Marcheret, V. Libal, G. Ramaswamy, and G. Potamianos, “Audio-visual speech synchronization detection using a bimodal linear prediction model,” Proc. IEEE Comp. Soc. Works. Biometrics, Held in Conjunction with CVPR, Miami Beach, FL, 2009.
- V. Libal, B. Ramabhadran, N. Mana, F. Pianesi, P. Chippendale, O. Lanz, and G. Potamianos, “Multimodal classification of activities of daily living inside smart homes,” Int. Works. Ambient Assisted Living (IWAAL), LNCS vol. 5518, Part II, pp. 687–694, Salamanca, Spain, 2009.
- X. Zhuang, J. Huang, G. Potamianos, and M. Hasegawa-Johnson, “Acoustic fall detection using Gaussian mixture models and GMM supervectors,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 69–72, Taipei, Taiwan, 2009.
- J. Huang, X. Zhuang, V. Libal, and G. Potamianos, “Long-time span acoustic activity analysis from far-field sensors in smart homes,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 4173–4176, Taipei, Taiwan, 2009.
- S.M. Chu, V. Goel, E. Marcheret, and G. Potamianos, Method for Likelihood Computation in Multi-Stream HMM Based Speech Recognition, Patent No.: US007480617B2, Jan. 20, 2009.
2008
- G. Potamianos, J. Huang, E. Marcheret, V. Libal, R. Balchandran, M. Epstein, L. Seredi, M. Labsky, L. Ures, M. Black, and P. Lucey, “Far-field multimodal speech processing and conversational interaction in smart spaces,” Proc. Joint Work. Hands-Free Speech Communication and Microphone Arrays (HSCMA), Trento, Italy, 2008.
- A Leone, G. Diraco, C. Distante, P. Siciliano, M. Grassi, A. Lombardi, G. Rescio, P. Malcovati, M. Malfatti, L. Gonzo, V. Libal, J. Huang, and G. Potamianos, “A multi-sensor approach for people fall detection in home environment,” Proc. 10th Europ. Conf. Computer Vision (ECCV), Marseille, France, 2008.
- P. Lucey, G. Potamianos, and S. Sridharan, “Patch-based analysis of visual speech from multiple views,” Proc. Int. Conf. Auditory-Visual Speech Process. (AVSP), pp. 69–73, Tangalooma, Australia, 2008.
- R. Balchandran, M. Epstein, G. Potamianos, and L. Seredi, “A multi-modal spoken dialog system for interactive TV,” Proc. Int. Conf. Multimodal Interfaces (ICMI) – Demo Papers, pp. 191–192, Chania, Greece, 2008.
- M. Grassi, A. Lombardi, G. Rescio, P. Malcovati, A Leone, G. Diraco, C. Distante, P. Siciliano, M. Malfatti, L. Gonzo, V. Libal, J. Huang, and G. Potamianos, “A hardware-software framework for high-reliability people fall detection,” Proc. 7th IEEE Conf. on Sensors (SENSORS), pp. 1328–1331, Lecce, Italy, 2008.
- A. Tyagi, J.W. Davis, and G. Potamianos, “Steepest descent for efficient covariance tracking,” Proc. IEEE Work. Motion and Video Computing (WMVC), Copper Mountain, Colorado, 2008.
- J. Huang, E. Marcheret, K. Visweswariah, and G. Potamianos, “The IBM RT07 evaluation systems for speaker diarization on lecture meetings,” in Multimodal Technologies for Perception of Humans: International Evaluation Workshops CLEAR 2007 and RT 2007, Baltimore, Maryland, May 2007, LNCS vol. 4625, pp. 497–508, Springer, Berlin, 2008.
- J. Huang, E. Marcheret, K. Visweswariah, V. Libal, and G. Potamianos, “The IBM Rich Transcription Spring 2007
speech-to-text systems for lecture meetings,” in Multimodal Technologies for Perception of Humans: International Evaluation Workshops CLEAR 2007 and RT 2007, Baltimore, Maryland, May 2007, LNCS vol. 4625, pp. 429–441, Springer, Berlin, 2008. - S. Deligne, C.V. Neti, and G. Potamianos, Audio-Visual Codebook Dependent Cepstral Normalization, Patent No.: US007319955B2, Jan. 15, 2008.
2007
- D. Mostefa, N. Moreau, K. Choukri, G. Potamianos, S.M. Chu, A. Tyagi, J.R. Casas, J. Turmo, L. Christoforetti, F. Tobia, A. Pnevmatikakis, V. Mylonakis, F. Talantzis, S. Burger, R. Stiefelhagen, K. Bernardin, and C. Rochet, “The CHIL audiovisual corpus for lecture and meeting analysis inside smart rooms,” Journal of Language Resources and Evaluation, vol. 41, pp. 389–407, 2007.
- Z. Zhang, G. Potamianos, A.W. Senior, and T.S. Huang, “Joint face and head tracking inside multi-camera smart rooms,” Signal, Image and Video Processing, vol. 1, pp. 163–178, 2007.
- V. Libal, J. Connell, G. Potamianos, and E. Marcheret, “An embedded system for in-vehicle visual speech activity detection,” Proc. Int. Work. Multimedia Signal Process. (MMSP), pp. 255–258, Chania, Greece, 2007.
- P. Lucey, G. Potamianos, and S. Sridharan, “A unified approach to multi-pose audio-visual ASR,” Proc. Conf. Int. Speech Comm. Assoc. (Interspeech), pp. 650–653, Antwerp, Belgium, 2007.
- J. Huang, E. Marcheret, K. Visweswariah, V. Libal, and G. Potamianos, “Detection, diarization, and transcription of far-field lecture speech,” Proc. Conf. Int. Speech Comm. Assoc. (Interspeech), pp. 2161–2164, Antwerp, Belgium, 2007.
- P. Lucey, G. Potamianos, and S. Sridharan, “Pose-invariant audio-visual automatic speech recognition,” Proc. Work. Audio-Visual Speech Process. (AVSP), pp. 176–180, Hilvarenbeek, The Netherlands, 2007.
- A. Tyagi, M. Keck, J.W. Davis, and G. Potamianos, “Kernel-based 3D tracking,” Proc. IEEE Int. Work. Visual Surveillance (VS/CVPR), Minneapolis, Minnesota, 2007.
- A. Tyagi, G. Potamianos, J.W. Davis, and S.M. Chu, “Fusion of multiple camera views for kernel-based 3D tracking,” Proc. IEEE Work. Motion and Video Computing (WMVC), Austin, Texas, 2007.
- E. Marcheret, V. Libal, and G. Potamianos, “Dynamic stream weight modeling for audio-visual speech recognition,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), vol. 4, pp. 945–948, Honolulu, Hawaii, 2007.
- J.H. Connell, N. Haas, E. Marcheret, C.V. Neti, and G. Potamianos, Audio-Only Backoff in Audio-Visual Speech Recognition System, Patent No.: US007251603B2, July 31, 2007.
- U.V. Chaudhari, C. Neti, G. Potamianos, and G.N. Ramaswamy, Automated Decision Making Using Time-Varying Stream Reliability Prediction, Patent No.: US007228279B2, June 5, 2007.
2006
- G. Potamianos, “Audio-visual speech recognition,” Short Article, Encyclopedia of Language and Linguistics, Second Edition, (Speech Technology Section – Computer Understanding of Speech), K. Brown (Ed. In Chief), Elsevier, Oxford, United Kingdom, ISBN: 0-08-044299-4, vol. 11, pp. 800–805, 2006.
- G. Potamianos and P. Lucey, “Audio-visual ASR from multiple views inside smart rooms,” Proc. Int. Conf. Multisensor Fusion and Integration for Intelligent Systems (MFI), pp. 35–40, Heidelberg, Germany, 2006.
- Z. Zhang, G. Potamianos, S.M. Chu, J. Tu, and T.S. Huang, “Person tracking in smart rooms using dynamic programming and adaptive subspace learning,” Proc. Int. Conf. Multimedia Expo. (ICME), pp. 2061–2064, Toronto, Canada, 2006.
- P. Lucey and G. Potamianos, “Lipreading using profile versus frontal views,” Proc. IEEE Work. Multimedia Signal Process. (MMSP), pp. 24–28, Victoria, Canada, 2006.
- A.W. Senior, G. Potamianos, S. Chu, Z. Zhang, and A. Hampapur, “A comparison of multicamera person-tracking algorithms,” Proc. IEEE Int. Work. Visual Surveillance (VS/ECCV), Graz, Austria, 2006.
- G. Potamianos and Z. Zhang, “A joint system for single-person 2D-face and 3D-head tracking in CHIL seminars,” Multimodal Technologies for Perception of Humans: First Int. Eval. Work. on Classification of Events, Activities and Relationships, CLEAR 2006, R. Stiefelhagen and J. Garofolo (Eds.), LNCS vol. 4122, pp. 105–118, Southampton, United Kingdom, 2006.
- Z. Zhang, G. Potamianos, M. Liu, and T. Huang, “Robust multi-view multi-camera face detection inside smart rooms using spatio-temporal dynamic programming,” Proc. Int. Conf. Automatic Face and Gesture Recog. (FGR), Southampton, United Kingdom, 2006.
- E. Marcheret, G. Potamianos, K. Visweswariah, and J. Huang, “The IBM RT06s evaluation system for speech activity detection in CHIL seminars,” Proc. RT06s Evaluation Work. – held with Joint Work. on Multimodal Interaction and Related Machine Learning Algorithms (MLMI), S. Renals, S. Bengio, and J.G. Fiscus (Eds.), LNCS vol. 4299, pp. 323–335, Washington DC, 2006.
- J. Huang, M. Westphal, S. Chen, O. Siohan, D. Povey, V. Libal, A. Soneiro, H. Schulz, T. Ross, and G. Potamianos, “The IBM Rich Transcription Spring 2006 speech-to-text system for lecture meetings,” Proc. RT06s Evaluation Work.– held with Joint Work. on Multimodal Interaction and Related Machine Learning Algorithms (MLMI), S. Renals, S. Bengio, and J.G. Fiscus (Eds.), LNCS vol. 4299, pp. 432–443, Washington DC, 2006.
2005
- P.S. Aleksic, G. Potamianos, and A.K. Katsaggelos, “Exploiting visual information in automatic speech processing,” Handbook of Image and Video Processing, Second Edition, Al. Bovic (Ed.), ch. 10.8, pp. 1263–1289, Elsevier Academic Press, Burlington, MA, ISBN: 0-12-119792-1, 2005.
- G. Potamianos and P. Scanlon, “Exploiting lower face symmetry in appearance-based automatic speechreading,” Proc. Work. Audio-Visual Speech Process. (AVSP), pp. 79–84, Vancouver Island, Canada, 2005.
- S.M. Chu, E. Marcheret, and G. Potamianos, “Automatic speech recognition and speech activity detection in the CHIL smart room,” Proc. Joint Work. on Multimodal Interaction and Related Machine Learning Algorithms (MLMI), LNCS vol. 3869, pp. 332–343, Edinburgh, United Kingdom, 2005.
- Z. Zhang, G. Potamianos, A. Senior, S. Chu, and T. Huang, “A joint system for person tracking and face detection,” Proc. Int. Work. Human-Computer Interaction (ICCV 2005 Work. on HCI), pp. 47–59, Beijing, China, 2005.
- E. Marcheret, K. Visweswariah, and G. Potamianos, “Speech activity detection fusing acoustic phonetic and energy features,” Proc. Europ. Conf. Speech Comm. Technol. (Interspeech), pp. 241–244, Lisbon, Portugal, 2005.
- J. Jiang, G. Potamianos, and G. Iyengar, “Improved face finding in visually challenging environments,” Proc. Int. Conf. Multimedia Expo. (ICME), Amsterdam, The Netherlands, 2005.
- D. Macho, J. Padrell, A. Abad, C. Nadeu, J. Hernando, J. McDonough, M. Wolfel, U. Klee, M. Omologo, A. Brutti, P. Svaizer, G. Potamianos, and S.M. Chu, “Automatic speech activity detection, source localization, and speech recognition on the CHIL seminar corpus,” Proc. Int. Conf. Multimedia Expo. (ICME), Amsterdam, The Netherlands, 2005.
2004
- J. Huang, G. Potamianos, J. Connell, and C. Neti. “Audio-visual speech recognition using an infrared headset,” Speech Communication, vol. 44, no. 4, pp. 83–96, 2004.
- P. Scanlon, G. Potamianos, V. Libal, and S.M. Chu, “Mutual information based visual feature selection for lipreading,” Proc. Int. Conf. Spoken Lang. Process. (ICSLP), Jeju Island, Korea, 2004.
- E. Marcheret, S.M. Chu, V. Goel, and G. Potamianos, “Efficient likelihood computation in multi-stream HMM based audio-visual speech recognition,” Proc. Int. Conf. Spoken Lang. Process. (ICSLP), Jeju Island, Korea, 2004.
- G. Potamianos, C. Neti, J. Huang, J.H. Connell, S. Chu, V. Libal, E. Marcheret, N. Haas, and J. Jiang, “Towards practical deployment of audio-visual speech recognition,” Invited, Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), vol. 3, pp. 777–780, Montreal, Canada, 2004.
- J. Jiang, G. Potamianos, H. Nock, G. Iyengar, and C. Neti, “Improved face and feature finding for audio-visual speech recognition in visually challenging environments,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), vol. 5, pp. 873–876, Montreal, Canada, 2004.
- S.M. Chu, V. Libal, E. Marcheret, C. Neti, and G. Potamianos, “Multistage information fusion for audio-visual speech recognition,” Proc. Int. Conf. Multimedia Expo. (ICME), Taipei, Taiwan, 2004.
- P. de Cuetos, G.R. Iyengar, C.V. Neti, and G. Potamianos, System and Method for Microphone Activation Using Visual Speech Cues, Patent No.: US006754373B1, June 22, 2004.
2003
- G. Potamianos, C. Neti, G. Gravier, A. Garg, and A.W. Senior, “Recent advances in the automatic recognition of audio-visual speech,” Invited, Proceedings of the IEEE, vol. 91, no. 9, pp. 1306–1326, 2003.
- G. Potamianos, C. Neti, and S. Deligne, “Joint audio-visual speech processing for recognition and enhancement,” Proc. Work. Audio-Visual Speech Process. (AVSP), pp. 95–104, St. Jorioz, France, 2003.
- J. Huang, G. Potamianos, and C. Neti, “Improving audio-visual speech recognition with an infrared headset,” Proc. Work. Audio-Visual Speech Process. (AVSP), pp. 175–178, St. Jorioz, France, 2003.
- G. Potamianos and C. Neti, “Audio-visual speech recognition in challenging environments,” Proc. Europ. Conf. Speech Comm. Technol. (Eurospeech), pp. 1293–1296, Geneva, Switzerland, 2003.
- J.H. Connell, N. Haas, E. Marcheret, C. Neti, G. Potamianos, and S. Velipasalar, “A real-time prototype for small-vocabulary audio-visual ASR,” Proc. Int. Conf. Multimedia Expo. (ICME), vol. II, pp. 469-472, Baltimore, MD, 2003.
- U.V. Chaudhari, G.N. Ramaswamy, G. Potamianos, and C. Neti, “Information fusion and decision cascading for audio-visual speaker recognition based on time varying stream reliability prediction,” Proc. Int. Conf. Multimedia Expo. (ICME), pp. 9–12, Baltimore, MD, 2003.
- A. Garg, G. Potamianos, C. Neti, and T.S. Huang, “Frame-dependent multi-stream reliability indicators for audio-visual speech recognition,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), vol. I, pp. 24–27, Hong Kong, China, 2003.
- U.V. Chaudhari, G.N. Ramaswamy, G. Potamianos, and C. Neti, “Audio-visual speaker recognition using time-varying stream reliability prediction,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), vol. V, pp. 712–715, Hong Kong, China, 2003.
- E. Cosatto, H.P. Graf, G. Potamianos, and J. Schroeter, Audio-Visual Selection Process for the Synthesis of Photo-Realistic Talking-Head Animations, Patent No.: US006654018B1, Nov. 25, 2003.
2002
- C. Neti, G. Potamianos, J. Luettin, and E. Vatikiotis-Bateson, “Editorial of the special issue on joint audio-visual speech processing,” EURASIP Journal on Applied Signal Processing, vol. 2002, no. 11, 2002.
- S. Deligne, G. Potamianos, and C. Neti, “Audio-visual speech enhancement with AVCDCN (audio-visual codebook dependent cepstral normalization)”, Proc. Int. Conf. Spoken Lang. Process. (ICSLP), vol. 3, pp. 1449–1452, Denver, CO, 2002.
- R. Goecke, G. Potamianos, and C. Neti, “Noisy audio feature enhancement using audio-visual speech data,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 2025–2028, Orlando, FL, 2002.
- G. Gravier, S. Axelrod, G. Potamianos, and C. Neti, “Maximum entropy and MCE based HMM stream weight estimation for audio-visual ASR,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 853–856, Orlando, FL, 2002.
- G. Gravier, G. Potamianos, and C. Neti, “Asynchrony modeling for audio-visual speech recognition,” Proc. Human Lang. Techn. Conf. (HLT), pp. 1–6, San Diego, CA, 2002.
2001
- G. Potamianos, C. Neti, G. Iyengar, A.W. Senior, and A. Verma, “A cascade visual front end for speaker independent automatic speechreading,” International Journal of Speech Technology, Special Issue on Multimedia, vol. 4, pp. 193–208, 2001.
- G. Potamianos, C. Neti, G. Iyengar, and E. Helmuth, “Large-vocabulary audio-visual speech recognition by machines and humans,” Proc. Europ. Conf. Speech Comm. Technol. (Eurospeech), pp. 1027–1030, Aalborg, Denmark, 2001.
- G. Potamianos and C. Neti, “Automatic speechreading of impaired speech,” Proc. Work. Audio-Visual Speech Process. (AVSP), pp. 177–182, Aalborg, Denmark, 2001.
- G. Potamianos and C. Neti, “Improved ROI and within frame discriminant features for lipreading,” Proc. Int. Conf. Image Process. (ICIP), pp. 250–253, Thessaloniki, Greece, 2001.
- C. Neti, G. Potamianos, J. Luettin, I. Matthews, H. Glotin, and D. Vergyri, “Large-vocabulary audio-visual speech recognition: A summary of the Johns Hopkins Summer 2000 Workshop,” Proc. IEEE Work. Multimedia Signal Process. (MMSP), pp. 619–624, Cannes, France, 2001.
- G. Iyengar, G. Potamianos, C. Neti, T. Faruquie, and A. Verma, “Robust detection of visual ROI for automatic
speechreading,” Proc. IEEE Work. Multimedia Signal Process. (MMSP), pp. 79–84, Cannes, France, 2001. - I. Matthews, G. Potamianos, C. Neti, and J. Luettin, “A comparison of model and transform-based visual features for audio-visual LVCSR,” Proc. Int. Conf. Multimedia Expo. (ICME), Tokyo, Japan, 2001.
- G. Potamianos, J. Luettin, and C. Neti, “Hierarchical discriminant features for audio-visual LVCSR,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), vol. 1, pp. 165–168, Salt Lake City, UT, 2001.
- J. Luettin, G. Potamianos, and C. Neti, “Asynchronous stream modeling for large-vocabulary audio-visual speech recognition,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), vol. 1, pp. 169–172, Salt Lake City, UT, 2001.
- H. Glotin, D. Vergyri, C. Neti, G. Potamianos, and J. Luettin, “Weighting schemes for audio-visual fusion in speech recognition,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), vol. 1, pp. 173–176, Salt Lake City, UT, 2001.
2000
- G. Potamianos and C. Neti, “Stream confidence estimation for audio-visual speech recognition,” Proc. Int. Conf. Spoken Language Process. (ICSLP), vol. III, pp. 746–749, Beijing, China, 2000.
- C. Neti, G. Iyengar, G. Potamianos, A. Senior, and B. Maison, “Perceptual interfaces for information interaction: Joint processing of audio and visual information for human-computer interaction”, Proc. Int. Conf. Spoken Language Process. (ICSLP), vol III, pp. 11-14, Beijing, China, 2000.
- G. Potamianos, A. Verma, C. Neti, G. Iyengar, and S. Basu, “A cascade image transform for speaker independent automatic speechreading,” Proc. IEEE Int. Conf. Multimedia Expo. (ICME), vol. II, pp. 1097–1100, New York, NY, 2000.
- E. Cosatto, G. Potamianos, and H.P. Graf, “Audio-visual unit selection for the synthesis of photo-realistic talking-heads,” Proc. IEEE Int. Conf. Multimedia Expo. (ICME), vol. II, pp. 619–622, New York, NY, 2000.
- E. Cosatto, H.P. Graf, and G. Potamianos, Robust Multi-Modal Method for Recognizing Objects, Patent No.: US006118887A, Sep. 12, 2000.
Past Millennium
- G. Potamianos and A. Potamianos, “Speaker adaptation for audio-visual automatic speech recognition,” Proc. Europ. Speech Comm. Technol. (Eurospeech), Budapest, Hungary, vol. 3, pp. 1291–1294, 1999.
- G. Potamianos and F. Jelinek, “A study of n-gram and decision tree letter language modeling methods,” Speech Communication, vol. 24, no. 3, pp. 171–192, 1998.
- G. Potamianos and H.P. Graf, “Linear discriminant analysis for speechreading,” Proc. IEEE Work. Multimedia Signal Process. (MMSP), Los Angeles, CA, pp. 221–226, 1998.
- G. Potamianos, H.P. Graf, and E. Cosatto, “An image transform approach for HMM based automatic lipreading,” Proc. Int. Conf. Image Process. (ICIP), Chicago, IL, pp. 173–177, 1998.
- G. Potamianos and H.P. Graf, “Discriminative training of HMM stream exponents for audio-visual speech recognition,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), Seattle, WA, vol. 6, pp. 3733–3736, 1998.
- H.P. Graf, E. Cosatto, and G. Potamianos, “Machine vision of faces and facial features,” Proc. R.I.E.C. Int. Symp. Design Archit. Inform. Process. Systems Based Brain Inform. Princ., Sendai, Japan, pp. 48–53, 1998.
- G. Potamianos and J. Goutsias, “Stochastic approximation algorithms for partition function estimation of Gibbs random fields,” IEEE Transactions on Information Theory, vol. 43, no. 6, pp. 1948–1965, 1997.
- G. Potamianos, E. Cosatto, H.P. Graf, and D.B. Roe, “Speaker independent audio-visual database for bimodal ASR,” Proc. Europ. Tutorial Research Work. Audio-Visual Speech Process. (AVSP), Rhodes, Greece, pp. 65–68, 1997.
- H.P. Graf, E. Cosatto, and G. Potamianos, “Robust recognition of faces and facial features with a multi-modal system,” Proc. Int. Conf. Systems Man Cybern. (ICSMC), Orlando, FL, pp. 2034–2039, 1997.
- G. Potamianos, “Efficient Monte Carlo estimation of partition function ratios of Markov random field images,” Proc. Conf. Inform. Sci. Systems (CISS), Princeton, NJ, vol. II, pp. 1212–1215, 1996.
- G. Potamianos and J. Goutsias, “A unified approach to Monte Carlo likelihood estimation of Gibbs random field images,” Proc. Conf. Inform. Sci. Systems (CISS), Princeton, NJ, vol. I, pp. 84–90, 1994.
- G.G. Potamianos and J. Goutsias, “Partition function estimation of Gibbs random field images using Monte Carlo simulations,” IEEE Transactions on Information Theory, vol. 39, no. 4, pp. 1322–1332, 1993.
- G. Potamianos and J. Goutsias, “An analysis of Monte Carlo methods for likelihood estimation of Gibbsian images,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), Minneapolis, MN, vol. V, pp. 519–522, 1993.
- G. Potamianos and J. Goutsias, “On computing the likelihood function of partially observed Markov random field images using Monte Carlo simulations,” Proc. Conf. Inform. Sci. Systems (CISS), Princeton, vol. I, pp. 357–362, 1992.
- G. Potamianos and J. Goutsias, “A novel method for computing the partition function of Markov random field images using Monte Carlo simulations,” Proc. Int. Conf. Acoust. Speech Signal Process. (ICASSP), Toronto, Canada, vol. 4, pp. 2325–2328, 1991.
- G. Potamianos and J. Diamessis, “Frequency sampling design of 2-D IIR filters using continued fractions,” Proc. Int. Symp. Circuits Systems (ISCAS), New Orleans, LA, pp. 2454–2457, 1990.
- J. Diamessis and G. Potamianos, “A novel method for designing IIR filters with nonuniform samples,” Proc. Conf. Inform. Sci. Systems (CISS), Princeton, NJ, vol. 1, pp. 192–195, 1990.
- J. Diamessis and G. Potamianos, “Modeling unequally spaced 2-D discrete signals by rational functions,” Proc. Int. Symp. Circuits Systems (ISCAS), Portland, OR, pp. 1508–1511, 1989.
[Last update: 04/04/2024]
Research Interests: Multimodal speech processing with applications to human-computer interaction and ambient intelligence. Particular emphasis on audio-visual speech processing, far-field automatic speech recognition, multichannel acoustic scene analysis, multimedia signal processing and fusion, sign language recognition, computer vision for human tracking and activity recognition, incorporating deep learning methods.
Undergraduate Courses:Postgraduate Courses of PGS in "Science and Technology of ECE":
Ph.D. Graduates:
Ph.D. Candidates:
Katerina Papadimitriou
Supervisor | Gerasimos Potamianos, Associate Professor |
---|---|
Phd Thesis Title | Automatic Sign Language Recognition Algorithms and their Incorporation into a Greek Sign Language Educational Platform |
Year | 2024 |
Spyridon Thermos
Supervisor | Gerasimos Potamianos, Associate Professor |
---|---|
Phd Thesis Title | A Deep Learning Approach to 2D/3D Object Affordance Understanding |
Year | 2020 |
Alexandros Koumparoulis
Supervisor | Gerasimos Potamianos, Associate Professor |
---|---|
Phd Thesis Title | Audio Visual Speech Processing in the Wild |
alkoumpa@e-ce.uth.gr |
Nikolaos Tsikrikonis
Supervisor | Gerasimos Potamianos, Associate Professor |
---|---|
Phd Thesis Title | Conformal Microstrip Antenna with Beam Steering |
ntsikrikonis@e-ce.uth.gr |