no code implementations • 19 Jul 2023 • Javad Peymanfard, Vahid Saeedi, Mohammad Reza Mohammadi, Hossein Zeinali, Nasser Mozayani
We evaluate our approach on various tasks, including word-level and sentence-level lip reading, and audiovisual speech recognition using the Arman-AV dataset, a largescale Persian corpus.
no code implementations • 8 Apr 2023 • Javad Peymanfard, Ali Lashini, Samin Heydarian, Hossein Zeinali, Nasser Mozayani
Lip-reading has made impressive progress in recent years, driven by advances in deep learning.
no code implementations • 21 Jan 2023 • Javad Peymanfard, Samin Heydarian, Ali Lashini, Hossein Zeinali, Mohammad Reza Mohammadi, Nasser Mozayani
In addition, we have proposed a technique to detect visemes (a visual equivalent of a phoneme) in Persian.
Audio-Visual Speech Recognition Automatic Speech Recognition +5
1 code implementation • 11 May 2022 • Hossein Parineh, Nasser Mozayani
In this paper we proposed a new approach for protein structure prediction by using agent-based modeling (ABM) in two dimensional hydrophobic-hydrophilic model.
no code implementations • 10 Apr 2021 • Javad Peymanfard, Mohammad Reza Mohammadi, Hossein Zeinali, Nasser Mozayani
Lip-reading is the operation of recognizing speech from lip movements.
no code implementations • 17 Feb 2019 • Babak Badnava, Mona Esmaeili, Nasser Mozayani, Payman Zarkesh-Ha
Among the literature of both the transfer learning and the potential-based reward shaping, a subject that has never been addressed is the knowledge gathered during the learning process itself.
no code implementations • 17 Dec 2016 • Sajad Mousavi, Michael Schukat, Enda Howley, Ali Borji, Nasser Mozayani
Bottom-Up (BU) saliency models do not perform well in complex interactive environments where humans are actively engaged in tasks (e. g., sandwich making and playing the video games).