Publicaciones de conferencias: comunicaciones, ponencias, pósters, etc (conferenceObject). 2023

Interpreting Sign Language Recognition Using Transformers and MediaPipe Landmarks

Archivo Digital UPM
oai:oa.upm.es:76743
Archivo Digital UPM
  • Luna Jiménez, Cristina
  • Gil Martín, Manuel
  • Kleinlein, Ricardo
  • San Segundo Hernández, Rubén
  • Fernández Martínez, Fernando
Sign Language Recognition (SLR) is a challenging task that aims to bridge the communication gap between the deaf and hearing communities. In recent years, deep learning-based approaches have shown promising results in SLR. However, the lack of interpretability remains a significant challenge. In this paper, we seek to understand which hand and pose MediaPipe Landmarks are deemed the most important for prediction as estimated by a Transformer model. We propose to embed a learnable array of parameters into the model that performs an element-wise multiplication of the inputs. This learned array highlights the most informative input features that contributed to solve the recognition task. Resulting in a human-interpretable vector that lets us interpret the model predictions. We evaluate our approach on public datasets called WLASL100 (SRL) and IPNHand (gesture recognition). We believe that the insights gained in this way could be exploited for the development of more efficient SLR pipelines.
 

DOI: https://oa.upm.es/76743/
Archivo Digital UPM
oai:oa.upm.es:76743

HANDLE: https://oa.upm.es/76743/
Archivo Digital UPM
oai:oa.upm.es:76743
 
Ver en: https://oa.upm.es/76743/
Archivo Digital UPM
oai:oa.upm.es:76743

1033