Download Full Text (4.6 MB)


Technological University Dublin


Gender stereotypes are perceptions about the typical physical, emotional, and social characteristics of individuals. Unlike gender bias which can result in the systematic and unfair treatment of individuals based on their gender, gender stereotypes do not always perpetuate a negative impact. Hence, there is no solid definition that frames what is considered a gender-stereotype in text. In addition, there is also a lack of labelled gender-stereotype datasets. This has led to most of the work in literature being about gender bias and not gender stereotypes. Therefore, in our research, we frame a clear definition of what constitutes a gender-stereotype in text. And using fairness benchmark datasets that were not originally intended for gender-stereotype studies, we propose an end-to-end approach to automatically identify language features that influence a gender-stereotypical sentence or text. We do this using the learned attention weights of a model. Our evaluations on the features that the model was using to predict gender-stereotypes aligned with our definition of gender-stereotypes.

Publication Date



Attention, gender-stereotypes, explainable ai


Computer Sciences


First Annual Teaching and Research Showcase 2023


Creative Commons License

Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License
This work is licensed under a Creative Commons Attribution-NonCommercial-Share Alike 4.0 International License.

Attention-based Gender-Stereotype Detection



To view the content in your browser, please download Adobe Reader or, alternately,
you may Download the file to your hard drive.

NOTE: The latest versions of Adobe Reader do not support viewing PDF files within Firefox on Mac OS and if you are using a modern (Intel) Mac, there is no official plugin for viewing PDF files within the browser window.