Mostrar el registro sencillo del ítem
Understanding stance classification of BERT models : an attention-based mechanism
dc.contributor.advisor | Becker, Karin | pt_BR |
dc.contributor.author | Sáenz, Carlos Abel Córdova | pt_BR |
dc.date.accessioned | 2022-08-20T04:55:52Z | pt_BR |
dc.date.issued | 2022 | pt_BR |
dc.identifier.uri | http://hdl.handle.net/10183/247549 | pt_BR |
dc.description.abstract | BERT produces state-of-the-art solutions for many natural language processing tasks at the cost of interpretability. As works discuss the value of BERT’s attention weights to this purpose, we contribute with an attention-based interpretability framework to identify the most influential words for stance classification using BERT-based models. Unlike related work, we develop a broader level of interpretability focused on the overall model behavior instead of single instances. We aggregate tokens’ attentions into words’ attention weights that are more meaningful and can be semantically related to the domain. We propose attention metrics to assess words’ influence in the correct classification of stances. We use three case studies related to COVID-19 to assess the proposed framework in a broad experimental setting encompassing six datasets and four BERT pre-trained models for Portuguese and English languages, resulting in sixteen stance classification models. Through establishing five different research questions, we obtained valuable insights on the usefulness of attention weights to interpret stance classification that allowed us to generalize our findings. Our results are independent of a particular pre-trained BERT model and comparable to those obtained using an alternative baseline method. High attention scores improve the probability of finding words that positively impact the model performance and influence the correct classification (up to 82% of identified influential words contribute to correct predictions). The influential words represent the domain and can be used to identify how the model leverages the arguments expressed to predict a stance. | en |
dc.format.mimetype | application/pdf | pt_BR |
dc.language.iso | por | pt_BR |
dc.rights | Open Access | en |
dc.subject | Interpretability | en |
dc.subject | Processamento de linguagem natural | pt_BR |
dc.subject | COVID-19 | pt_BR |
dc.subject | BERT | en |
dc.subject | Attention | en |
dc.subject | Stance classification | en |
dc.title | Understanding stance classification of BERT models : an attention-based mechanism | pt_BR |
dc.type | Dissertação | pt_BR |
dc.identifier.nrb | 001147927 | pt_BR |
dc.degree.grantor | Universidade Federal do Rio Grande do Sul | pt_BR |
dc.degree.department | Instituto de Informática | pt_BR |
dc.degree.program | Programa de Pós-Graduação em Computação | pt_BR |
dc.degree.local | Porto Alegre, BR-RS | pt_BR |
dc.degree.date | 2022 | pt_BR |
dc.degree.level | mestrado | pt_BR |
Ficheros en el ítem
Este ítem está licenciado en la Creative Commons License
-
Ciencias Exactas y Naturales (5129)Computación (1764)