Interpretable machine learning in natural language processing for misinformation data

Show simple item record

dc.contributor.advisor Marivate, Vukosi
dc.contributor.postgraduate Nkalashe, Yolanda
dc.date.accessioned 2023-10-09T08:02:21Z
dc.date.available 2023-10-09T08:02:21Z
dc.date.created 2023-04
dc.date.issued 2022-11
dc.description Mini Dissertation (MIT (Big Data Science))--University of Pretoria, 2022. en_US
dc.description.abstract The interpretability of models has been one of the focal research topics in the machine learning community due to a rise in the use of black box models and complex state-of-the-art models [6]. Most of these models are debugged through trial and error, based on end-to-end learning [7, 48]. This creates some uneasiness and distrust among the end-user consumers of the models, which has resulted in limited use of black box models in disciplines where explainability is required [33]. However, alternative models, ”white-box models,” come with a trade-off of accuracy and predictive power [7]. This research focuses on interpretability in natural language processing for misinformation data. First, we explore example-based techniques through prototype selection to determine if we can observe any key behavioural insights from a misinformation dataset. We use four prototype selection techniques: Clustering, Set Cover, MMD-critic, and Influential examples. We analyse the quality of each technique’s prototype set and use two prototype sets that have the optimal quality to further process for word analysis, linguistic characteristics, and together with the LIME technique for interpretability. Secondly, we compare if there are any critical insights in the South African disinformation context. en_US
dc.description.availability Unrestricted en_US
dc.description.degree MIT (Big Data Science) en_US
dc.description.department Computer Science en_US
dc.identifier.citation * en_US
dc.identifier.other A2023 en_US
dc.identifier.uri http://hdl.handle.net/2263/92768
dc.language.iso en en_US
dc.publisher University of Pretoria
dc.rights © 2021 University of Pretoria. All rights reserved. The copyright in this work vests in the University of Pretoria. No part of this work may be reproduced or transmitted in any form or by any means, without the prior written permission of the University of Pretoria.
dc.subject UCTD en_US
dc.subject Disinformation en_US
dc.subject Interpretability en_US
dc.subject Prototypes en_US
dc.subject Example-based en_US
dc.subject Interpretable Machine Learning en_US
dc.subject Natural Language Processing en_US
dc.title Interpretable machine learning in natural language processing for misinformation data en_US
dc.type Mini Dissertation en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record