Skip to content

aljubrmj/CS388-Final-Project

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Contributors Forks Stargazers Issues LinkedIn


University of Texas at Austin CS388 Class Final Project

This project explores the application of pretrained contextual embedding (PCE) models in comparison to the bidirectional attentive reader (baseline) implemented by the CS-388 staff. Particularly, I applied DistilBERT (Sanh et al., 2019) as it is 60% faster to train and retains 97% of the base BERT performance. Model comparison was conducted on the SQuAD 1.1 (Rajpurkar et al., 2016) and SQuAD adverserial datasets (Jia and Liang, 2017). The HuggingFace implementation of DistilBERT was utilized, hence I first had to ensure the SQuAD datasets used match to that provided by the class staff. Whereas the baseline model achieved 60.63% and 46.76% F1 scores on the SQuAD 1.1 and adverserial SQuAD datasets, respectively, the fine-tuned DistilBERT model achieved 85.4% and 69.4%, respectively. Analysis shows that both models struggle the most with ”why” questions, and achieve lower F1 scores with increasing answer lengths.
Explore the docs »

Report Bug · Request Feature

License

Distributed under the MIT License. See LICENSE for more information.

Contact

Name: @MJAljubran - [email protected]

Project Link: https://github.com/aljubrmj/CS388-Final-Project

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published