Skip to content

Understanding and Evaluating Commonsense Reasoning in Transformer-based Architectures. Paper Presented at IEEE ICCCNT '21

License

Notifications You must be signed in to change notification settings

manavgakhar/ComVE

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Understanding and Evaluating Commonsense Reasoning in Transformer-based Architectures PAPER

Ascertaining the reason for and identifying the differences between sensical and nonsensical statements is a task that humans are naturally adept at. NLP methods, however, might struggle with such a task. This paper aims to evaluate various transformer-based architectures on common sense validation and explanation task, methods to improve their performance, as well as interpreting how fine-tuned language models perform such tasks, using the attention distribution of sentences at inference time. The tasks entail identifying the nonsensical statement from a given pair of similar statements (validation), followed by selecting the correct reason for its nonsensical nature(explanation). An accuracy of 83.90 and 88.42 is achieved on the respective tasks, using RoBERTa (large) language model fine-tuned on the datasets using the ULMFiT training approach.

Inter-task transfer learning framework

image

Explaining results with attention distribution

image

Results

image

About

Understanding and Evaluating Commonsense Reasoning in Transformer-based Architectures. Paper Presented at IEEE ICCCNT '21

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published