WNLI
4 papers with code • 0 benchmarks • 0 datasets
Benchmarks
These leaderboards are used to track progress in WNLI
Most implemented papers
A Hybrid Neural Network Model for Commonsense Reasoning
An HNN consists of two component models, a masked language model and a semantic similarity model, which share a BERT-based contextual encoder but use different model-specific input and output layers.
A Surprisingly Robust Trick for Winograd Schema Challenge
The Winograd Schema Challenge (WSC) dataset WSC273 and its inference counterpart WNLI are popular benchmarks for natural language understanding and commonsense reasoning.
WikiCREM: A Large Unsupervised Corpus for Coreference Resolution
We use a language-model-based approach for pronoun resolution in combination with our WikiCREM dataset.
Time Travel in LLMs: Tracing Data Contamination in Large Language Models
To estimate contamination of individual instances, we employ "guided instruction:" a prompt consisting of the dataset name, partition type, and the random-length initial segment of a reference instance, asking the LLM to complete it.