模型:
neulab/gpt2-large-finetuned-wikitext103
This is a gpt2-large model, finetuned on the Wikitext-103 dataset.
It achieves a perplexity of 10.56 using a "sliding window" context, using the run_clm.py script at https://github.com/neulab/knn-transformers .
| Base LM: | distilgpt2 | gpt2 | 
|---|---|---|
| base perplexity | 18.25 | 14.84 | 
| + kNN-LM | 15.03 | 12.57 | 
| + RetoMaton | 14.70 | 12.46 | 
This model was released as part of the paper "Neuro-Symbolic Language Modeling with Automaton-augmented Retrieval" (ICML'2022).
For more information, see: https://github.com/neulab/knn-transformers
If you use this model, please cite:
@inproceedings{alon2022neuro,
  title={Neuro-Symbolic Language Modeling with Automaton-augmented Retrieval},
  author={Alon, Uri and Xu, Frank and He, Junxian and Sengupta, Sudipta and Roth, Dan and Neubig, Graham},
  booktitle={International Conference on Machine Learning},
  pages={468--485},
  year={2022},
  organization={PMLR}
}