-
How Do Large Language Models Acquire Factual Knowledge During Pretraining?
Paper • 2406.11813 • Published • 31 -
From RAGs to rich parameters: Probing how language models utilize external knowledge over parametric information for factual queries
Paper • 2406.12824 • Published • 21 -
Tokenization Falling Short: The Curse of Tokenization
Paper • 2406.11687 • Published • 16 -
Iterative Length-Regularized Direct Preference Optimization: A Case Study on Improving 7B Language Models to GPT-4 Level
Paper • 2406.11817 • Published • 13
![Mayank Sharma's picture](https://cdn-avatars.huggingface.co/v1/production/uploads/noauth/xWH9VCl_uz6IFflBMgSY5.jpeg)
Mayank Sharma
mayank1729
AI & ML interests
None yet
Organizations
None yet
Collections
1
models
None public yet
datasets
None public yet