Test-Time Fine-Tuning with SIFT + Llama-3.2 (3B) | 0.557 | Efficiently Learning at Test-Time: Active Fine-Tuning of LLMs | - |
GPT-3 Davinci 175B (pre-trained) | 0.7177 | The Pile: An 800GB Dataset of Diverse Text for Language Modeling | |
Test-Time Fine-Tuning with SIFT + GPT-2 (124M) | 0.862 | Efficiently Learning at Test-Time: Active Fine-Tuning of LLMs | - |
Larger Transformer 771M (fine-tuned) | - | Need a Small Specialized Language Model? Plan Early! | - |
Larger Transformer 771M (pre-trained) | - | Need a Small Specialized Language Model? Plan Early! | - |
Test-Time Fine-Tuning with SIFT + GPT-2 (774M) | 0.762 | Efficiently Learning at Test-Time: Active Fine-Tuning of LLMs | - |
GPT-2 Medium 355M (pre-trained) | 1.0928 | The Pile: An 800GB Dataset of Diverse Text for Language Modeling | |
GPT-2 XL 1.5B (pre-trained) | 1.0468 | The Pile: An 800GB Dataset of Diverse Text for Language Modeling | |