Published on Jun 20, 2023
· Submitted by akhaliq on Jun 21, 2023 #1 Paper of the day Authors:,
,
,
,
,
,
,
,
,
,
AbstractA new compact Transformer-based large language model for code, phi-1, achieves high accuracy on coding benchmarks despite having fewer parameters than competing models.
AI-generated summary
We introduce phi-1, a new large language model for code, with significantly smaller size than competing models: phi-1 is a Transformer-based model with 1.3B parameters, trained for 4 days on 8 A100s, using a selection of ``textbook quality" data from the web (6B tokens) and synthetically generated textbooks and exercises with GPT-3.5 (1B tokens). Despite this small scale, phi-1 attains pass@1 accuracy 50.6% on HumanEval and 55.5% on MBPP. It also displays surprising emergent properties compared to phi-1-base, our model before our finetuning stage on a dataset of coding exercises, and phi-1-small, a smaller model with 350M parameters trained with the same pipeline as phi-1 that still achieves 45% on HumanEval.
Models citing this paper 10 microsoft/phi-1Text Generation • 1B • Updated Apr 29, 2024 • 9.61k • 213
kenhktsui/llm-data-textbook-quality-fasttext-classifier-v2Text Classification • Updated 20 days ago • 2.32k • 28
michaelfeil/ct2fast-phi-1Text Generation • Updated Nov 30, 2023 • 9
OpenNMT/phi-1-ct2-int8Text Generation • Updated Nov 30, 2023 • 8
Browse 10 models citing this paper Datasets citing this paper 14 HuggingFaceTB/cosmopediaViewer • Updated Aug 12, 2024 • 31.1M • 4.58k • 628
maywell/korean_textbooksViewer • Updated Jan 10, 2024 • 4.42M • 1.51k • 116
fzmnm/TinyEncyclopedias-ChineseViewer • Updated Aug 1, 2024 • 11.1k • 711 • 1
goendalf666/sales-conversationsViewer • Updated Oct 4, 2023 • 3.41k • 193 • 40
Browse 14 datasets citing this paper Spaces citing this paper 52 Collections including this paper 34 Browse 34 collections that include this paperRetroSearch is an open source project built by @garambo | Open a GitHub Issue
Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo
HTML:
3.2
| Encoding:
UTF-8
| Version:
0.7.4