We introduce phi-1, a new large language model for code, with significantly smaller size than competing models: phi-1 is a Transformer-based model with 1.3B parameters, trained for 4 days on 8 A100s, ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results