MANIFOLD
BrowseUS ElectionNewsAbout
Is code-davinci-002 just the largest non-GPT-4 model in the GPT-4 scaling law experiment?
Mini
3
Ṁ35
Feb 17
44%
chance
1D
1W
1M
ALL

https://arxiv.org/pdf/2303.08774.pdf


Resolves Yes/No when I have >90% confidence.

Get
Ṁ1,000
and
S1.00
1 Comment
Sort by:

It's 4 orders of magnitude less compute than GPT-4! GPT-3 was 2 OOMs less compute.

Related questions

Will GPT-4 be trained (roughly) compute-optimally using the best-known scaling laws at the time?
30% chance
What hardware will GPT-5 be trained on?
What will be true about GPT-5?
Will the performance jump from GPT4->GPT5 be less than the one from GPT3->GPT4?
71% chance
Will GPT-5 have over 1 trillion parameters?
87% chance
GPT-4 #5: Will GPT-4 be a dense model?
1% chance
GPT-5 trained with >=24k GPUs?
82% chance
How many parameters does GPT4o have?
Is GPT-4.5 the base model for o3?
4% chance
Will GPT-5 have more than 10 trillion parameters?
31% chance

Related questions

Will GPT-4 be trained (roughly) compute-optimally using the best-known scaling laws at the time?
30% chance
GPT-4 #5: Will GPT-4 be a dense model?
1% chance
What hardware will GPT-5 be trained on?
GPT-5 trained with >=24k GPUs?
82% chance
What will be true about GPT-5?
How many parameters does GPT4o have?
Will the performance jump from GPT4->GPT5 be less than the one from GPT3->GPT4?
71% chance
Is GPT-4.5 the base model for o3?
4% chance
Will GPT-5 have over 1 trillion parameters?
87% chance
Will GPT-5 have more than 10 trillion parameters?
31% chance
Terms & Conditions•Privacy Policy•Sweepstakes Rules
BrowseElectionNewsAbout