Will a LLM trained with FP4 have competitive performance in 2 years time?
Plus
14
Ṁ1401Jan 21
23%
chance
1D
1W
1M
ALL
"Currently, the technology for 4-bit training does not exists, but research looks promising and I expect the first high performance FP4 Large Language Model (LLM) with competitive predictive performance to be trained in 1-2 years time." (see: https://timdettmers.com/2023/01/16/which-gpu-for-deep-learning/)
Granted, the model must be open source for us to know, so the market will resolve based on publicly available information.
Get
1,000
and1.00
Sort by:
This seems important @typedfemale
Will this resolve YES if scaling laws suggest a 4-bit model would be competitive if compute-matched to a SOTA 16-bit model?
@NoaNabeshima Yes, you need to be better than everything else, but be trained in 4-bit (to some extent)
Related questions
Related questions
At EOY 2024, who will have the best LLM?
Will any LLM outrank GPT-4 by 150 Elo in LMSYS chatbot arena before 2025?
20% chance
Will a LLM beat human experts on GPQA by Jan 1, 2025?
91% chance
Will an LLM improve its own ability along some important metric well beyond the best trained LLMs before 2026?
58% chance
Will any LLM released in the next year double my coding productivity?
24% chance
Which company will have the best LLM by the end of 2024?
Will an LLM be able to match the ground truth >85% of the time when performing PII detection by 2024 end?
84% chance
Will LLM progress stall in 2024?
6% chance
Will a home grown LLM developed by Chinese company be competitive by end of 2025?
58% chance
Who will be ahead in the AI/LLM war by the end of 2024?