Will a Mamba 7b model trained on 2 trillion tokens outperform Llama2-13B
Plus
21
Ṁ738Jul 1
66%
chance
1D
1W
1M
ALL
Question will resolve positive if someone trains a Mamba (https://twitter.com/tri_dao/status/1731728602230890895) language model with <=7.5billion parameters on <=2 trillion tokens that outperforms Llama2-13B on the huggingface open llm leaderboard (https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
This question is managed and resolved by Manifold.
Get
1,000
and3.00
Related questions
Related questions
Will the Jan 2024 version of the LLM detector "Binoculars" be effective against OpenAI's best model at end 2024?
59% chance
Will a open source pure Mamaba LLM surpass 82 MMLU on MMLU (5-shot) before end of year 2024?
25% chance
Will Llama 3-multimodal be natively mixed-multimodal? (VQ-VAE+next token prediction)
50% chance
When will OpenAI release a more capable LLM?
Will anyone train a TokenFormer model at scale before 2026?
25% chance
Will the next LLM released by OpenAI be worse than GPT-4 at MMLU?
16% chance
Will a 15 billion parameter LLM match or outperform GPT4 in 2024?
7% chance
Will a flagship (>60T training bytes) open-weights LLM from Meta which doesn't use a tokenizer be released in 2025?
43% chance
What will be true of the first model to cross 1400 on lmarena.ai?
Will the next major LLM by OpenAI use a new tokenizer?
77% chance