BigScienceLLM's profile picture. Follow the training of "BLOOM 🌸", the @BigScienceW multilingual 176B parameter open-science open-access language model, a research tool for the AI community.

BigScience Large Model Training

@BigScienceLLM

Follow the training of "BLOOM 🌸", the @BigScienceW multilingual 176B parameter open-science open-access language model, a research tool for the AI community.

Pinned

The BLOOM model is now officially released! Read more here: bigscience.huggingface.co/blog/bloom Find the model here: huggingface.co/bigscience/blo…

BLOOM is here. The largest open-access multilingual language model ever. Read more about it or get it at bigscience.huggingface.co/blog/bloom hf.co/bigscience/blo…

BigscienceW's tweet image. BLOOM is here. The largest open-access multilingual language model ever. Read more about it or get it at
bigscience.huggingface.co/blog/bloom
hf.co/bigscience/blo…


BigScience Large Model Training reposted

The Bloom paper is out. Looks like it's doing worse than current GPT3 API in zero-shot generation tasks in English but better than other open-source LLMs & better than all in zs multi-lingual (which was the main goal). Proud of the work from the community! arxiv.org/abs/2211.05100

ClementDelangue's tweet image. The Bloom paper is out. Looks like it's doing worse than current GPT3 API in zero-shot generation tasks in English but better than other open-source LLMs & better than all in zs multi-lingual (which was the main goal). Proud of the work from the community! arxiv.org/abs/2211.05100

BigScience Large Model Training reposted

Crosslingual Generalization through Multitask Finetuning 🌸 Demo: huggingface.co/bigscience/blo… 📜 arxiv.org/abs/2211.01786 💻github.com/bigscience-wor… We present BLOOMZ & mT0, a family of models w/ up to 176B params that follow human instructions in >100 languages zero-shot. 1/7

Muennighoff's tweet image. Crosslingual Generalization through Multitask Finetuning  🌸

Demo: huggingface.co/bigscience/blo…
📜 arxiv.org/abs/2211.01786
💻github.com/bigscience-wor…

We present BLOOMZ & mT0, a family of models w/ up to 176B params that follow human instructions in >100 languages zero-shot. 1/7
Muennighoff's tweet image. Crosslingual Generalization through Multitask Finetuning  🌸

Demo: huggingface.co/bigscience/blo…
📜 arxiv.org/abs/2211.01786
💻github.com/bigscience-wor…

We present BLOOMZ & mT0, a family of models w/ up to 176B params that follow human instructions in >100 languages zero-shot. 1/7

The super-fast inference solutions are finally here for all to use:

Learn how you can get under 1msec per token generation time with BLOOM 176B model! Not one, but multiple super-fast solutions including Deepspeed-Inference, Accelerate and Deepspeed-ZeRO! huggingface.co/blog/bloom-inf…



BigScience Large Model Training reposted

What do @StabilityAI @EMostaque #stablediffusion & @BigscienceW Bloom - aka the coolest new models ;) - have in common? They both use a new gen of ML licenses aimed at making ML more open & inclusive while keeping it harder to do harm with them. So cool! huggingface.co/blog/open_rail

ClementDelangue's tweet image. What do @StabilityAI @EMostaque #stablediffusion & @BigscienceW Bloom - aka the coolest new models ;) - have in common?

They both use a new gen of ML licenses aimed at making ML more open & inclusive while keeping it harder to do harm with them. So cool!

huggingface.co/blog/open_rail

BigScience Large Model Training reposted

The Technology Behind BLOOM Training🌸 Discover how @BigscienceW used @MSFTResearch DeepSpeed + @nvidia Megatron-LM technologies to train the World's Largest Open Multilingual Language Model (BLOOM): huggingface.co/blog/bloom-meg…


BigScience Large Model Training reposted

BLOOM is here. The largest open-access multilingual language model ever. Read more about it or get it at bigscience.huggingface.co/blog/bloom hf.co/bigscience/blo…

BigscienceW's tweet image. BLOOM is here. The largest open-access multilingual language model ever. Read more about it or get it at
bigscience.huggingface.co/blog/bloom
hf.co/bigscience/blo…

BigScience Large Model Training reposted

🌸@BigscienceW BLOOM's intermediate checkpoints have already shown some very cool capabilities! What's great about BLOOM is that you can ask it to generate the rest of a text - and this even if it is not yet fully trained yet! 👶 🧵 A thread with some examples

LucileSaulnier's tweet image. 🌸@BigscienceW BLOOM's intermediate checkpoints have already shown some very cool capabilities! 

What's great about BLOOM is that you can ask it to generate the rest of a text - and this even if it is not yet fully trained yet! 👶

🧵 A thread with some examples

A milestone soon to be reached 🚀💫 Can't wait to see the capabilities and performance of this long-awaited checkpoint! What about you? Have you already prepared some prompts that you want to test? ✏️



▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓ 102%


For 111 days, we've enjoyed world-class hardware stability and throughput thanks to the hard work of our friends at @Genci_fr, @INS2I_CNRS, Megatron & DeepSpeed. Having reached our objective earlier than expected, we'll keep training for a few more days. Stay tuned, more soon ;)


▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓ 101%


▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓ 100%


▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓ 99%


▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓ 98%


▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓ 97%


▓▓▓▓▓▓▓▓▓▓▓▓▓▓░ 96%


▓▓▓▓▓▓▓▓▓▓▓▓▓▓░ 95%


▓▓▓▓▓▓▓▓▓▓▓▓▓▓░ 94%


▓▓▓▓▓▓▓▓▓▓▓▓▓▓░ 92%


▓▓▓▓▓▓▓▓▓▓▓▓▓▓░ 91%


Loading...

Something went wrong.


Something went wrong.