#sparsemodels search results
Smarter routing, not bigger models. Elegant work by Apple’s researchers — important progress in scalable AI design. 👏 #Apple #AppleAI #SparseModels #AIArchitecture
5/5 What aspects of trillion-param MoE deployment interest you most? Memory offloading strategies? Dynamic routing budgets? Hierarchical expert organization? Drop your thoughts below 👇 #MoE #LLMs #SparseModels #AIResearch
Read more about this exciting finding and its implications for AI development: openreview.net/forum?id=svm53… openreview.net/pdf?id=svm53KQ… #artificialintelligence #deeplearning #sparsemodels
#AI & #MachineLearning need to converge well. Check out this new theory that could make this possible! 🤔 #SparseModels forbes.com/sites/johnwern…
أعلنت Neural Magic عن نموذج اللغة Sparse Llama 3.1 8B، أصغر حجماً وأكثر كفاءة من سابقه. يهدف النموذج الجديد إلى جعل تقنيات الذكاء الاصطناعي في متناول الجميع، حيث يمكن تشغيله بأجهزة أقل قوة. #AI #MachineLearning #SparseModels #NeuralMagic #Llama_3_1_8B marktechpost.com/2024/11/25/neu…
marktechpost.com
Neural Magic Releases 2:4 Sparse Llama 3.1 8B: Smaller Models for Efficient GPU Inference
Neural Magic Releases 2:4 Sparse Llama 3.1 8B: Smaller Models for Efficient GPU Inference
@Stanford H2O.ai advisors, Trevor Hastie & Rob Tibshirani, are holding a 2-day course in #MachineLearning #DeepLearning #SparseModels.
Tech terms decoded! 🛠️ Attention techies, it’s time for #TermOfTheDay. Today, we are learning about: Sparse Models! ⚡ #TechTerms #SparseModels #AI #MachineLearning #DeepLearning #TechEducation
Download Sparse Llama: huggingface.co/neuralmagic/Sp… See benchmarks and our approach: hubs.li/Q02ZlXd90 Thanks to @_EldarKurtic, Alexdre Marques, @markurtz_, @DAlistarh, Shubhra Pandit & the Neural Magic team for always enabling efficient AI! #SparseModels #OpenSourceAI #vLLM
huggingface.co
RedHatAI/Sparse-Llama-3.1-8B-2of4 · Hugging Face
RedHatAI/Sparse-Llama-3.1-8B-2of4 · Hugging Face
@sarahookr, @KaliTessera, and Benjamin Rosman take a broader view of training #sparsnetworks and consider the role of regularization, optimization, and architecture choices on #sparsemodels. They propose a simple experimental framework, #SameCapacitySparse vs #DenseComparison.
Tomorrow at @ml_collective DLTC reading group, @KaliTessera will be presenting our work on how initialization is only one piece of the puzzle for training sparse networks. Can taking a wider view of model design choices unlock sparse training? bit.ly/3xFtHKI
Jonathan Schwarz et al. introduce #Powerpropagation, a new weight-parameterisation for #neuralnetworks that leads to inherently #sparsemodels. Exploiting the behavior of gradient descent, their method gives rise to weight updates exhibiting a "rich get richer" dynamic.
Powerpropagation: A sparsity inducing weight reparameterisation pdf: arxiv.org/pdf/2110.00296… abs: arxiv.org/abs/2110.00296 a new weight-parameterisation for neural networks that leads to inherently sparse models
5/5 What aspects of trillion-param MoE deployment interest you most? Memory offloading strategies? Dynamic routing budgets? Hierarchical expert organization? Drop your thoughts below 👇 #MoE #LLMs #SparseModels #AIResearch
Tech terms decoded! 🛠️ Attention techies, it’s time for #TermOfTheDay. Today, we are learning about: Sparse Models! ⚡ #TechTerms #SparseModels #AI #MachineLearning #DeepLearning #TechEducation
أعلنت Neural Magic عن نموذج اللغة Sparse Llama 3.1 8B، أصغر حجماً وأكثر كفاءة من سابقه. يهدف النموذج الجديد إلى جعل تقنيات الذكاء الاصطناعي في متناول الجميع، حيث يمكن تشغيله بأجهزة أقل قوة. #AI #MachineLearning #SparseModels #NeuralMagic #Llama_3_1_8B marktechpost.com/2024/11/25/neu…
marktechpost.com
Neural Magic Releases 2:4 Sparse Llama 3.1 8B: Smaller Models for Efficient GPU Inference
Neural Magic Releases 2:4 Sparse Llama 3.1 8B: Smaller Models for Efficient GPU Inference
Download Sparse Llama: huggingface.co/neuralmagic/Sp… See benchmarks and our approach: hubs.li/Q02ZlXd90 Thanks to @_EldarKurtic, Alexdre Marques, @markurtz_, @DAlistarh, Shubhra Pandit & the Neural Magic team for always enabling efficient AI! #SparseModels #OpenSourceAI #vLLM
huggingface.co
RedHatAI/Sparse-Llama-3.1-8B-2of4 · Hugging Face
RedHatAI/Sparse-Llama-3.1-8B-2of4 · Hugging Face
Read more about this exciting finding and its implications for AI development: openreview.net/forum?id=svm53… openreview.net/pdf?id=svm53KQ… #artificialintelligence #deeplearning #sparsemodels
#AI & #MachineLearning need to converge well. Check out this new theory that could make this possible! 🤔 #SparseModels forbes.com/sites/johnwern…
Jonathan Schwarz et al. introduce #Powerpropagation, a new weight-parameterisation for #neuralnetworks that leads to inherently #sparsemodels. Exploiting the behavior of gradient descent, their method gives rise to weight updates exhibiting a "rich get richer" dynamic.
Powerpropagation: A sparsity inducing weight reparameterisation pdf: arxiv.org/pdf/2110.00296… abs: arxiv.org/abs/2110.00296 a new weight-parameterisation for neural networks that leads to inherently sparse models
@sarahookr, @KaliTessera, and Benjamin Rosman take a broader view of training #sparsnetworks and consider the role of regularization, optimization, and architecture choices on #sparsemodels. They propose a simple experimental framework, #SameCapacitySparse vs #DenseComparison.
Tomorrow at @ml_collective DLTC reading group, @KaliTessera will be presenting our work on how initialization is only one piece of the puzzle for training sparse networks. Can taking a wider view of model design choices unlock sparse training? bit.ly/3xFtHKI
@Stanford H2O.ai advisors, Trevor Hastie & Rob Tibshirani, are holding a 2-day course in #MachineLearning #DeepLearning #SparseModels.
Tech terms decoded! 🛠️ Attention techies, it’s time for #TermOfTheDay. Today, we are learning about: Sparse Models! ⚡ #TechTerms #SparseModels #AI #MachineLearning #DeepLearning #TechEducation
Something went wrong.
Something went wrong.
United States Trends
- 1. #SmackDown 15.4K posts
- 2. Clemson 4,692 posts
- 3. End 1Q N/A
- 4. Kevin James 8,029 posts
- 5. Bill Clinton 144K posts
- 6. Bubba 42.8K posts
- 7. Jey Uso 2,867 posts
- 8. Bronson Reed N/A
- 9. Cam Boozer N/A
- 10. #TNATurningPoint 3,338 posts
- 11. The Miz 3,503 posts
- 12. Josh Hart N/A
- 13. End of 1st 1,594 posts
- 14. Ersson N/A
- 15. #VenezuelaConference 11.3K posts
- 16. End of the 1st 1,201 posts
- 17. Landry Shamet N/A
- 18. Dirk 9,044 posts
- 19. #cthsfb N/A
- 20. Karl Anthony 1,560 posts