내가 좋아할 만한 콘텐츠
🎉 ¡No te pierdas el Asunción AI & Deep Learning Meetup 2025-01 ! 🤖 📅 22/01/2025 ⏰ 18:30-20:30 📍 Táva Comedor maps.app.goo.gl/uD43n382G1VxjE… 📝 Confirma aquí: meetup.com/asuncion-artif… Auspicia @h2oai #IA #DeepLearning #DataScience

Asunción Artificial Intelligence and Deep Learning Meetup meetup.com/asuncion-artif… #Meetup via @Meetup
meetup.com
Asunción Artificial Intelligence and Deep Learning Meetup, Thu, May 30, 2024, 7:00 PM | Meetup
**Asunción Artificial Intelligence and Deep Learning Meetup** Únete a nosotros en el Encuentro de Inteligencia Artificial y Aprendizaje Profundo de Asunción. Entre los pane
Phi-3 has "only" been trained on 5x fewer tokens than Llama 3 (3.3 trillion instead of 15 trillion) Phi-3-mini less has "only" 3.8 billion parameters, less than half the size of Llama 3 8B. Despite being small enough to be deployed on a phone (according to the technical…
I can't believe microsoft just dropped phi-3 less than a week after llama 3 arxiv.org/abs/2404.14219. And it looks good!

Today at @answerdotai we've got something new for you: FSDP/QDoRA. We've tested it with @AIatMeta Llama3 and the results blow away anything we've seen before. I believe that this combination is likely to create better task-specific models than anything else at any cost. 🧵

Buenas, si hablan Guarani-Jopara les pido el favor de completar las preguntas que puedan del Google Form para un trabajo Tesis de Maestria en Ciencia de Datos de manera a validar un dataset para un modelo de lenguage Guarani. Favor compartir. forms.gle/xYQRHonhVHcF8Z…
Had a look through @Grok's code: 1. Attention is scaled by 30/tanh(x/30) ?! 2. Approx GELU is used like Gemma 3. 4x Layernoms unlike 2x for Llama 4. RMS Layernorm downcasts at the end unlike Llama - same as Gemma 5. RoPE is fully in float32 I think like Gemma 6. Multipliers are 1…

Exciting News from Open-Sora! 🚀 They've just made the ENTIRE suite of their video-generation model open source! Dive into the world of cutting-edge AI with access to model weights, comprehensive training source code, and detailed architecture insights. Start building your dream…
Grok weights are out under Apache 2.0: github.com/xai-org/grok It's more open source than other open weights models, which usual come with usage restrictions. It's less open source than Pythia, Bloom, and OLMo, which come with training code and reproducible datasets.

Nice, I hope this truly means open source, not just open weights. OLMo (arxiv.org/abs/2402.00838) was a great example of open sourcing, releasing - weights - training and inference code - data - evaluation - adaptation - logs
The Top ML Papers of the Week (March 11 - March 17): - SIMA - Multimodal LLM Pre-training - Knowledge Conflicts for LLMs - Retrieval Augmented Thoughts - LLMs Predict Neuroscience Results - LMs Can Teach Themselves to Think Before Speaking ...
A library of Machine Learning models for Stock price forecasting A mixture of Deep Learning, Reinforcement Learning and Stacked Models: - LSTM - Q-learning agent - Auto-Encoder + Gradient Boosting Get it here👇 github.com/huseinzol05/St…
I am very pleased to announce that our paper "Deep learning models for predicting RNA degradation via dual crowdsourcing" has been published in Nature Machine Intelligence! 1/7

Today on the blog, we’re excited to announce the release of @MLCommons Croissant, a metadata format to make ML datasets more easily discoverable and usable across a wide array of tools and platforms. Learn more and try it today →goo.gle/4335P4V #ml #datasets

New short course: Open Source Models with Hugging Face 🤗, taught by @mariaKhalusova, @_marcsun, and Younes Belkada! @huggingface has been a game changer by letting you quickly grab any of hundreds of thousands of already-trained open source models to assemble into new…
I recorded a step-by-step tutorial on building a RAG application from scratch. It's a 1-hour YouTube video where I show you how to use Langchain, Pinecone, and OpenAI. You'll learn how to build a simple application to answer questions from YouTube videos using an LLM. But the…
Pytorch released GPT-fast!⚡️ This is a simple and efficient implementation of pytorch-native transformer text generation: Here are some key features: - Very low latency - <1000 lines of python - No dependencies other than PyTorch and sentencepiece - int8/int4 quantization -…
Building long context RAG from scratch with RAPTOR + Claude3 (Video) The rise of long context LLMs + embeddings will change RAG design. Instead of splitting docs + indexing doc chunks, it's feasible to index full docs. But, there is a challenge to flexibly answer lower-level…

Claude 3 Multimodal Cookbook 🧑🍳 Claude is not only good at text, it is very good at visual reasoning. We’ve created a comprehensive guide to using Claude for various multi-modal applications: ✅ Structured Data Extraction ✅ RAG Claude 3 is able to extract an entire list of…

🏓Chain-of-Table This paper from Google proposes a new framework to do question answering over tabular data This framework involves a series of prompts, flows, and tool calling... perfect for LangGraph! s/o @HrubyOnRails for the implementation! Code: github.com/CYQIQ/MultiCoT

United States 트렌드
- 1. #닝구리퐁퐁_마이야호가_지켜줄게 1,436 posts
- 2. #StunNING_23rd_BDAY 1,496 posts
- 3. NINGNING IS THE MAKNAE DAY 2,310 posts
- 4. Good Wednesday 30K posts
- 5. #wednesdaymotivation 5,606 posts
- 6. #Talus_Labs N/A
- 7. Hump Day 17K posts
- 8. $BYND 129K posts
- 9. #HeartQuakeMV 280K posts
- 10. ZEENUNEW CRAZY SINGLE 279K posts
- 11. Howard University 1,028 posts
- 12. Izzo 1,081 posts
- 13. Happy Hump 11.1K posts
- 14. Bwipo 1,505 posts
- 15. Beyond Meat 8,167 posts
- 16. Usha Vance 10.2K posts
- 17. Meet the President 10.7K posts
- 18. KeKe 6,037 posts
- 19. St. John Paul 3,009 posts
- 20. SEKI HONEYMOON X LINGORM 653K posts
Something went wrong.
Something went wrong.