TechSamaya's profile picture. Tweets in Machine Learning, Technology and Startups

ML/NLP Weekly newsletter
http://newsletter.samaya.tech/

Samaya

@TechSamaya

Tweets in Machine Learning, Technology and Startups ML/NLP Weekly newsletter http://newsletter.samaya.tech/

Contrastive Learning has been shown to be an effective way to learn representations for various tasks. This post summarized some of the important papers for this method. blog.samaya.tech/2022/04/Contra… #DeepLearning

TechSamaya's tweet image. Contrastive Learning has been shown to be an effective way to learn representations for various tasks. This post summarized some of the important papers for this method.

blog.samaya.tech/2022/04/Contra…

#DeepLearning

It's tough to get GANs to train. Lots of problems can arise, from mode collapse to training instability.This post summarized the problems and solutions. blog.samaya.tech/2022/04/GAN-.h… #DeepLearning #MachineLearning

TechSamaya's tweet image. It's tough to get GANs to train. Lots of problems can arise, from mode collapse to training instability.This post summarized the problems and solutions. 

blog.samaya.tech/2022/04/GAN-.h…

#DeepLearning #MachineLearning

We integrated Decision Transformers, an Offline Reinforcement Learning method, into the 🤗 transformers library and the @huggingface Hub 🥳 ➕ 9 pre-trained models for continuous control tasks in Gym 🔥 We wrote a tutorial if you want to try it 👉bit.ly/36SI4Ds

ThomasSimonini's tweet image. We integrated Decision Transformers, an Offline Reinforcement Learning method, into the 🤗 transformers library and the @huggingface Hub 🥳

➕ 9 pre-trained models for continuous control tasks in Gym 🔥

We wrote a tutorial if you want to try it 👉bit.ly/36SI4Ds


Today, we’re sharing a roundup of Meta AI’s recent cutting-edge multimodal research, which we believe will collectively lead to more interactive, immersive, and smarter AI systems of the future: ai.facebook.com/blog/advances-…

AIatMeta's tweet image. Today, we’re sharing a roundup of Meta AI’s recent cutting-edge multimodal research, which we believe will collectively lead to more interactive, immersive, and smarter AI systems of the future: ai.facebook.com/blog/advances-…


An open-source solution has a 25% better accuracy than Amazon Forecast and is 20% more accurate than fbprophet. It also performs 4x faster than Amazon Forecast and is less expensive. github.com/Nixtla/nixtla #MachineLearning


Terraform is a powerful tool for building, changing, and versioning infrastructure. With its infrastructure as code approach, execution plans, and resource graph, Terraform is a must-have for any #infrastructure engineer! github.com/hashicorp/terr… #openSource #backend


Checkout how Yahoo, Netflix or Doordash designed their recommendation system blog.samaya.tech/2022/03/recomm…

TechSamaya's tweet image. Checkout how Yahoo, Netflix or Doordash designed their recommendation system

blog.samaya.tech/2022/03/recomm…

Introducing the Multimodal Bottleneck Transformer, a novel transformer-based model for multimodal fusion that restricts cross-modal attention flow to achieve state-of-the-art results on video classification tasks with less compute. Read more ↓ goo.gle/3MN3YZz



See how GPT3 explains Instance-Conditioned GAN which is a new paper but still the generated text is surprisingly good. blog.samaya.tech/2022/03/gpt3-i… #NLP #GPT3 #AI #RPA

TechSamaya's tweet image. See how GPT3 explains Instance-Conditioned GAN which is a new paper but still the generated text is surprisingly good.

blog.samaya.tech/2022/03/gpt3-i…
#NLP #GPT3 #AI #RPA

Samaya reposted

Fine-Tuning with Hugging Face Trainer. Check out my video tutorial to learn how you can improve and simplify ML model fine-tuning workflow with Hugging Face Trainer. This API is very well defined. @huggingface rocks 🚀 Video: youtu.be/L6Dr8AFXMd8 Code: github.com/katanaml/sparr…


A short post on overfitting and how to detect it in your model. blog.samaya.tech/2022/03/overfi…

TechSamaya's tweet image. A short post on overfitting and how to detect it in your model. 

blog.samaya.tech/2022/03/overfi…

Samaya reposted

LMs can learn via inference alone through demonstrations -- but how does it work? We find that LMs do not really need correct input-output pairs. Randomly replacing labels in the demonstrations barely hurts performance, consistently over 12 models. arxiv.org/abs/2202.12837

sewon__min's tweet image. LMs can learn via inference alone through demonstrations -- but how does it work?

We find that LMs do not really need correct input-output pairs. Randomly replacing labels in the demonstrations barely hurts performance, consistently over 12 models.

arxiv.org/abs/2202.12837

Samaya reposted

We’re pleased to announce new advances in SEER, Meta AI’s groundbreaking self-supervised #computervision model. SEER is now not only much more powerful, it also produces fairer, more robust computer vision models. Learn more: ai.facebook.com/blog/seer-10b-…


Samaya reposted

Checkout this easy to read tutorial on how to use a pre-trained Sentiment Analysis or train one with your data #NLP

Getting started with sentiment analysis has never been easier! 🚀 In this new post, you’ll learn how to use pre-trained models, how to fine-tune your own sentiment model and how to use these models to analyze tweets in just a few lines of Python code 🔥 huggingface.co/blog/sentiment…



Samaya reposted

A ready to use GUI to build your own prompt source It is also connected to @huggingface dataset to make the process smoother Paper: arxiv.org/abs/2202.01279 Code: github.com/bigscience-wor… #NLP #GPT3 #paper #OpenSource

arezae's tweet image. A ready to use GUI  to build your own prompt source
It is also connected to @huggingface dataset to make the process smoother
Paper: arxiv.org/abs/2202.01279
Code: github.com/bigscience-wor…
#NLP #GPT3 #paper #OpenSource

GLaM: Efficient Scaling of Language Models with Mixture-of-Experts 1.2T GLaM achieves better overall zero-shot perf over GPT-3 across 29 NLP tasks, while consuming only 1/3 of the energy used to train GPT-3. arxiv.org/abs/2112.06905

arankomatsuzaki's tweet image. GLaM: Efficient Scaling of Language Models with Mixture-of-Experts

1.2T GLaM achieves better overall zero-shot perf over GPT-3 across 29 NLP tasks, while consuming only 1/3 of the energy used to train GPT-3.

arxiv.org/abs/2112.06905


Sense raises $50M to bolster recruitment efforts with AI ift.tt/3IwYdNw



Loading...

Something went wrong.


Something went wrong.