ML_explorations's profile picture.

ML explorations

@ML_explorations

ML explorations reposted

In case you have missed, this week @huggingface released IDEFICS3Llama a vision language model state-of-the-art of it's size in many benchmarks 😍

Idefics3-Llama is out! 💥 It's a multimodal model based on Llama 3.1 that accepts arbitrary number of interleaved images with text with a huge context window (10k tokens!) 😍 Link to demo and model in the next one 😏



ML explorations reposted

New leaderboard powered by Decoding Trust (outstanding paper at Neurips!), to evaluate LLM safety, such as bias and toxicity, PII, and robustness 🚀 You can find it here: huggingface.co/spaces/AI-Secu… And the intro blog is here: huggingface.co/blog/leaderboa… Congrats to @uiuc_aisecure !


ML explorations reposted

Just one hour since Canva dropped its new AI and the design world will never be the same 🤯 10 new features to 10x your productivity 🧵👇


ML explorations reposted

It used to take me hours of work to turn my presentation slides into content. With Canva's newest update, you can get it done in seconds. I've created a short video walkthrough of how to use it. What's great about it: It allows you to repurpose assets you already have with a…


Lots of magic in @canva! So much #AI #ML tools in one place to edit photos, create presentations and much more… canva.com/newsroom/news/…


ML explorations reposted

LLMs can hallucinate and lie. They can be jailbroken by weird suffixes. They memorize training data and exhibit biases. 🧠 We shed light on all of these phenomena with a new approach to AI transparency. 🧵 Website: ai-transparency.org Paper: arxiv.org/abs/2310.01405

andyzou_jiaming's tweet image. LLMs can hallucinate and lie. They can be jailbroken by weird suffixes. They memorize training data and exhibit biases.

🧠 We shed light on all of these phenomena with a new approach to AI transparency. 🧵

Website: ai-transparency.org
Paper: arxiv.org/abs/2310.01405

ML explorations reposted

There are many known “foundation models” for chat models, but what about computer vision? 🧐 In this thread, we’ll talk about few of them 👇 🖼️ Segment Anything Model 🦉 OWLViT 💬 BLIP-2 🐕 IDEFICS 🧩 CLIP 🦖 Grounding DINO Let’s go! ✨


ML explorations reposted

Mistral 7B is out. It outperforms Llama 2 13B on every benchmark we tried. It is also superior to LLaMA 1 34B in code, math, and reasoning, and is released under the Apache 2.0 licence. mistral.ai/news/announcin…

GuillaumeLample's tweet image. Mistral 7B is out. It outperforms Llama 2 13B on every benchmark we tried. It is also superior to LLaMA 1 34B in code, math, and reasoning, and is released under the Apache 2.0 licence.
mistral.ai/news/announcin…

magnet:?xt=urn:btih:208b101a0f51514ecf285885a8b0f6fb1a1e4d7d&dn=mistral-7B-v0.1&tr=udp%3A%2F%2Ftracker.opentrackr.org%3A1337%2Fannounce&tr=https%3A%2F%https://t.co/HAadNvH1t0%3A443%2Fannounce RELEASE ab979f50d7d406ab8d0b07d09806c72c



ML explorations reposted

Multimodal Foundation Models: From Specialists to General-Purpose Assistants paper page: huggingface.co/papers/2309.10… paper presents a comprehensive survey of the taxonomy and evolution of multimodal foundation models that demonstrate vision and vision-language capabilities,…

_akhaliq's tweet image. Multimodal Foundation Models: From Specialists to General-Purpose Assistants

paper page: huggingface.co/papers/2309.10…

paper presents a comprehensive survey of the taxonomy and evolution of multimodal foundation models that demonstrate vision and vision-language capabilities,…

ML explorations reposted

Canva now has incredible AI features. You can easily create visuals in seconds. I'll show you how to create AI-boosted designs on Canva:


ML explorations reposted

LLaSM: Large Language and Speech Model paper page: huggingface.co/papers/2308.15… Multi-modal large language models have garnered significant interest recently. Though, most of the works focus on vision-language multi-modal models providing strong capabilities in following…

_akhaliq's tweet image. LLaSM: Large Language and Speech Model

paper page: huggingface.co/papers/2308.15…

Multi-modal large language models have garnered significant interest recently. Though, most of the works focus on vision-language multi-modal models providing strong capabilities in following…

ML explorations reposted

We are thrilled to share our groundbreaking paper published today in @Nature: "Champion-Level Drone Racing using Deep Reinforcement Learning." We introduce "Swift," the first autonomous vision-based drone that beat human world champions in several fair head-to-head races! PDF…

davsca1's tweet image. We are thrilled to share our groundbreaking paper published today in @Nature: "Champion-Level Drone Racing using Deep Reinforcement Learning." We introduce "Swift," the first autonomous vision-based drone that beat human world champions in several fair head-to-head races! PDF…

ML explorations reposted

🚨 NLP News 🛠 Tool-Augmented LLMs open.substack.com/pub/nlpnewslet…

seb_ruder's tweet image. 🚨 NLP News

🛠 Tool-Augmented LLMs

open.substack.com/pub/nlpnewslet…

ML explorations reposted

New on @huggingface — CoTracker simultaneously tracks the movement of multiple points in videos using a flexible design based on a transformer network — it models correlation of the points in time via specialized attention layers. 🤗 Try CoTracker ➡️ bit.ly/3swQFqt


ML explorations reposted

We beat GPT-4 on HumanEval with fine-tuned CodeLlama-34B! Here's how we did it: phind.com/blog/code-llam… 🚀 Both models have been open-sourced on Huggingface: huggingface.co/Phind


ML explorations reposted

Today, we release CodeLlama, a collection of base and instruct-finetuned models with 7B, 13B and 34B parameters. For coding tasks, CodeLlama 7B is competitive with Llama 2 70B and CodeLlama 34B is state-of-the-art among open models. Paper and weights: ai.meta.com/research/publi…

b_roziere's tweet image. Today, we release CodeLlama, a collection of base and instruct-finetuned models with 7B, 13B and 34B parameters. For coding tasks, CodeLlama 7B is competitive with Llama 2 70B and CodeLlama 34B is state-of-the-art among open models. Paper and weights: ai.meta.com/research/publi…
b_roziere's tweet image. Today, we release CodeLlama, a collection of base and instruct-finetuned models with 7B, 13B and 34B parameters. For coding tasks, CodeLlama 7B is competitive with Llama 2 70B and CodeLlama 34B is state-of-the-art among open models. Paper and weights: ai.meta.com/research/publi…
b_roziere's tweet image. Today, we release CodeLlama, a collection of base and instruct-finetuned models with 7B, 13B and 34B parameters. For coding tasks, CodeLlama 7B is competitive with Llama 2 70B and CodeLlama 34B is state-of-the-art among open models. Paper and weights: ai.meta.com/research/publi…
b_roziere's tweet image. Today, we release CodeLlama, a collection of base and instruct-finetuned models with 7B, 13B and 34B parameters. For coding tasks, CodeLlama 7B is competitive with Llama 2 70B and CodeLlama 34B is state-of-the-art among open models. Paper and weights: ai.meta.com/research/publi…

ML explorations reposted

AutoGPTQ is now natively supported in transformers! 🤩 AutoGPTQ is a library for GPTQ, a post-training quantization technique to quantize autoregressive generative LLMs. 🦜 With this integration, you can quantize LLMs with few lines of code! Read more 👉 hf.co/blog/gptq-inte…

mervenoyann's tweet image. AutoGPTQ is now natively supported in transformers! 🤩
AutoGPTQ is a library for GPTQ, a post-training quantization technique to quantize  autoregressive generative LLMs. 🦜
With this integration, you can quantize LLMs with few lines of code!
Read more 👉 hf.co/blog/gptq-inte…

ML explorations reposted

Will be interesting how this LoRA-on-demand service will compare to open-source LoRA on prem. Here's a little reminder that open-source Llama 2 compares very favorably to ChatGPT / GPT 3.5

rasbt's tweet image. Will be interesting how this LoRA-on-demand service will compare to open-source LoRA on prem.

Here's a little reminder that open-source Llama 2 compares very favorably to ChatGPT / GPT 3.5

fine-tuning for GPT-3.5 turbo! (and coming this fall for GPT-4)



ML explorations reposted

fine-tuning for GPT-3.5 turbo! (and coming this fall for GPT-4)

We've just launched fine-tuning for GPT-3.5 Turbo! Fine-tuning lets you train the model on your company's data and run it at scale. Early tests have shown that fine-tuned GPT-3.5 Turbo can match or exceed GPT-4 on narrow tasks: openai.com/blog/gpt-3-5-t…

OpenAI's tweet image. We've just launched fine-tuning for GPT-3.5 Turbo! Fine-tuning lets you train the model on your company's data and run it at scale. Early tests have shown that fine-tuned GPT-3.5 Turbo can match or exceed GPT-4 on narrow tasks: openai.com/blog/gpt-3-5-t…


ML explorations reposted

Introducing SeamlessM4T, the first all-in-one, multilingual multimodal translation model. This single model can perform tasks across speech-to-text, speech-to-speech, text-to-text translation & speech recognition for up to 100 languages depending on the task. Details ⬇️


United States Trends

Loading...

Something went wrong.


Something went wrong.