pydatamunich's profile picture. We are a local group that meets regularly throughout Munich to learn data science, incorporating free and accessible tutorials through an engaging community,

PyData Munich

@pydatamunich

We are a local group that meets regularly throughout Munich to learn data science, incorporating free and accessible tutorials through an engaging community,

PyData Munich reposted

Launching Core v2 – The Future of Web-Based Video Rendering Today, we're rolling out a major update to our WebCodecs-powered video rendering engine. We've completely removed PixiJS and Pixi Filters, replacing them with a new, custom-built render system designed for speed,…

konstipaulus's tweet image. Launching Core v2 – The Future of Web-Based Video Rendering

Today, we're rolling out a major update to our WebCodecs-powered video rendering engine. We've completely removed PixiJS and Pixi Filters, replacing them with a new, custom-built render system designed for speed,…

PyData Munich reposted

Our Call for Proposals for Summit NY is closing EOD Sunday (11:59pm PT). Got a talk or workshop you want to get on our radar for Summit? Send it to us in the link in the next post!


PyData Munich reposted

Join us for the first dstack user group meetup in Berlin - a unique gathering diving deep into AI infrastructure & open-source AI! 🌟 Expect lightning talks, demos, and networking. RSVP now🔗lu.ma/m8arz807

dstackai's tweet image. Join us for the first dstack user group meetup in Berlin - a unique gathering diving deep into AI infrastructure & open-source AI! 🌟

Expect lightning talks, demos, and networking. 

RSVP now🔗lu.ma/m8arz807

PyData Munich reposted

Why do we treat train and test times so differently? Why is one “training” and the other “in-context learning”? Just take a few gradients during test-time — a simple way to increase test time compute — and get a SoTA in ARC public validation set 61%=avg. human score! @arcprize

akyurekekin's tweet image. Why do we treat train and test times so differently?

Why is one “training” and the other “in-context learning”?

Just take a few gradients during test-time — a simple way to increase test time compute — and  get a SoTA in ARC public validation set 61%=avg. human score! @arcprize

PyData Munich reposted

Gil free is such a good news. Tho I still be that we will have issues with decency for a decades 🙈

🐍💥Python 3.13.0 has been released! 🎉 This is the first version with 🧵experimental GIL-free mode, an experimental JIT compiler🔧, a slick new REPL 🖌️ and many new cool features! And it's faster, smarter, and more colorful than ever! 🚀 Get it here: python.org/downloads/rele…



PyData Munich reposted

quantized.w8a16 has the highest average recovery rate at around 99.8%

Muhtasham9's tweet image. quantized.w8a16 has the highest average recovery rate at around 99.8%

📢 4-bit Llama 3.1 405B, 70B, 8B Now Available! 📢 @AIatMeta's Llama 3.1 models are now quantized to 4 bits by @neuralmagic's research team and available with ~100% recovery. These enable 4X cheaper deployments (405B goes from 2 8x80GB nodes to 1 4x80GB). Continued in next...

markurtz_'s tweet image. 📢 4-bit Llama 3.1 405B, 70B, 8B Now Available! 📢

@AIatMeta's Llama 3.1 models are now quantized to 4 bits by @neuralmagic's research team and available with ~100% recovery. These enable 4X cheaper deployments (405B goes from 2 8x80GB nodes to 1 4x80GB). Continued in next...


PyData Munich reposted

On June 20th (6pm) we will host the next MUC🥨NLP meetup at Quality Minds GmbH. More Infos: munich-nlp.github.io/events/june-24… RSVP: meetup.com/pydata-munchen… Looking forward to seeing you all there!


PyData Munich reposted

GPU poor but have cool ideas? Say less Thread of GPU Grants below 🧵👇 Image by #SD3

MunichNlp's tweet image. GPU poor but have cool ideas? Say less

Thread of GPU Grants below 🧵👇

Image by #SD3

PyData Munich reposted

It's crazy how over time I have slowly replaced all of my command line tools with Rust equivalents 🦀 - cat → bat - pip → uv - grep → ripgrep - htop → zenith - fswatch → watchexec Any other good ones?


PyData Munich reposted

The AI Engineer World's Fair is confirmed for June 25 - 27 in San Francisco! Super Early Bird Tickets now on sale! Lock in the best rate by purchasing your ticket today, and use code TWEETY for an extra $100 off! Link in the next post🐦 Some more details about the event:…

aiDotEngineer's tweet image. The AI Engineer World's Fair is confirmed for June 25 - 27 in San Francisco! 

Super Early Bird Tickets now on sale! Lock in the best rate by purchasing your ticket today, and use code TWEETY for an extra $100 off! Link in the next post🐦

Some more details about the event:…
aiDotEngineer's tweet image. The AI Engineer World's Fair is confirmed for June 25 - 27 in San Francisco! 

Super Early Bird Tickets now on sale! Lock in the best rate by purchasing your ticket today, and use code TWEETY for an extra $100 off! Link in the next post🐦

Some more details about the event:…
aiDotEngineer's tweet image. The AI Engineer World's Fair is confirmed for June 25 - 27 in San Francisco! 

Super Early Bird Tickets now on sale! Lock in the best rate by purchasing your ticket today, and use code TWEETY for an extra $100 off! Link in the next post🐦

Some more details about the event:…
aiDotEngineer's tweet image. The AI Engineer World's Fair is confirmed for June 25 - 27 in San Francisco! 

Super Early Bird Tickets now on sale! Lock in the best rate by purchasing your ticket today, and use code TWEETY for an extra $100 off! Link in the next post🐦

Some more details about the event:…

PyData Munich reposted

A short thread about changes in the transformer architecture since 2017. Reading articles about LLMs, you can see phrases like “we use a standard transformer architecture.” But what does "standard" mean, and have there been changes since the original article? (1/6)

Muhtasham9's tweet image. A short thread about changes in the transformer architecture since 2017. 

Reading articles about LLMs, you can see phrases like “we use a standard transformer architecture.”

But what does "standard" mean, and have there been changes since the original article? 

(1/6)

Interestingly despite the 5 years(!) of hyper-growth of NLP space, Vanilla Transformer is holding to the Lindy Effects which is the idea that the older something is, the longer it's likely to be around in the future.



PyData Munich reposted

Start off the week by upgrading to Gradio 4.23.0 🔥 ⌨️ Big improvements to Gradio's Reload mode ⚡️ Performance improvements -- speeds up postprocessing/preprocessing for components. 🖼️ Redesigned "view API" page 🪄 Demos on the Gradio website now running on Gradio Lite

Gradio's tweet image. Start off the week by upgrading to Gradio 4.23.0 🔥

⌨️ Big improvements to Gradio's Reload mode
⚡️ Performance improvements -- speeds up postprocessing/preprocessing for components. 
🖼️ Redesigned "view API" page
🪄 Demos on the Gradio website now running on Gradio Lite

PyData Munich reposted

Here's a recipe from @ggerganov for running a GGUF of the new Mistral 7B 0.2 completion model on macOS using llama.cpp and Homebrew

quote reply with your results (note: downloads 3GB data)

ggerganov's tweet image. quote reply with your results

(note: downloads 3GB data)


PyData Munich reposted

Except it’s called AI engineering now Come to @aiDotEngineer conf to learn more

2013 — 2023: you were hired to do machine learning but do data engineering 2023 — : you were hired to do machine learning but do web dev



PyData Munich reposted

Btw: • All the model code is in *Python* • No compilation needed • Fully dynamic graph creation -> easy to debug • Simple code -> easy to extend

MLX 0.7 → 0.8 Tokens per second on M2 Ultra - 4-bit Starcoder 7B, 64.3 → 74 - 4-bit Mistral 7B, 73.1 → 83.2 Thanks to Nanobind and fast RMS and layer norms. Look at that Mistral go:



Loading...

Something went wrong.


Something went wrong.