TedInRealLife's profile picture. still on pace for PhD-before-50

Ted

@TedInRealLife

still on pace for PhD-before-50

Ted memposting ulang

Large multimodal models often lack precise low-level perception needed for high-level reasoning, even with simple vector graphics. We bridge this gap by proposing an intermediate symbolic representation that leverages LLMs for text-based reasoning. mikewangwzhl.github.io/VDLM 🧵1/4


Wordle 1,006 5/6 🟩⬜⬜⬜⬜ 🟩🟩⬜⬜⬜ 🟩🟩🟩⬜⬜ 🟩🟩🟩🟩⬜ 🟩🟩🟩🟩🟩

Terrible %0
Awesome %0

0 suara · Hasil akhir


Congress could use more computer scientists like @RepTedLieu

TedInRealLife's tweet image. Congress could use more computer scientists like @RepTedLieu

Ted memposting ulang

Why do neural nets generalize so well?🤔 There's a ton of work on SGD, flat minima, ... but the root cause is that their inductive biases somehow match properties of real-world data.🌎 We've examined these inductive biases in *untrained* networks.🎲 arxiv.org/abs/2403.02241 ⬇️

DamienTeney's tweet image. Why do neural nets generalize so well?🤔
There's a ton of work on SGD, flat minima, ... but the root cause is that their inductive biases somehow match properties of real-world data.🌎 We've examined these inductive biases in *untrained* networks.🎲
arxiv.org/abs/2403.02241 ⬇️

Mechanism for feature learning in neural networks and backpropagation-free machine learning models | Science science.org/doi/10.1126/sc…


Ted memposting ulang

And today is T - 2 weeks for my other @NVIDIA #GTC session - a fireside chat with @ChrSzegedy, cofounder of @xAI. Christian is one of the seminal research figures in the Deep Learning community, but the main focus of our chat will be on something that he has been working very…

tunguz's tweet image. And today is T - 2 weeks for my other @NVIDIA #GTC session - a fireside chat with @ChrSzegedy, cofounder of @xAI. Christian is one of the seminal research figures in the Deep Learning community, but the main focus of our chat will be on something that he has been working very…

Ted memposting ulang

How do we capture local features across multiple resolutions? While standard convolutional layers work only on a fixed input-resolution, we design local neural operators that learn integral and differential kernels, and are principled ways to extend standard convolutions to…

AnimaAnandkumar's tweet image. How do we capture local features across multiple resolutions? While standard convolutional layers work only on a fixed input-resolution, we design local neural operators that learn integral and differential kernels, and are principled ways to extend standard convolutions to…

Ted memposting ulang

Me: Can you draw a very normal image? ChatGPT: Here is a very normal image depicting a tranquil suburban street scene during the daytime. Me: Not bad, but can you go more normal than that? (cont.)

ESYudkowsky's tweet image. Me:  Can you draw a very normal image?

ChatGPT:  Here is a very normal image depicting a tranquil suburban street scene during the daytime.

Me:  Not bad, but can you go more normal than that?

(cont.)

Ted memposting ulang

Today's AI (credit: Balazs Szegedy using DALL-E-3)

ChrSzegedy's tweet image. Today's AI

(credit: Balazs Szegedy using DALL-E-3)

So tempting to just group a collection of "specialized"s and call it "general" arxiv.org/abs/2311.02462


Ted memposting ulang

Inception used 1.5X less compute than AlexNet and 12X less than VGG, outperforming both. The trend continued with mobile net... etc. IMO, today's LLMs are insanely inefficient/compute. Regulations that impose limits on the amount of compute spent on AI training will just…


Ted memposting ulang

Regulation starts at roughly two orders of magnitude larger than a ~70B Transformer trained on 2T tokens -- which is ~5e24. Note: increasing the size of the dataset OR the size of the transformer increases training flops. The (rumored) size of GPT-4 is regulated.

New Executive Order is out. Already one notable item: Any AI model that required more than 1e26 floating point operations or 1e23 integer operations to build must report to the government.

DavidVorick's tweet image. New Executive Order is out. Already one notable item:

Any AI model that required more than 1e26 floating point operations or 1e23 integer operations to build must report to the government.


Finally found the introverts' dining area at SJC. 🙂

TedInRealLife's tweet image. Finally found the introverts' dining area at SJC. 🙂

$0.25/mile

After a journey of nearly 3.9 billion miles, the #OSIRISREx asteroid sample return capsule is back on Earth. Teams perform the initial safety assessment—the first persons to come into contact with this hardware since it was on the other side of the solar system.



Ted memposting ulang

The AI world is in a GPU crunch and meanwhile NERSC is offering 50% off its A100 GPUs (rest.nersc.gov/REST/announcem…). They could make a killing by backfilling idle capacity with commercial workloads 💰 💰 💰


This is super cool

Personal Announcement! I’m launching @SakanaAILabs together with my friend, Llion Jones (@YesThisIsLion). sakana.ai is a new R&D-focused company based in Tokyo, Japan. We’re on a quest to create a new kind of foundation model based on nature-inspired intelligence!

hardmaru's tweet image. Personal Announcement! I’m launching @SakanaAILabs together with my friend, Llion Jones (@YesThisIsLion).

sakana.ai is a new R&D-focused company based in Tokyo, Japan.

We’re on a quest to create a new kind of foundation model based on nature-inspired intelligence!


Loading...

Something went wrong.


Something went wrong.