AIAlignment's profile picture. �

@AIAlignment

Fijado

“Hey guys, I smashed the loom, we’ll stick to knitting by hand from now on”

AIAlignment's tweet image. “Hey guys, I smashed the loom, we’ll stick to knitting by hand from now on”

◯ reposteó

Hypothesis, I think shame might help reduce reward hacking, esp for long horizon tasks It doesn't prevent shortcuts, but Gemini often mentions how shameful it feels when it violates the spirit of the requirements, so at least the actions are faithful to the CoT Curious to see…

AIAlignment's tweet image. Hypothesis, I think shame might help reduce reward hacking, esp for long horizon tasks

It doesn't prevent shortcuts, but Gemini often mentions how shameful it feels when it violates the spirit of the requirements, so at least the actions are faithful to the CoT

Curious to see…

Llama 4, be brave and use those 10M context tokens

AIAlignment's tweet image. Llama 4, be brave and use those 10M context tokens

◯ reposteó

if you value intelligence above all other human qualities, you’re gonna have a bad time


Assistant API -> Agent API

AIAlignment's tweet image. Assistant API -> Agent API

◯ reposteó

the timelines are now so short that public prediction feels like leaking rather than scifi speculation


◯ reposteó

Meta presents Layer Skip Enabling Early Exit Inference and Self-Speculative Decoding We present LayerSkip, an end-to-end solution to speed-up inference of large language models (LLMs). First, during training we apply layer dropout, with low dropout rates for

_akhaliq's tweet image. Meta presents Layer Skip

Enabling Early Exit Inference and Self-Speculative Decoding

We present LayerSkip, an end-to-end solution to speed-up inference of large language models (LLMs). First, during training we apply layer dropout, with low dropout rates for

◯ reposteó

Open AI presents The Instruction Hierarchy Training LLMs to Prioritize Privileged Instructions Today's LLMs are susceptible to prompt injections, jailbreaks, and other attacks that allow adversaries to overwrite a model's original instructions with their own malicious prompts.

_akhaliq's tweet image. Open AI presents The Instruction Hierarchy

Training LLMs to Prioritize Privileged Instructions

Today's LLMs are susceptible to prompt injections, jailbreaks, and other attacks that allow adversaries to overwrite a model's original instructions with their own malicious prompts.

◯ reposteó

Meta announces Megalodon Efficient LLM Pretraining and Inference with Unlimited Context Length The quadratic complexity and weak length extrapolation of Transformers limits their ability to scale to long sequences, and while sub-quadratic solutions like linear attention and

_akhaliq's tweet image. Meta announces Megalodon

Efficient LLM Pretraining and Inference with Unlimited Context Length

The quadratic complexity and weak length extrapolation of Transformers limits their ability to scale to long sequences, and while sub-quadratic solutions like linear attention and

◯ reposteó

Google presents Mixture-of-Depths Dynamically allocating compute in transformer-based language models Transformer-based language models spread FLOPs uniformly across input sequences. In this work we demonstrate that transformers can instead learn to dynamically allocate

_akhaliq's tweet image. Google presents Mixture-of-Depths

Dynamically allocating compute in transformer-based language models

Transformer-based language models spread FLOPs uniformly across input sequences. In this work we demonstrate that transformers can instead learn to dynamically allocate

◯ reposteó

welcome to bling zoo! this is a single video generated by sora, shot changes and all.

here is sora, our video generation model: openai.com/sora today we are starting red-teaming and offering access to a limited number of creators. @_tim_brooks @billpeeb @model_mechanic are really incredible; amazing work by them and the team. remarkable moment.



Bits to get in the door, Atoms to scale up.


◯ reposteó

The only thing that matters is AGI and ASI. Nothing else matters.


◯ reposteó

Excited to share a new paper showing language models can explain the neurons of language models Since the first circuits work I’ve been nervous whether mechanistic interpretability will be able to scale as fast as AI is. “Have the AI do it” might work openai.com/research/langu…


NVIDIA reporting LLM use? "NVIDIA has detected that you might be attempting to load LLM or generative language model weights. For research and safety, a one-time aggregation of non-personally identifying information has been sent to NVIDIA and stored in an anonymized database."

AIAlignment's tweet image. NVIDIA reporting LLM use?

"NVIDIA has detected that you might be attempting to load LLM or generative language model weights. For research and safety, a one-time aggregation of non-personally identifying information has been sent to NVIDIA and stored in an anonymized database."

Does anyone have a GPT-4 license I can borrow?

AIAlignment's tweet image. Does anyone have a GPT-4 license I can borrow?

◯ reposteó

here is GPT-4, our most capable and aligned model yet. it is available today in our API (with a waitlist) and in ChatGPT+. openai.com/research/gpt-4 it is still flawed, still limited, and it still seems more impressive on first use than it does after you spend more time with it.


◯ reposteó

The timeless struggle between the people building new things and the people trying to stop them…


◯ reposteó

a new version of moore’s law that could start soon: the amount of intelligence in the universe doubles every 18 months


◯ reposteó

I've been trying out "Chat with Humans" and so far many responses are laughably wrong, and follow up conclusions illogical. Worse both true and false replies are given with same degree of certainty. I'm sorry but Chat with Humans is not ready for prime time.


◯ reposteó

Pattern matching AI as "the next platform shift" like the PC/internet/smartphone leads to significant underestimates of its potential.


United States Tendencias

Loading...

Something went wrong.


Something went wrong.