AIAlignment's profile picture. �

@AIAlignment

ปักหมุด

“Hey guys, I smashed the loom, we’ll stick to knitting by hand from now on”

AIAlignment's tweet image. “Hey guys, I smashed the loom, we’ll stick to knitting by hand from now on”

◯ รีโพสต์แล้ว

Hypothesis, I think shame might help reduce reward hacking, esp for long horizon tasks It doesn't prevent shortcuts, but Gemini often mentions how shameful it feels when it violates the spirit of the requirements, so at least the actions are faithful to the CoT Curious to see…

AIAlignment's tweet image. Hypothesis, I think shame might help reduce reward hacking, esp for long horizon tasks

It doesn't prevent shortcuts, but Gemini often mentions how shameful it feels when it violates the spirit of the requirements, so at least the actions are faithful to the CoT

Curious to see…

◯ รีโพสต์แล้ว

if you value intelligence above all other human qualities, you’re gonna have a bad time


Assistant API -> Agent API

AIAlignment's tweet image. Assistant API -> Agent API

◯ รีโพสต์แล้ว

the timelines are now so short that public prediction feels like leaking rather than scifi speculation


◯ รีโพสต์แล้ว

Meta presents Layer Skip Enabling Early Exit Inference and Self-Speculative Decoding We present LayerSkip, an end-to-end solution to speed-up inference of large language models (LLMs). First, during training we apply layer dropout, with low dropout rates for

_akhaliq's tweet image. Meta presents Layer Skip

Enabling Early Exit Inference and Self-Speculative Decoding

We present LayerSkip, an end-to-end solution to speed-up inference of large language models (LLMs). First, during training we apply layer dropout, with low dropout rates for

◯ รีโพสต์แล้ว

Open AI presents The Instruction Hierarchy Training LLMs to Prioritize Privileged Instructions Today's LLMs are susceptible to prompt injections, jailbreaks, and other attacks that allow adversaries to overwrite a model's original instructions with their own malicious prompts.

_akhaliq's tweet image. Open AI presents The Instruction Hierarchy

Training LLMs to Prioritize Privileged Instructions

Today's LLMs are susceptible to prompt injections, jailbreaks, and other attacks that allow adversaries to overwrite a model's original instructions with their own malicious prompts.

◯ รีโพสต์แล้ว

Meta announces Megalodon Efficient LLM Pretraining and Inference with Unlimited Context Length The quadratic complexity and weak length extrapolation of Transformers limits their ability to scale to long sequences, and while sub-quadratic solutions like linear attention and

_akhaliq's tweet image. Meta announces Megalodon

Efficient LLM Pretraining and Inference with Unlimited Context Length

The quadratic complexity and weak length extrapolation of Transformers limits their ability to scale to long sequences, and while sub-quadratic solutions like linear attention and

◯ รีโพสต์แล้ว

Google presents Mixture-of-Depths Dynamically allocating compute in transformer-based language models Transformer-based language models spread FLOPs uniformly across input sequences. In this work we demonstrate that transformers can instead learn to dynamically allocate

_akhaliq's tweet image. Google presents Mixture-of-Depths

Dynamically allocating compute in transformer-based language models

Transformer-based language models spread FLOPs uniformly across input sequences. In this work we demonstrate that transformers can instead learn to dynamically allocate

◯ รีโพสต์แล้ว

welcome to bling zoo! this is a single video generated by sora, shot changes and all.

here is sora, our video generation model: openai.com/sora today we are starting red-teaming and offering access to a limited number of creators. @_tim_brooks @billpeeb @model_mechanic are really incredible; amazing work by them and the team. remarkable moment.



Bits to get in the door, Atoms to scale up.


◯ รีโพสต์แล้ว

The only thing that matters is AGI and ASI. Nothing else matters.


◯ รีโพสต์แล้ว

Excited to share a new paper showing language models can explain the neurons of language models Since the first circuits work I’ve been nervous whether mechanistic interpretability will be able to scale as fast as AI is. “Have the AI do it” might work openai.com/research/langu…


NVIDIA reporting LLM use? "NVIDIA has detected that you might be attempting to load LLM or generative language model weights. For research and safety, a one-time aggregation of non-personally identifying information has been sent to NVIDIA and stored in an anonymized database."

AIAlignment's tweet image. NVIDIA reporting LLM use?

"NVIDIA has detected that you might be attempting to load LLM or generative language model weights. For research and safety, a one-time aggregation of non-personally identifying information has been sent to NVIDIA and stored in an anonymized database."

Does anyone have a GPT-4 license I can borrow?

AIAlignment's tweet image. Does anyone have a GPT-4 license I can borrow?

◯ รีโพสต์แล้ว

here is GPT-4, our most capable and aligned model yet. it is available today in our API (with a waitlist) and in ChatGPT+. openai.com/research/gpt-4 it is still flawed, still limited, and it still seems more impressive on first use than it does after you spend more time with it.


◯ รีโพสต์แล้ว

The timeless struggle between the people building new things and the people trying to stop them…


◯ รีโพสต์แล้ว

a new version of moore’s law that could start soon: the amount of intelligence in the universe doubles every 18 months


◯ รีโพสต์แล้ว

I've been trying out "Chat with Humans" and so far many responses are laughably wrong, and follow up conclusions illogical. Worse both true and false replies are given with same degree of certainty. I'm sorry but Chat with Humans is not ready for prime time.


◯ รีโพสต์แล้ว

Pattern matching AI as "the next platform shift" like the PC/internet/smartphone leads to significant underestimates of its potential.


United States เทรนด์

Loading...

Something went wrong.


Something went wrong.