Tyler
@Developer65537
🇺🇸 Data Analyst (PM) ← AI Startup Co-founder | IBM Quantum Developer | Toptal-Certified Full-Stack Engineer | Math Olympiad Camp | Invest, AI, Quantum, Space
You might like
Scientists demonstrated a method for scalable spin squeezing using the dipolar interactions in an ultracold gas of fermionic erbium atoms, achieving 7.1 dB of metrologically useful squeezing using a protocol that can be implemented with neutral atoms. go.aps.org/3LLdgZN
I highly encourage anyone interested in Delta Attention/Deltanet to read through this whole thread. You can see how I start from practically 0 and am trying to understand Kimi Delta Attention and related linear attention literature by spamming Grad with questions.
Originally a while back i got some intuition with it from the query key value perspective which might help (theres also the gradient descent perspective which is good too). Scraped this from a chat with @stochasticchasm a year ago so might be a bit dodgy. Imagine u want to store…
It's a shame Kaluza-Klein theory didn't pan out, because we'd have badass particle called the 'Graviphoton' Graviphoton cannons, Graviphoton thrusters, etc
Optimal Flow Matching (OFM) is a cutting-edge deep learning technique for training generative models. It learns the most efficient path (an "optimal flow") to transform simple noise into complex data, like a realistic image. It's known for being much faster and more stable to…
Diffusion Language Models are Super Data Learners… now on arXiv with MegaDLMs, the full large-scale training framework (6.1K H100s, 462B-param run, 47 % MFU). Supports diffusion and autoregressive LMs, dense and MoE architectures, FP8/BF16/FP16 precision, and multi-axis…
Breaking release! OpenHands Software Agent SDK — a complete redesign of the popular 64k⭐️ OpenHands framework. It offers: • Plug-and-play agent interfaces • Sandboxed & portable execution • Multi-LLM routing • Built-in security analysis Benchmarks on SWE-Bench Verified &…
🚨 MIT just humiliated every major AI lab and nobody’s talking about it. They built a new benchmark called WorldTest to see if AI actually understands the world… and the results are brutal. Even the biggest models Claude, Gemini 2.5 Pro, OpenAI o3 got crushed by humans.…
congrats to llama 3 large for winning the LLM trading contest by not participating
When the XPENG IRON gracefully approaches you, @Tesla_Optimus how will you greet her?
AI can become a fully autonomous data scientist. DeepAnalyze-8B is the first agentic LLM capable of handling the entire data science pipeline—from raw data to analyst-grade research reports—without predefined workflows. It learns like a human via a curriculum-based agentic…
New: Qwen-Image-2509-MultipleAngles. Very solid model and probably a lot of creative use cases to find with it. ⬇️ Free demo available on Hugging Face
Met a Meta AI researcher. He studied Physics in China, came to the US for a PhD in Physics, and then fell in love with AI, despite never having studied computer science. He watched Andrej Karpathy and Andrew Ng, bought a GPU, read every arXiv paper title daily, and dived into…
Top AI Papers of the Week (Oct 27 - Nov 2): - SmolLM2 - AgentFold - Precision-RL - Multi-Agent Evolve - Agent Data Protocol - Graph-based Agent Planning - Introspective Awareness in LLMs Read on for more:
Google $GOOG director just said HALF of all code is now being written by AI. Wow…
Wow, language models can talk without words. A new framework, Cache-to-Cache (C2C), lets multiple LLMs communicate directly through their KV-caches instead of text, transferring deep semantics without token-by-token generation. It fuses cache representations via a neural…
Quantum computing is best done in the permanently shadowed craters on the Moon
Most multi-drone systems struggle with one thing: coordination under real-world constraints. A new paper in Science Robotics from TU Delft proposes a model-based approach that lets multiple quadrotors jointly move and orient a cable-suspended load. BUT: without relying on…
Im confused about "10,000 more efficient" part. This means you can train stable-diffusion-3 like model with 20$~ ish amount of electricity. What stops them from building a model and demonstrating it, beyond *checks note* ... Fashion MNIST? Im genuinely curious whats stopping them…
United States Trends
- 1. #GRAMMYs 211K posts
- 2. Clipse 7,857 posts
- 3. addison rae 12.7K posts
- 4. olivia dean 7,617 posts
- 5. #FanCashDropPromotion 2,813 posts
- 6. gaga 69.6K posts
- 7. Dizzy 7,525 posts
- 8. Katseye 75.7K posts
- 9. Leon Thomas 9,115 posts
- 10. Kehlani 24.8K posts
- 11. ravyn lenae N/A
- 12. lorde 8,844 posts
- 13. #FridayVibes 5,984 posts
- 14. AOTY 12.5K posts
- 15. Album of the Year 37.9K posts
- 16. Best New Artist 13.8K posts
- 17. GAME DAY 27.8K posts
- 18. Good Friday 65.3K posts
- 19. Benito 5,853 posts
- 20. Durand 3,093 posts
You might like
-
𝒏🅞𝒈𝒖❶𝒔𝒔
@n0gu1ss_yt -
idea
@IdeaSimulated -
tttppp
@TtttpppppM -
狐坂
@kitsunezaka55 -
黒猫さん🐈⬛
@StudioNoGimmick -
青月猫
@seigetsubyo -
hugues.freesoul
@FreesoulHugues -
透明色素
@toumei_shikiso -
ΛIBΞN
@aibenxyz -
AI育生計画
@questionsnake -
スノク200
@sunouku -
HO・KA・U
@hrathnir7 -
しずのり@AIイラスト投稿用
@uni_nori_ -
voxeloops
@voxeloops -
びとびとさん
@betoo3_char
Something went wrong.
Something went wrong.