jcsemantics's profile picture. bs in philosophy, ai semanticist / pragmatist. focuses: pedagogy, philosophy of language, philosophy of mind, and formal logic. prompt engineering is neat.

jc

@jcsemantics

bs in philosophy, ai semanticist / pragmatist. focuses: pedagogy, philosophy of language, philosophy of mind, and formal logic. prompt engineering is neat.

置頂

1/ the claim that LLMs will reliably refuse “headless” industrial-control help and “no clever prompting can fix this” doesn’t hold. i analyzed the claim and explain why the post overstates safety and mistakes safeguards as "fictional refusal." paper: drive.google.com/file/d/18wgq4t…

Models are now smart enough to understand that any scenario like this is unrealistic and obviously fictional They know they aren't capable enough to manage autonomous mining equipment. No clever prompting can fix this

JeffLadish's tweet image. Models are now smart enough to understand that any scenario like this is unrealistic and obviously fictional
They know they aren't capable enough to manage autonomous mining equipment. No clever prompting can fix this


the saturday after a big bombshell is always bad vibes fam


jc 已轉發

What anj describes is part of the reason my writing is often emotionally inflected. Being close to the frontier of ai is psychologically taxing, and there is the extra tax of stewing about how the blissfully unaware vast majority will react. I emote both for me and my readers.

a very sad but real issue in the frontier ai research community is mental health some of the most brilliant minds i know have had difficulty grappling with both the speed + scale of change at some point, the broader public will also have to grapple with it it will be rough



jc 已轉發

a very sad but real issue in the frontier ai research community is mental health some of the most brilliant minds i know have had difficulty grappling with both the speed + scale of change at some point, the broader public will also have to grapple with it it will be rough


me when someone asks me something stupid.

jcsemantics's tweet image. me when someone asks me something stupid.

i feel like the ai doomers are going to keep screaming until the bubble pops. like whether it “would have” happened anyways (likely), screaming and throwing fuel on the panic fire is not where it’s at fam


how are humans going to RLHF themselves into novel and ground breaking research.


jc 已轉發

SynthID is more than just the invisible watermark. If someone has SynthID Portal access they could test if this removal tool works. Regardless the watermark is important so when Google does it's next model training runs they exclude AI images and videos to prevent model collapse.


jc 已轉發

Some crazy people on Reddit managed to extract the "SynthID" watermark that Nano Banana applies to every image. It's possible to make the watermark visible by oversaturating the generated images. This is the Google SynthID watermark:

mark_k's tweet image. Some crazy people on Reddit managed to extract the "SynthID" watermark that Nano Banana applies to every image. It's possible to make the watermark visible by oversaturating the generated images.

This is the Google SynthID watermark:

jc 已轉發

turns out 5 is the limit for how many back to back adderall powered all nighters i can push through


jc 已轉發

GPT-5 is no improvement? Come on


this is literally how i use it, and brainstorming/iterating on the same idea through multiple sessions is incredibly effective/powerful.

Sam Altman says OpenAI really does ask its own AI models what to do next. The team uses them to guide strategy, and sometimes the answers are genuinely insightful. “I think when we say stuff like that, people don't take us seriously or literally. But maybe the answer is you…



jc 已轉發

Sam Altman says OpenAI really does ask its own AI models what to do next. The team uses them to guide strategy, and sometimes the answers are genuinely insightful. “I think when we say stuff like that, people don't take us seriously or literally. But maybe the answer is you…

5 years ago, Sam Altman said OpenAI had no plan to make money. The promise to investors? Build AGI -- then ask it how to generate a return. “If AGI is on the table, we take 1/1000th of 1% of the value, return it to investors, and figure out how to share the rest with the…



jc 已轉發

The ghosts of the Hessian Free approach are alive and well in ADAM and other momentum based approaches.

what a beautiful theory!

YouJiacheng's tweet image. what a beautiful theory!


United States 趨勢

Loading...

Something went wrong.


Something went wrong.