CSProfKGD's profile picture. #CS Assoc Prof @YorkUniversity, #ComputerVision Scientist Samsung #AI, @VectorInst Faculty Affiliate, TPAMI AE, @ELLISforEurope Member #ICCV2025 Publicity Chair

Kosta Derpanis (sabbatical @ CMU)

@CSProfKGD

#CS Assoc Prof @YorkUniversity, #ComputerVision Scientist Samsung #AI, @VectorInst Faculty Affiliate, TPAMI AE, @ELLISforEurope Member #ICCV2025 Publicity Chair

置頂

Lecture slides for my "Introduction to #ComputerVision" and "#DeepLearning in Computer Vision" courses. 🆕 Gaussian Splatting 🆕 Flow Matching The included videos do not contain voiceovers yet, planned for a future revision.

CSProfKGD's tweet image. Lecture slides for my "Introduction to #ComputerVision" and "#DeepLearning in Computer Vision" courses.

🆕 Gaussian Splatting
🆕 Flow Matching 

The included videos do not contain voiceovers yet, planned for a future revision.
CSProfKGD's tweet image. Lecture slides for my "Introduction to #ComputerVision" and "#DeepLearning in Computer Vision" courses.

🆕 Gaussian Splatting
🆕 Flow Matching 

The included videos do not contain voiceovers yet, planned for a future revision.

Kosta Derpanis (sabbatical @ CMU) 已轉發

I would like you to tell me based on the coffee pattern whether my research proposal will be successful. ChatGPT5: General Impression The residue forms a thick central column leading upward, almost like a tree trunk or a road. This is traditionally read as a path of effort…

KostasPenn's tweet image. I would like you to tell me based on the coffee pattern whether my research proposal will be successful.
ChatGPT5: 
General Impression
The residue forms a thick central column leading upward, almost like a tree trunk or a road. This is traditionally read as a path of effort…

Just submitted my @NSERC_CRSNG Discovery application 😅 Wish me luck 🤞

CSProfKGD's tweet image. Just submitted my @NSERC_CRSNG Discovery application 😅  Wish me luck 🤞

Kosta Derpanis (sabbatical @ CMU) 已轉發

Ranked worst times to open a "decision email": 1. Your birthday 2. Vacation 3. Dinner with loved ones


It’ll never work …

We've trained an unsupervised language model that can generate coherent paragraphs and perform rudimentary reading comprehension, machine translation, question answering, and summarization — all without task-specific training: blog.openai.com/better-languag…



Kosta Derpanis (sabbatical @ CMU) 已轉發

Google and ETH have joined the large scale localization effort with a banger I really did not expect this And now I'm really hoping to make it to NeurIPS where the paper will be presented I'll read it and report a summary here in the next few days

gabriberton's tweet image. Google and ETH have joined the large scale localization effort with a banger

I really did not expect this

And now I'm really hoping to make it to NeurIPS where the paper will be presented

I'll read it and report a summary here in the next few days

the legendary Canadian milk bag 🇨🇦

CSProfKGD's tweet image. the legendary Canadian milk bag 🇨🇦

Kosta Derpanis (sabbatical @ CMU) 已轉發

This week’s #CVPR2026 abstract deadline is FIRM. NO extensions. NO exceptions! Important dates and deadlines: cvpr.thecvf.com/Conferences/20…


Kosta Derpanis (sabbatical @ CMU) 已轉發

Adaptable Intelligence. Multiple possible paths to an objective.


Kosta Derpanis (sabbatical @ CMU) 已轉發

Imagine losing first authorship because you got hit by a blue shell on the last lap 💀

luismbat's tweet image. Imagine losing first authorship because you got hit by a blue shell on the last lap 💀

LLMs are injective and invertible. In our new paper, we show that different prompts always map to different embeddings, and this property can be used to recover input tokens from individual embeddings in latent space. (1/6)

GladiaLab's tweet image. LLMs are injective and invertible.

In our new paper, we show that different prompts always map to different embeddings, and this property can be used to recover input tokens from individual embeddings in latent space.

(1/6)


Kosta Derpanis (sabbatical @ CMU) 已轉發

What really matters in matrix-whitening optimizers (Shampoo/SOAP/PSGD/Muon)? We ran a careful comparison, dissecting each algorithm. Interestingly, we find that proper matrix-whitening can be seen as *two* transformations, and not all optimizers implement both. Blog:…


Kosta Derpanis (sabbatical @ CMU) 已轉發

The Computer Science section of @arxiv is now requiring prior peer review for Literature Surveys and Position Papers. Details in a new blog post


Kosta Derpanis (sabbatical @ CMU) 已轉發

Tesfaldet et al., "Generative Point Tracking with Flow Matching" Tracking, waaaaaay back in the days, used to be solved using sampling methods. They are now back. Also reminds me of my first major conference work, where I looked into how much impact the initial target point has.


Kosta Derpanis (sabbatical @ CMU) 已轉發

I contemplated whether I should post this, because it seems kind of obvious. But it's often taken for granted, so we might underestimate the impact: e.g. these days, diffusion papers don't usually show samples without guidance anymore (figures from GLIDE arxiv.org/abs/2112.10741)

sedielem's tweet image. I contemplated whether I should post this, because it seems kind of obvious. But it's often taken for granted, so we might underestimate the impact: e.g. these days, diffusion papers don't usually show samples without guidance anymore (figures from GLIDE arxiv.org/abs/2112.10741)
sedielem's tweet image. I contemplated whether I should post this, because it seems kind of obvious. But it's often taken for granted, so we might underestimate the impact: e.g. these days, diffusion papers don't usually show samples without guidance anymore (figures from GLIDE arxiv.org/abs/2112.10741)
sedielem's tweet image. I contemplated whether I should post this, because it seems kind of obvious. But it's often taken for granted, so we might underestimate the impact: e.g. these days, diffusion papers don't usually show samples without guidance anymore (figures from GLIDE arxiv.org/abs/2112.10741)
sedielem's tweet image. I contemplated whether I should post this, because it seems kind of obvious. But it's often taken for granted, so we might underestimate the impact: e.g. these days, diffusion papers don't usually show samples without guidance anymore (figures from GLIDE arxiv.org/abs/2112.10741)

Generative modelling used to be about capturing the training data distribution. Interestingly, this stopped being the case when we started actually using them🤔 We tweak temps, use classifier-free guidance and post-train to get a distribution better than the training data.



Kosta Derpanis (sabbatical @ CMU) 已轉發

Our new method for diffusion stitching allows us to generate ultra-long video sequences that follow a long, pre-defined camera trajectory! All segments are generated in parallel (not auto-regressive) and so the model never generates walls that it has to later step through!

Introducing Generative View Stitching (GVS), a non-autoregressive sampling method for length extrapolation of video diffusion models. GVS enables collision-free camera-guided video generation for predefined trajectories, including Oscar Reutersvärd's Impossible Staircase (1/9).



Kosta Derpanis (sabbatical @ CMU) 已轉發

📢 New Paper PointSt3R: Point Tracking through 3D Grounded Correspondence Can point tracking be re-formulated as pairwise frame correspondence solely? We fine-tuning MASt3R with dynamic correspondences and a visibility loss and achieve competitive point tracking results 1/3


Kosta Derpanis (sabbatical @ CMU) 已轉發

This Halloween's HOTTEST costume 🔥🔥🔥

CSProfKGD's tweet image. This Halloween's HOTTEST costume 🔥🔥🔥

Loading...

Something went wrong.


Something went wrong.