micdillon's profile picture. Applied Scientist @ Amazon / Math Nerd

Michael Dillon

@micdillon

Applied Scientist @ Amazon / Math Nerd

Michael Dillon reposted

Releasing the Sparse Transformer, a network which sets records at predicting what comes next in a sequence — whether text, images, or sound. Improvements to neural 'attention' let it extract patterns from sequences 30x longer than possible previously: openai.com/blog/sparse-tr…


Michael Dillon reposted

1. Compute matters, specially for RL agents, but neither convnets nor LSTMs were invented when focusing on scale.

Rich Sutton has a new blog post entitled “The Bitter Lesson” (incompleteideas.net/IncIdeas/Bitte…) that I strongly disagree with. In it, he argues that the history of AI teaches us that leveraging computation always eventually wins out over leveraging human knowledge.



Michael Dillon reposted

One could argue for these loops being genuine topological features of the underlying ImageNet data, since neural nets are a continuous map and t-sne/umap try to preserve neighborhood structure. (Counter argument: be cautious to draw conclusions from t-sne/umap layouts!)


Michael Dillon reposted

Our new paper is out! We proposed a Gaussian-like distribution on hyperbolic space whose density can be evaluated and differentiated analytically. Our Hyperbolic VAEs capture hierarchical structure well. w/ @guguchi_yama, @mooopan, @Masomatics. arxiv.org/abs/1902.02992

ny_ganow's tweet image. Our new paper is out! We proposed a Gaussian-like distribution on hyperbolic space whose density can be evaluated and differentiated analytically. Our Hyperbolic VAEs capture hierarchical structure well. w/ @guguchi_yama, @mooopan, @Masomatics. arxiv.org/abs/1902.02992
ny_ganow's tweet image. Our new paper is out! We proposed a Gaussian-like distribution on hyperbolic space whose density can be evaluated and differentiated analytically. Our Hyperbolic VAEs capture hierarchical structure well. w/ @guguchi_yama, @mooopan, @Masomatics. arxiv.org/abs/1902.02992
ny_ganow's tweet image. Our new paper is out! We proposed a Gaussian-like distribution on hyperbolic space whose density can be evaluated and differentiated analytically. Our Hyperbolic VAEs capture hierarchical structure well. w/ @guguchi_yama, @mooopan, @Masomatics. arxiv.org/abs/1902.02992

Michael Dillon reposted

Working on a little differential geometry notebook with @TensorFlow 2.0 and @matplotlib's pleasant 3d interface.

colindcarroll's tweet image. Working on a little differential geometry notebook with @TensorFlow 2.0 and @matplotlib's pleasant 3d interface.

Michael Dillon reposted

Very nice disambiguation of statistical notation, thanks to @avehtari

ericnovik's tweet image. Very nice disambiguation of statistical notation, thanks to @avehtari

Michael Dillon reposted

Interesting in quickly experimenting with BNNs, GPs, and flows? Check out Bayesian Layers, a simple layer API for designing and scaling up architectures. #NeurIPS2018 Bayesian Deep Learning, Happening now.

dustinvtran's tweet image. Interesting in quickly experimenting with BNNs, GPs, and flows? Check out Bayesian Layers, a simple layer API for designing and scaling up architectures. #NeurIPS2018 Bayesian Deep Learning, Happening now.

Michael Dillon reposted

A fun way to describe Nesterov's momentum:

SebastienBubeck's tweet image. A fun way to describe Nesterov's momentum:

Michael Dillon reposted

We just open-sourced a suite of ODE solvers in PyTorch: github.com/rtqichen/torch… Everything happens on the GPU and is differentiable. Now you can use ODEs in your deep learning models! Credit to @rtqichen.


Michael Dillon reposted

So far we used ODEs for image classification, time series models, and density estimation: arxiv.org/abs/1806.07366. We implemented backprop using the adjoint method, which has constant memory cost. This let us scale ODE-based normalizing flows to SOTA: arxiv.org/abs/1810.01367


Michael Dillon reposted

"Simple, Distributed, and Accelerated Probabilistic Programming". The #NIPS2018 paper for Edward2. Scaling probabilistic programs to 512 TPUv2 cores and 100+ million parameter models. arxiv.org/abs/1811.02091 github.com/google-researc…

dustinvtran's tweet image. "Simple, Distributed, and Accelerated Probabilistic Programming". The #NIPS2018 paper for Edward2. Scaling probabilistic programs to 512 TPUv2 cores and 100+ million parameter models. arxiv.org/abs/1811.02091 github.com/google-researc…
dustinvtran's tweet image. "Simple, Distributed, and Accelerated Probabilistic Programming". The #NIPS2018 paper for Edward2. Scaling probabilistic programs to 512 TPUv2 cores and 100+ million parameter models. arxiv.org/abs/1811.02091 github.com/google-researc…

Michael Dillon reposted

#SundayClassicPaper📜: A Decision-based View of Causality. Lots of work going on in causality, so useful to reflect on this view defining causation in terms of unresponsiveness, and on Howard-form influence diagrams (aistats95). 🍂🎃 citeseerx.ist.psu.edu/viewdoc/downlo…

shakir_za's tweet image. #SundayClassicPaper📜: A Decision-based View of Causality. Lots of work going on in causality, so useful to reflect on this view defining causation in terms of unresponsiveness, and on Howard-form influence diagrams (aistats95). 🍂🎃
citeseerx.ist.psu.edu/viewdoc/downlo…

Michael Dillon reposted

The videos from the 3 lectures I gave at #MLSS2018 are online🌱. If you watch any part of the 4.5 hours leave me some feedback: what's missing, should be left out, reordered, etc Video 1: youtu.be/BgZ_UFsjCD4 Video 2: youtu.be/YdwbVgEsg2M Video 3: youtu.be/B_F8PtU2XMU

shakir_za's tweet card. MLSS 2018 Madrid - Shakir Mohamed 3

youtube.com

YouTube

MLSS 2018 Madrid - Shakir Mohamed 3

The attendees at #mlss2018 Madrid were one of the most engaged I've ever had: so many questions and thoughts😻. I gave 3 lectures on 'Planting the Seeds of Probabilistic Thinking'. 🌱🎲 See the slides here: shakirm.com/slides/MLSS201…

shakir_za's tweet image. The attendees at #mlss2018 Madrid were one of the most engaged I've ever had: so many questions and thoughts😻. I gave 3 lectures on 'Planting the Seeds of Probabilistic Thinking'. 🌱🎲 See the slides here: 
shakirm.com/slides/MLSS201…


Michael Dillon reposted

Thought he’d just focus on Kavanaugh. Instead he gave best response on rape culture I’ve ever heard from a man! I aspire to be this reflective. Damn! #CancelKavanaugh #StopKanavaugh


Michael Dillon reposted

Yordan Zaykov at #PROBPROG: "Infer.NET is now open-source." That's incredible news for the probabilistic programming community! github.com/dotnet/infer


Michael Dillon reposted

Taming VAEs: A theoretical analysis of their properties and behaviour in the high-capacity regime. We also argue for a different way of training these models for robust control of key properties. It was fun thinking about this with @fabiointheuk arxiv.org/abs/1810.00597


Michael Dillon reposted

New book on probabilistic programming on arXiv. I’m sure the authors @hyang144 @jwvdm @frankdonaldwood will welcome feedback. arxiv.org/abs/1809.10756


Loading...

Something went wrong.


Something went wrong.