Author Interview - VOS: Learning What You Don't Know by Virtual Outlier Synthesis

  Рет қаралды 6,805

Yannic Kilcher

Yannic Kilcher

Күн бұрын

Пікірлер: 23
@YannicKilcher
@YannicKilcher 2 жыл бұрын
Watch the paper review video here: kzbin.info/www/bejne/n16tZYdpqrF2b68 OUTLINE: 0:00 - Intro 2:20 - What was the motivation behind this paper? 5:30 - Why object detection? 11:05 - What's the connection to energy-based models? 12:15 - Is a Gaussian mixture model appropriate for high-dimensional data? 16:15 - What are the most important components of the method? 18:30 - What are the downstream effects of the regularizer? 22:00 - Are there severe trade-offs to outlier detection? 23:55 - Main experimental takeaways? 26:10 - Why do outlier detection in the last layer? 30:20 - What does it take to finish a research projects successfully? Paper: arxiv.org/abs/2202.01197 Code: github.com/deeplearning-wisc/vos
@brandom255
@brandom255 2 жыл бұрын
I think this new format with 1. Yannic explains paper (1½. Authors watch Yannic's video) 2. Common discussion is perfect! I really love it!
@EternalKernel
@EternalKernel 2 жыл бұрын
Love this format, this brings so much more value to us. Thank you!
@deepblender
@deepblender 2 жыл бұрын
In my view, quantifying uncertainty is going to play a more and more important role. The core idea of this paper to introduce uncertainty as a regularization loss is unexpected and very creative! Love it!
@markmorgan5568
@markmorgan5568 2 жыл бұрын
Agreed.
@markmorgan5568
@markmorgan5568 2 жыл бұрын
This was great. I plan to try it on a very practical problem I’m facing at work. I was glad to see that the “Gaussian” objection turned out to be not really an objection at all - they weren’t saying the feature distribution was Gaussian, just that the method implicitly makes that assumption. One thing I’d really like to hear (these researchers, especially!) thoughts on is whether we might be able to use autoencoders to identify OOD samples. The basic idea being that if a set of features can be reliably reconstructed after passing through a learned bottleneck, those features are likely to be “typical”, which in my mind is basically the same as “in distribution”. And if the reconstruction error is large, then the features are likely OOD. It seems like this would be true at any layer. I’d honestly be interested in thoughts from anyone reading this - I do think this is a reasonable approach, but I had limited success with it and am not sure if there’s a fundamental problem with the approach or if I probably just did something wrong.
@phaZZi6461
@phaZZi6461 2 жыл бұрын
good insights, i was looking forward to this interview!
@alan2here
@alan2here 2 жыл бұрын
Image classifiers in self driving cars could probably do with broad backup categories: solid mover tumbleweed agent animal large animal small animal small spikey animal flying animal With score where applicable for "run", "freeze", "attack", "block", "run on lights", "run on horn", "freeze on lights", "freeze on horn", "car needs to flee", "car needs to close windows".
@Addoagrucu
@Addoagrucu 2 жыл бұрын
last bit about starting the loss after the embedding space starts to take shape, reminds me of KL annealing in training a VAE
@oncedidactic
@oncedidactic 2 жыл бұрын
In retrospect the purpose and techniques of this paper are obvious in their practical utility and theoretical basis. But that always happens after many ideas have been tried and then the “a ha” approach pops out. This really gets me thinking about the other system 1 vs 2 and inter- vs extrapolation conversation going on at MLST for quite a while now. It seems like this outlier or uncertainty measurement will be of course necessary for combining NN (system 1 perception/“autopilot decisions”) with a system 2 process, which is evidently both harder and more expensive. Because you want to offload to system 1 as possible, and save system 2 for when it will pay off with knowledge/policy discovery. But to do that you need to know when system 1 is not giving a good result or is not applicable. Humans seem to have this capability (when they choose to exercise it, haha) and can fluidly extend and recombine system 1 and 2 as best fits the problem. A botanist can immediately tell the difference between “indistinguishable” weeds at 20 meters because they’ve trained themselves to have a plant detection system 1 eye-brain, by focusing on these data. The approach in this paper is a first step for making OD algo a moose-ologist in the same vein.
@ChocolateMilkCultLeader
@ChocolateMilkCultLeader 2 жыл бұрын
Great interview. Lovely to see your work getting attention and recognition.
@johnpope1473
@johnpope1473 2 жыл бұрын
Well done. Great questions and answers.
@zhenfang174
@zhenfang174 2 жыл бұрын
A very good work!
@shawnlee6633
@shawnlee6633 2 жыл бұрын
Kudos to Sharon!
@HD-Grand-Scheme-Unfolds
@HD-Grand-Scheme-Unfolds 2 жыл бұрын
Hello Yannic Kilcher, always a pleasure and benefits watching your channel. Please though may I ask: With "VOS" does this means that it assists against the problem with insufficient robustness to adversarial attacks?
@ofirshifman6472
@ofirshifman6472 2 жыл бұрын
Great paper and interview, thanks! Did you ended up asking them about your comment on the p(x,b,y) notation? where in your first video you said you think there is a typo. I'm asking since I felt the same reading the paper and I'm trying to figure it out. I would like an answer if you know it :-) Thanks again and keep on the good work!
@YannicKilcher
@YannicKilcher 2 жыл бұрын
Yes pretty sure it's a typo :)
@MASTER-qc3ei
@MASTER-qc3ei 2 жыл бұрын
Cat is all you need
@josephharvey1762
@josephharvey1762 2 жыл бұрын
Just want to let you know, Yannic, in case you haven't noticed, the camera is focusing on your mic and not on you.
@abdjahdoiahdoai
@abdjahdoiahdoai 9 ай бұрын
#suggestion maybe ask what they are working on,
@qichaoying4478
@qichaoying4478 Жыл бұрын
Mr Cat distracted me for several times :(
@ketalesto
@ketalesto 2 жыл бұрын
I love the format of two videos. Although, I think now both could be shortened a bit in length. Maybe shorten the paper review video and leave it at about 20 minutes, leaving some details for the authors to explain them. Just as a suggestion. :)
@black-snow
@black-snow 2 жыл бұрын
I object. He already left out a lot, didn't really go into details of the evaluation and didn't even look at the appendix. Not that it was bad, it was great. I just don't think the review could be any shorter without loosing essential information.
Electric Flying Bird with Hanging Wire Automatic for Ceiling Parrot
00:15
女孩妒忌小丑女? #小丑#shorts
00:34
好人小丑
Рет қаралды 100 МЛН
Throwing Swords From My Blue Cybertruck
00:32
Mini Katana
Рет қаралды 11 МЛН
Zombie Boy Saved My Life 💚
00:29
Alan Chikin Chow
Рет қаралды 35 МЛН
One Model For All The Tasks - BLIP (Author Interview)
48:34
Yannic Kilcher
Рет қаралды 7 М.
Can Wikipedia Help Offline Reinforcement Learning? (Paper Explained)
38:35
Has Generative AI Already Peaked? - Computerphile
12:48
Computerphile
Рет қаралды 979 М.
Dynamic Inference with Neural Interpreters (w/ author interview)
1:22:37
AlphaCode - with the authors!
53:46
Yannic Kilcher
Рет қаралды 12 М.
Terence Tao at IMO 2024: AI and Mathematics
57:24
AIMO Prize
Рет қаралды 278 М.
I programmed my own YOLO model to detect Football players (oh it went so well)
21:01
Neural Breakdown with AVB
Рет қаралды 1 М.
Electric Flying Bird with Hanging Wire Automatic for Ceiling Parrot
00:15