-
Notifications
You must be signed in to change notification settings - Fork 71
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
why large influence is harmful #11
Comments
There is no absolute guidance on what influence numbers are harmful / helpful to the model to make a single prediction. Rather, those are in relation to one another meaning you're able to rank them. |
I have the same question. I'm not sure whether in implementation you calculate I_{up,loss} or -1/n I_{up,loss} in original paper. And I use it in MINIST and find out that the avg contribution is very negative if the chosen train set and test set have the same number. I'm not sure whether something wrong occurs in my experiments. But hope you can check this part.. |
I have the same doubts as @QingXuTHU . It seems from pytorch_influence_functions/pytorch_influence_functions/calc_influence_function.py Line 334 in 66c9a9e
helpful and smallest as harmful . However, I think it should be the opposite: the points with negative I_{up, loss} (z, z_test) should be helpful, and those positive I_{up, loss} (z, z_test) should be harmful (refer to the title of Page 3, Figure 1 of the Koh and Liang paper). I wonder if someone could help check it, thanks!
|
I agree @chengrunyang. On page 6 of the paper, the authors also clarify that the most helpful images are those with the "most positive -I_{up, loss}" (or equivalently the most negative I_{up, loss}). Since pytorch_influence_functions/pytorch_influence_functions/calc_influence_function.py Lines 346 to 347 in 66c9a9e
I stumbled upon this issue while testing how often a training sample is its own most influential sample. Most of the time, the training sample came out as its own most harmful sample, which is counter-intuitive. This bug in the code explains that behavior. |
does there have any evaluation for influence function values? I found the most helpful train image label in CIFAR10 is 2 while the test image label 4.
The text was updated successfully, but these errors were encountered: