You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I've got a short question about the network data input shapes during training (and testing).
Regarding the original pytorch implementation of Ilse et al. it seems that there is a bag of instances [n,x,y,z] with a single label attached to it as a single value. Then you have multiple bags m and in total m labels. Makes sense so far when reading the description in the original publication.
However, now in this implementation we have the same bag of instances [n,x,y,z] but instead of a single label there is array of label instances [n,]. This label instance is indeed derived from the binary bag label (0 or 1) but extended to each instance.
Why is that the case? Is it not possible in tensorflow to train the model in a similar way to the original publication? Maybe I am missing something here but I would be happy if anyone could help me out.
Thank you for your work on this implementation!
Kind regards,
Joshua
The text was updated successfully, but these errors were encountered:
The label instance [n,] is the hidden instance label defined in multiple instance learning, but actually you don't need to know the ground-truth value during your training. In implementation, we set batch size to 1, and each time, Attention MIL pooling will pool those n instance features to one bag feature. After FC layers, the output will be the one predicted bag label for this bag.
Ok thank you I wondered about this a lot. So it's more like a tensorflow-specific implementation detail because it's not possible otherwise?
Is there any reason this could be hindering to extend this approach to multiclass MIL?
Thank you so much for your fast answer :)
Hello,
I've got a short question about the network data input shapes during training (and testing).
Regarding the original pytorch implementation of Ilse et al. it seems that there is a bag of instances [n,x,y,z] with a single label attached to it as a single value. Then you have multiple bags m and in total m labels. Makes sense so far when reading the description in the original publication.
However, now in this implementation we have the same bag of instances [n,x,y,z] but instead of a single label there is array of label instances [n,]. This label instance is indeed derived from the binary bag label (0 or 1) but extended to each instance.
Why is that the case? Is it not possible in tensorflow to train the model in a similar way to the original publication? Maybe I am missing something here but I would be happy if anyone could help me out.
Thank you for your work on this implementation!
Kind regards,
Joshua
The text was updated successfully, but these errors were encountered: