-
Notifications
You must be signed in to change notification settings - Fork 2.5k
I have some problem about object query. #596
Comments
That's also my confusion. I guess instead of using random values, the embedding weights was used and reshaped. Maybe it's the same. But is it trainable?. I'll appreciate an answer if you've gotten the answer |
I have implemented DETR and found that embedding weights is more convenient than random values when building a model. |
Thanks for helping. |
Yes, so that the gradient of embedding weight will have the same value. |
Thanks for your answer. |
I think it's in (num_queries, batch_size, dim) and not (batch_size, num_queries, dim) |
You are right! |
@Zhong-Zi-Zeng What do you mean with more convenient? Is it that the results are better? Because, as shown in the DETR colab notebook, if you use, |
The last page of the original paper shows a simple code for DETR. The decoder's input is just a random value of size (100, 256). However, on your GitHub, I can't understand what you did about the object query. Why did you use the embedding layer's weight instead of a random value?
The text was updated successfully, but these errors were encountered: