Questions on DAM creation
See original GitHub issueHi! Thank you for releasing such a wonderful work. How DAM is generated was a bit unclear to me when reading the paper. Assuming there are N
tokens in total from the encoder (considering one feature level, then N = H x W
), and M
object queries:
- Regarding “In the case of the dense attention, DAM can be easily obtained by summing up attention maps from every decoder layer”, do you mean the cross-attention with shape
N x M
? - Regarding “produces a single map of the same size as the feature map from the backbone”, how is this achieved? Could you help walk through the calculation and the shapes of the tensors?
- Why not directly use the DAM to select the top-k tokens and why have a separate scoring network?
Thanks! I look forward to your reply.
Issue Analytics
- State:
- Created 2 years ago
- Comments:6
Top Results From Across the Web
Questions and Answers About Large Dams | International Rivers
Small dams can be a sustainable and economic source of electricity, especially in rural areas. Q: Are dams an effective method of stopping...
Read more >Questions and Answers about Dams
Q Will dam removal lower the level of water in the river? an impoundment depend on the height of the dam, the shape...
Read more >Dam Engineering - Science topic - ResearchGate
asked a question related to Dam Engineering. Is construction of a huge hydroelectric power generation dam like the GERD of Ethiopia a profitable...
Read more >Basic Knowledge Of Dam Engineering || Questions & Answers
In this article, we explained some general questions answer related to dam engineering. like what is dam, reservoir, maximum flood level, ...
Read more >Basic Civil Engineering Questions and Answers – Dams
This set of Basic Civil Engineering Multiple Choice Questions & Answers (MCQs) focuses on “Dams”. 1. Which of the below is the earliest...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
Yeah, I get this now. The goal is to obtain the
attention_weights
of reference points for each query, instead of theattention_weights
* value of reference points.@JWoong-Shin Thanks for your quick response.
Yes, for some query q, it will obtain value by
A * G((x,y), (x1, y1)) * v(x1, y1) + ...
, whereA
isattention_weights
,G
is a bilinear interpolation kernel, andv
is the value at the point (x1, y1).In other words, for the query q, it references (x1, y1) by A * G((x,y), (x1, y1)). (I think we are having different understanding here).
Therefore, in the perspective of gridpoint (x1, y1), the DAM value is accumulated by A * G((x,y), (x1, y1)) for the query q, and summing over every query, DAM is created. (Sum is not conducted inside the
attn_map_to_flat_grid
method. The method obtains interpolated attention weights in the grid shape, and then obtain DAM by summing over decoder queries and decoder layers: https://github.com/kakaobrain/sparse-detr/blob/1ea7a062ca6d1dd57768d65b14352cfd1a65ab52/models/deformable_detr.py#L408-L409)