-
Notifications
You must be signed in to change notification settings - Fork 28.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Plain-DETR #27496
Comments
@amyeroberts @rafaelpadilla this open for contribution? |
@kamathis4 Sure! If you'd like to contribute this model feel free to open a PR and ping us when ready for review :) |
@amyeroberts is anyone working on this issue ?? |
@qubvel if no one is working on this isssue would like to raise pr on this...pls let me knew |
Hi @sushmanthreddy, it seems like no one is working on it. Feel free to take this up. 🤗 |
@qubvel Should I follow the modular file approach for this model? I see that Plain DETR is the same as DETR, but with a small addition— a global attention mechanism. |
Hey @sushmanthreddy, sure, the modular would be a great choice in that case 👍 |
Model description
Plain-DETR is an object detector that maintains a "plain" nature: using a single-scale feature map and global cross-attention calculations without specific locality constraints.
In contrast to previous leading DETR-based detectors that reintroduce architectural inductive biases of multi-scale and locality into the decoder.
By leveraging the Object365 dataset for pre-training, it achieved 63.9 mAP accuracy using a Swin-L backbone, which is highly competitive with state-of-the-art detectors which all heavily rely on multi-scale feature maps and region-based feature extraction
Published in the proceedings of ICCV 2023.
Open source status
Provide useful links for the implementation
Official implementation: code
Paper: DETR Does Not Need Multi-Scale or Locality Design
The text was updated successfully, but these errors were encountered: