Transformers for detection
Multi-Scale Multi-head Attention (MSA)
Attention mechanism that operates on fused features from multiple levels of the feature map, allowing the model to simultaneously capture local and global information for better detection of objects of various sizes.
← Wstecz