Abstract
Unsupervised domain adaptation aims to mitigate the domain gap between the source and the target domains. Despite domain shifts, we have observed intrinsic knowledge that spans across domains for object detection in urban driving scenes. First, it includes consistent characteristics of objects within the same category of extracted ROIs. Second, it encompasses the similarity of patterns within the extracted ROIs, relating to the positions of the foreground and background during object detection. To utilize these, we present DuPDA, a method that effectively adapts object detectors to target domains by leveraging domain invariant knowledge to separable objectness for training. Specifically, we construct categorical and regional prototypes, each of which operates through their specialized moving alignments. These prototypes serve as valuable references for training unlabeled target objects using similarity. Leveraging these prototypes, we determine and utilize a boundary that trains separately the foreground and background regions within the target ROIs, thereby transferring the knowledge to focus on each respective region. Our DuPDA surpasses previous state-of-the-art methods in various evaluation protocols on six benchmarks.
Original language | English |
---|---|
Title of host publication | Computer Vision – ACCV 2024 |
Publisher | Springer |
Publication status | Accepted/In press - 25 Sept 2024 |
Event | 17th Asian Conference on Computer Vision - Hanoi, Viet Nam Duration: 8 Dec 2024 → 12 Dec 2024 https://accv2024.org/ |
Publication series
Name | Lecture Notes in Computer Science |
---|---|
Publisher | Springer |
ISSN (Print) | 0302-9743 |
ISSN (Electronic) | 1611-3349 |
Conference
Conference | 17th Asian Conference on Computer Vision |
---|---|
Abbreviated title | ACCV 2024 |
Country/Territory | Viet Nam |
City | Hanoi |
Period | 8/12/24 → 12/12/24 |
Internet address |