Full metadata record
DC FieldValueLanguage
dc.contributor.authorLin, Yong-Xiangen_US
dc.contributor.authorTan, Daniel Stanleyen_US
dc.contributor.authorCheng, Wen-Huangen_US
dc.contributor.authorHua, Kai-Lungen_US
dc.date.accessioned2020-01-02T00:03:28Z-
dc.date.available2020-01-02T00:03:28Z-
dc.date.issued2019-01-01en_US
dc.identifier.isbn978-1-5386-9552-4en_US
dc.identifier.issn1945-7871en_US
dc.identifier.urihttp://dx.doi.org/10.1109/ICME.2019.00046en_US
dc.identifier.urihttp://hdl.handle.net/11536/153327-
dc.description.abstractTraining a deep neural network for semantic segmentation relies on pixel-level ground truth labels for supervision. However, collecting large datasets with pixel-level annotations is very expensive and time consuming. One workaround is to utilize synthetic data where we can generate potentially unlimited data with their corresponding ground truth labels. Unfortunately, networks trained on synthetic data perform poorly on real images due to the domain shift problem. Domain adaptation techniques have shown potential in transferring the knowledge learned from synthetic data to real world data. Prior works have mostly leveraged on adversarial training to perform a global aligning of features. However, we observed that background objects have lesser variations across different domains as opposed to foreground objects. Using this insight, we propose a method for domain adaptation that models and adapts foreground objects and background objects separately. Our approach starts with a fast style transfer to match the appearance of the inputs. This is followed by a foreground adaptation module that learns a foreground mask that is used by our gated discriminator in order to adapt the foreground and background objects separately. We demonstrate in our experiments that our model outperforms several state-of-the-art baselines in terms of mean intersection over union (mIoU).en_US
dc.language.isoen_USen_US
dc.subjectSemantic segmentationen_US
dc.subjectDomain adaptationen_US
dc.subjectGated-convolutionen_US
dc.titleADAPTING SEMANTIC SEGMENTATION OF URBAN SCENES VIA MASK-AWARE GATED DISCRIMINATORen_US
dc.typeProceedings Paperen_US
dc.identifier.doi10.1109/ICME.2019.00046en_US
dc.identifier.journal2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME)en_US
dc.citation.spage218en_US
dc.citation.epage223en_US
dc.contributor.department電子工程學系及電子研究所zh_TW
dc.contributor.departmentDepartment of Electronics Engineering and Institute of Electronicsen_US
dc.identifier.wosnumberWOS:000501820600038en_US
dc.citation.woscount0en_US
Appears in Collections:Conferences Paper