Binocular Fusion Net: Deep Learning Visual Comfort Assessment for Stereoscopic 3D

Cited 17 time in webofscience Cited 17 time in scopus
  • Hit : 593
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorKim, Hak Guko
dc.contributor.authorJeong, Hyunwookko
dc.contributor.authorLim, Heoun-taekko
dc.contributor.authorRo, Yong Manko
dc.date.accessioned2019-04-29T01:32:08Z-
dc.date.available2019-04-29T01:32:08Z-
dc.date.created2018-03-28-
dc.date.created2018-03-28-
dc.date.created2018-03-28-
dc.date.issued2019-04-
dc.identifier.citationIEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, v.29, no.4, pp.956 - 967-
dc.identifier.issn1051-8215-
dc.identifier.urihttp://hdl.handle.net/10203/261606-
dc.description.abstractIn this paper we propose a novel deep learning-based visual comfort assessment (VCA) for stereoscopic images. To assess the overall degree of visual discomfort in stereoscopic viewing, we devise a binocular fusion deep network (BFN) learning binocular characteristics between stereoscopic images. The proposed BFN learns the latent binocular feature representations for visual comfort score prediction. In the BFN, the binocular feature is encoded by fusing the spatial features extracted from left and right views. Finally, visual comfort score is predicted by projecting the binocular feature onto the subjective score space. In addition, we devise a disparity regularization network (DRN) for improving prediction results. The proposed DRN takes the binocular feature from the BFN and estimates disparity maps from the feature in order to embed disparity relations between left and right views into the deep network. The proposed deep network with BFN and DRN is end-to-end trained in a unified framework where the DRN acts as disparity regularization. We evaluated the prediction performance of the proposed deep network for VCA by the comparison of existing objective VCA metrics. Further, we demonstrated that the proposed BFN showed various factors causing visual discomfort by using network visualization.-
dc.languageEnglish-
dc.publisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC-
dc.titleBinocular Fusion Net: Deep Learning Visual Comfort Assessment for Stereoscopic 3D-
dc.typeArticle-
dc.identifier.wosid000464149700004-
dc.identifier.scopusid2-s2.0-85044034038-
dc.type.rimsART-
dc.citation.volume29-
dc.citation.issue4-
dc.citation.beginningpage956-
dc.citation.endingpage967-
dc.citation.publicationnameIEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY-
dc.identifier.doi10.1109/TCSVT.2018.2817250-
dc.contributor.localauthorRo, Yong Man-
dc.contributor.nonIdAuthorJeong, Hyunwook-
dc.contributor.nonIdAuthorLim, Heoun-taek-
dc.description.isOpenAccessN-
dc.type.journalArticleArticle-
dc.subject.keywordAuthorBinocular fusion-
dc.subject.keywordAuthordeep learning-
dc.subject.keywordAuthorstereoscopic images-
dc.subject.keywordAuthorvisual comfort assessment-
dc.subject.keywordAuthorvisual discomfort-
dc.subject.keywordPlusDISCOMFORT-
dc.subject.keywordPlusIMAGES-
dc.subject.keywordPlusFATIGUE-
dc.subject.keywordPlusDISPARITY-
Appears in Collection
EE-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 17 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0