Scenery Image Extension via Inpainting with a Mirrored Input

Naofumi Akimoto, Daiki Ito, Yoshimitsu Aoki

研究成果: Article査読

3 被引用数 (Scopus)


Generative image extension has the advantage of extending the overall image size while preserving the target image because, unlike other image extensions using interpolation, it completes the surroundings of the target image. However, existing generative image extension methods tend to have poor quality in the generation of outer pixels. One method deals only with a limited number of scene classes because the extension repeats the same semantics. We propose a mirrored input, which sandwiches the extended region by mirroring a part of the target image. This replaces generative image extension with an image inpainting problem and thus helps to achieve higher quality pixel generation and can extend semantics with more complex shapes than horizontal repetition. Experimental results show that our proposed method achieves a scenery image extension that exceeds the state-of-the-art generative image extension methods in both visual quality and FID score for datasets containing diverse scenes.

ジャーナルIEEE Access
出版ステータスPublished - 2021

ASJC Scopus subject areas

  • コンピュータ サイエンス(全般)
  • 材料科学(全般)
  • 工学(全般)
  • 電子工学および電気工学


「Scenery Image Extension via Inpainting with a Mirrored Input」の研究トピックを掘り下げます。これらがまとまってユニークなフィンガープリントを構成します。