Existing image inpainting methods have shown promising performance in filling the missing regions with visually plausible contents. However, these methods tend to produce distorted structure and blurry texture. To address these issues, in this paper we propose a two-stage inpainting network that combines texture generation and image completion. In the first stage, a texture generator is used to hallucinate texture of the missing regions to guide the reconstruction in the next stage. In the second stage, considering the texture prior would gradually lose its guiding role with the deepening of the network, we adopt residual texture prior to generate fine details. We also introduce a cross-layer contextual attention module which can not only learn contextual attention in decoder feature map, but also benefit from the similar feature shifted from the encoder, generating reasonable structure and realistic texture. Our comparison results of both qualitative analysis and quantitative experiments on Paris StreetView and CelebA datasets demonstrate our proposed method has better inpainting performance than existing methods.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.