NVIDAI的AI画笔又进化了,这个视频演示了这款新画笔的威力:NVIDIA’s New AI: Next Level Image Editing! 👌 – YouTube

https://www.youtube.com/watch?v=cS4jCvzey-4

The paper “EditGAN: High-Precision Semantic Image Editing” is available here:


以下为NVIDIA的论文:EditGAN (nv-tlabs.github.io)


EditGAN: High-Precision Semantic Image Editing
 

Huan Ling *1,2,3
Karsten Kreis *1
Daiqing Li 1
Seung Wook Kim 1,2,3
Antonio Torralba4
Sanja Fidler1,2,3

 

1NVIDIA
2University of Toronto
3Vector Institute
4MIT

 

NeurIPS 2021

 

 



 (1) EditGAN builds on a GAN framework that jointly models images and their semantic segmentations. (2 & 3) Users can modify segmentation masks, based on which we perform optimization in the GAN’s latent space to realize the edit. (4) Users can perform editing simply by applying previously learnt editing vectors and manipulate images at interactive rates.

 

Generative adversarial networks (GANs) have recently found applications in image editing. However, most GAN-based image editing methods often require large-scale datasets with semantic segmentation annotations for training, only provide high level control, or merely interpolate between different images. Here, we propose EditGAN, a novel method for high-quality, high-precision semantic image editing, allowing users to edit images by modifying their highly detailed part segmentation masks, e.g., drawing a new mask for the headlight of a car. EditGAN builds on a GAN framework that jointly models images and their semantic segmentations (DatasetGAN), requiring only a handful of labeled examples – making it a scalable tool for editing. Specifically, we embed an image into the GAN’s latent space and perform conditional latent code optimization according to the segmentation edit, which effectively also modifies the image. To amortize optimization, we find “editing vectors” in latent space that realize the edits. The framework allows us to learn an arbitrary number of editing vectors, which can then be directly applied on other images at interactive rates. We experimentally show that EditGAN can manipulate images with an unprecedented level of detail and freedom, while preserving full image quality. We can also easily combine multiple edits and perform plausible edits beyond EditGAN’s training data. We demonstrate EditGAN on a wide variety of image types and quantitatively outperform several previous editing methods on standard editing benchmark tasks.

EditGAN is the first GAN-driven image editing framework, which simultaneously (i) offers very high-precision editing, (ii) requires only very little annotated training data (and does not rely on external classifiers), (iii) can be run interactively in real time, (iv) allows for straightforward compositionality of multiple edits, (v) and works on real embedded, GAN-generated, and even out-of-domain images.

 



 

News

 

  • [February 2022] Code released on Github!
  • [January 2022] EditGAN featured in NVIDIA Developer Blog!
  • [November 2021] Project page released!
  • [November 2021] Paper released on arXiv!
  • [September 2021] Paper accepted to Advances in Neural Information Processing Systems (NeurIPS) 2021!


 

Paper

 

EditGAN:
High-Precision Semantic Image Editing
Huan Ling*, Karsten Kreis*, Daiqing Li,
Seung Wook Kim, Antonio Torralba, Sanja Fidler* Authors contributed equallyAdvances in Neural Information Processing Systems (NeurIPS), 2021

[Paper]      [Code and Interactive Editing Tool]

For feedback and questions please reach out to Huan Ling and Karsten Kreis.

 



 

Demos

 
Left: The video showcases EditGAN in an interacitve demo tool. Right: The video showcases EditGAN where we apply multiple edits and exploit pre-defined editing vectors.
 
Left: The video shows interpolation of editing vectors. Right: The video shows the result of applying EditGAN editing vectors on out-of-domain images.



 

Results

Examples of segmentation-driven edits with EditGAN. Blue boxes: Original images. Orange boxes: Zoom-in view.
We combine multiple edits. Blue boxes: Original images. Edits in detail: Second row, first person: open eyes, add hair, add mustache. Second person: smile, look left. Third row, first car: remove mirror, remove door handle, shrink wheels. Second car: remove license plate, enlarge wheels. Third row, bird: longer beak, bigger belly, head up. Third row, cat: open mouth, bigger ears, bigger eyes
 

Beak size editing. First row: Image and mask pair to learn editing vector. Images are images before editing and after editing. Segmentation masks are before editing and target segmentation mask after manual modification. Second to fourth rows: Applying the learnt edit on new images.
License plate deletion editing. First row: Image and mask pair to learn editing vector. Images are images before editing and after editing. Segmentation masks are before editing and target segmentation mask after manual modification. Second to fourth rows: Applying the learnt edit on new images
 



 

Citation

 

If you find this work useful for your research, please consider citing it as:

Also see prior work on efficient semantic segmentation using GANs, which EditGAN builds on:

DatasetGAN:


SemanticGAN:

发表回复

您的电子邮箱地址不会被公开。