Christopher Hesse hat mit pix2pix ein Teil gecodet, dass Zeichnungen die im Browser entstehen, aus einem Pool von Fotos zu Bildern zusammensetzt.
Ich habe mich mal an einer Katze versucht. Wahrscheinlich sieht das schon besser aus, wenn man eine ordentliche Katze abliefert. Außerdem lassen sich Fassaden, Taschen und Schuhe generieren.
„The pix2pix model works by training on pairs of images such as building facade labels to building facades, and then attempts to generate the corresponding output image from any input image you give it. The idea is straight from the pix2pix paper, which is a good read.
[…]
Trained on about 2k stock cat photos and edges automatically generated from those photos. Generates cat-colored objects, some with nightmare faces. The best one I’ve seen yet was a cat-beholder.
Some of the pictures look especially creepy, I think because it’s easier to notice when an animal looks wrong, especially around the eyes. The auto-detected edges are not very good and in many cases didn’t detect the cat’s eyes, making it a bit worse for training the image translation model.“
(via Martin)