Digital Artist: Creative Adversarial Networks(CAN)
Last Updated on March 8, 2021 by Editorial Team
Author(s): Daksh Trehan
Computer Vision, DeepΒ Learning
An interesting approach to make AI better at the art of βfakingβ!
Artificial Intelligence has surely stormed mankind in past years, the machines are extremely good at imitating what we tell them to do. But AI and creativity are counterparts, creativity is the abstract concept that is still missing from the core AIΒ field.
For the past few years, researchers have been trying to decode the machineβs ability to mimic human-level intelligence to generate creative products such as jokes, poetry, problems, paintings, music, etc. The integral aim is to show that AI algorithms are in-fact intelligent enough to produce art without involving human artists but taking into account the human creative products in the learningΒ process.
Several interesting algorithms like GANs(Generative Adversarial Networks) have been introduced to explore the creative space. GANs took the AI field by storm last year by generating human-like fake faces. It implements implicit techniques, that is, it learns with no data being passed to theΒ network.
GANs constitute two networks i.e. Discriminator and Generator. The goal of the generator is to generate βfakeβ images and the discriminator tries to bust βfakeβ images developed by the Generator with the help of provided training data. The generator generates random images and asks for the feedback of the discriminator i.e. whether the discriminator finds it real or fake. At composure, the discriminator wonβt be able to differentiate between randomly generated images and actual images in the training, and the goal of the generator is accomplished. Both Generator and Discriminator are independent of each other yet the iterative process helps both Discriminator and Generator to constantly learn from each otherβs shortcomings to generate even betterΒ images.
Let us infer that we are training our GAN on images of the painting. The motive of the generator is to generate images from training distribution so the discriminator can label them as βrealβ. Ultimately, the generator will start simulating existing art and the discriminator will be fooled right away, the model is generating the new images but the images are far from being called novel or creative. Thus, it can be concluded that GANs ability to generate creative ideas in their original design isΒ limited.
The secret ingredient to developing an algorithm that can think creatively is to link their creative process to human art development throughout time. In simple words, we are trying to mimic exactly how humans develop art. Humans, throughout their lives, are exposed to the various art form and thatβs where they get an impression for their new art, thatβs the exact workflow we want our model toΒ follow.
Experience andΒ Create.
Principle of βLeastΒ Effortsβ
CANs are derived from GANs and are based on Martindaleβs principle, where he argued that the arts are made attractive by increasing their arousing potential that pushes it against its habituation. In simpler words, the art is perceived by viewers when they are presented with something exclusive yet associated with the historical pieces. But, the level of arousal must be controlled and not grow exponentially to gain a negative reaction.
The goal of the model is to satisfy βarousing potentialβ which refers to the levels of excitement in human beings. The level of arousal is least when a person is slept or relaxed and is highest when a person faces fury/danger/violence. Thus, too little arousing potential can be boring, and too much of it can activate the hostile situations. The situation can be best explained by the WundtΒ curve.
Besides, if an artist keeps on producing a similar sort of work then it automatically decreases the arousing potential and leads to the viewerβs distaste. This can be concluded that habituation builds a constant pressure to generate a higher level of arousing potential.
The model tries to boost the arousing potential by increasing the βstylistic ambiguityβ and still avoiding moving too far away from what we accept as art. The architecture of CANs is inherited from GANs with a slightly averted workflow. The discriminator is provided with a large set of art that have been experienced over mankind with their associated style labels as Renaissance, Baroque, Impressionism, Expressionism, etc. The generator is provided with no training example similar to GANs but is designed to accept two signals from the discriminator.
The first signal depicts whether the discriminator classifies the generated image as βart or not artβ. In the traditional GANs, this signal will make the generator change the weights and try again to fool the discriminator by making it believe that the art is coming from the same custom distribution. But in CANs, the discriminator is trained over a large dataset of art, it can accurately distinguish the generated image as βart or not artβ and will only ask the generator to converge into images that are already been accept asΒ βartβ.
The second signal tells the level of accuracy of the discriminator to classify generated art into already defined classes. If the generator, can generate a piece of art that can be regarded as βartβ and can also be easily distinguished by the discriminator into an already defined class then it has successfully fooled the discriminator by generating some art that can be included in so-called humans accepted art. The generator tries to fool the discriminator by making it think of the generate piece as an βartβ and also confuses it about the styles of work that has been generated.
The two signals work opposite to each other, the first one pushes the generator to generate art that can be accepted as an βartβ and if it succeeds it allows the discriminator to classify the images easily. However, the second signal will substantially penalize the generator for doing so, as the motive was to generate style-ambiguous art i.e. something that canβt be easily classified by discriminator but can still be regarded as βartβ. Both signals work independently yet plays an essential role in perfecting each other just like inΒ GAN.
Quantitative Validation
In the above table, the human viewers tried to rate four sets of artwork. The DCGAN is a standard GAN that mimics the established artwork closely but lacks creativity. The Abstract Expresionist dataset is a collection of art that was created between 1945 and 2007, whereas Art Basel 2016 constitutes images displayed in Art Basel 2016, a leading-edge artΒ show.
Unsurprisingly, the images generated from CANs rank highest in arousal potential techniques i.e. Novelty, Surprising, Ambiguity, and Complexity. Also, they fooled the human viewers better into making them believe that the art was generated byΒ humans.
References:
- CAN: Creative Adversarial Networks Generating βArtβ by Learning About Styles and Deviating from StyleΒ Norms
- Creative Adversarial Network
- What are Creative Adversarial NetworksΒ (CANs)?
- CAN (Creative Adversarial Network)βββExplained
- PR-122: CAN: Creative Adversarial Networks
- Creative Adversarial Networks for Art Generation with Ahmed ElgammalβββTWiML TalkΒ #265
- Creative Adversarial Networks (CAN) and artificial intelligence asΒ artist
Feel free toΒ connect:
Portfolio ~ https://www.dakshtrehan.com
LinkedIn ~ https://www.linkedin.com/in/dakshtrehan
Follow for further Machine Learning/ Deep LearningΒ blogs.
Medium ~ https://medium.com/@dakshtrehan
Want to learnΒ more?
Are You Ready to Worship AI Gods?
Detecting COVID-19 Using Deep Learning
The Inescapable AI Algorithm: TikTok
GPT-3: AI overruling started?
Tinder+AI: A perfect Matchmaking?
An insiderβs guide to Cartoonization using Machine Learning
Reinforcing the Science Behind Reinforcement Learning
Decoding science behind Generative Adversarial Networks
Understanding LSTMβs and GRUβs
Recurrent Neural Network for Dummies
Convolution Neural Network forΒ Dummies
Cheers
Digital Artist: Creative Adversarial Networks(CAN) was originally published in Towards AI on Medium, where people are continuing the conversation by highlighting and responding to this story.
Published via Towards AI