My earliest ventures into Digital Brushes was during my time at Adobe Design Labs where I created the project Brush Bounty which explored animated brushes and connecting data with them. I showcased the same at Adobe Max 2018:
The AI Brushes project is a project exploring how can we create more expressive and functional digital brushes using Artificial Intelligence . After my time with Adobe, I decided to independently explore how new kind of creative expressions could be created with AI. Here are some of those works:
1. Lava Brush
Lava brush was created by training a pix2pix network with images of lava. I used Google's Tensorflow+Keras for this. I tried it out with a direct feed from the Adobe Fresco app and also by using AR to augment real-life acrylic based painting using the brush.
2. Cloth brush
This was a brush I made to draw textiles or cloth. I keyed out a green screen video of a cloth and used its overall shape as reference for the pix2pix network. I could'nt get detailed folds of the cloth, mainly because of my limited computational power. The possibility of using such brushes in fashion design prototyping is exciting.
3. Product Brush
This was not exactly a brush. This was my attempt to render shaded solids from flat drawings. As such it seemed fitting to use in product design prototyping and I explored this using the use case of designing a chair. The flat vector drawings were made in Adobe XD.
3. Character Action Brush
What if you could draw stick figures and it automatically creates fully formed character illustrations? Imagine drawing characters with different poses as easy as drawing stick figures. Imagine animating them frame by frame using just stick figures. Character animators, I feel would love it. Also it will help lazy artists like me
I combined all the brushes created so far into making an illustration (with some editing in Photoshop)
4. Belle Brush
This is a brush which is derived from annotating images of 47DaysOfDevanagariType (the Devanagari version of 36DaysOfType)
I was able to get this series of images. It looked like inkblots or watercolours. Maybe I could use this for something else?
Here are the images in detail:
What if I could translate line drawings into inkblots/watercolour patches? I ran the same model with a drawing inspired from a stock image (vectorstock). Here is the creation.
After some light tweaks, i got this brilliant watercolour effect. This seemed like an Aha! moment (atleast with this brush). AI Art for the win!
5. Scenery Brush
So I started using a different algorithm and a combination of new tools. I used these algorithms and some pre-existing models to augment physical painting with real-life objects. (That is Augmenting Reality with more reality using Artificial Intelligence assisting human creativity (Fancy :D) ). Although these aren't realtime and works with overlaid video, this could be the future of painting:
This video went viral in Twitter:
6. Object Brush (SPADE)
So I have started enjoying the Gaugan algorithm, that I have started composing imagery with it. Following are the ones I made so far. You can follow thes on my Instagram. These are a bit rugged and low res, but hey, its the beginning of a new era.
Next I made an animation of the segments in After Effects. Each frames were then translated using Gaugan to a real image to create a real life animation render of vector art. This particular piece was done in honour of Hachiko, the loyal dog who waited 9 years for its owner to return. The music is also generated using AI.
To create better drawings, I ported the drawing experience to the iPad using RunwaML and p5js.
Can this be used for interior design? Trials here:
Redesigned the UI and made it a responsive webapp.
An artwork created using the painted segments to augmented real image technique. Featured exclusively on www.makersplace.com/fabinrasheed
I started trying out how one could draw strokes while colour gets added simultaneously using style2paints. Here is a video of an example of this live sketching with colour. Although colourization is common, the simultaneous sketch with colour gave it a fresh expression in the video.
8. Segment and realise
I tried segmenting imagery using the deeplab model in RunwayML and then generating imagery from the segments using Spade by Nvidia. (Not exactly a brush, and by now I have mixed up brushes with art.) Here are some works using the same:
I converted a multiple video montage from pexels using the technique. What I found is that the machine recognises and segments not fully correctly. And due to some segment conversion difference in generation, the subsequent translated video seemed interesting. As if the machine had delusions.
Next up, I used the technique to visualise the effect of climate change using Artificial Intelligence. The realised segments were replaced by equivalent climate after-effect segments, eg. river-gravel, road-river, snowy mountain-mountain, grass-sand etc. Here are some imagery created using the same:
9. Data Visualization
I used Nvidia Gaugan to do a real life visualization of data. What if variation in data can be represented with real-life objects in terms of temporal quantity and distribution. Check out the video below: