Research is an essential element in Fausto’s practice.
Amplified and driven by the use of Creative Coding, his research is centred on prototyping and aimed to test visual ideas.
It becomes a creative tool, that provides tangible starting points to develop artistic solutions, explore concepts that will nurture and become integrated into broader designing projects.
Fausto's designs and research include also custom made objects and lighting, tailored to meet specific aesthetics and production needs, driving the director work and set design towards an evocative and unique space.
Fake Identities (May 2019)
Working on an app that generates fake identities.
As part of my research for an upcoming project I have started investigating and using coding for topics concerning privacy, personal data, and fake identities.
Digital Self Portrait (May 2019)
A self portrait using generative coding.
Visualisation Tool for a Lighting Installation (May 2019)
First drafts of a visualisation tool I am developing for a lighting installation. These effects have been generated and manipulated in real time.
Prototype Surface Transducer Research (May 2019)
Initial tests using transducer speakers to be used in an installation I am devising in team with a composer. This type of speaker uses any type of solid surface (glass, timber, metal…) to reproduce sound in a wide variety of textures and qualities.
Custom Made Lighting Software and Interface (April 2019)
A couple of screenshots of the software I have developed to control my lighting design during a an artistic residency in April for a contemporary dance show. The lighting effects were produced and manipulated in real time, alongside being triggered live by several sound parameters, achieving unique lighting effects, cohesive and in complete sync with the live soundtrack.
This production has received very positive feedback and will be presented in the first half of 2020.
Tests for a Lighting Installation (Jan 2019)
This is a GIF from one of the tests that I have been running recently. It has been generated by a software I am prototyping to control in-ground lighting for an installation currently in proposal stage. Each dot on the screen will correspond to a lighting fixture. In its final version, the software will be able to control each light individually (1800 individual lights in the example below), allowing the creative team to control their brightness and colour. The software will be used also for pre-visualisation, programming, and to refine the lighting effects, in order to achieve the desired aesthetic and look.
As part of this project, I am also developing a software that can count and provide an accurate x y position for each light/dot, which is a crucial parameter for a correct lighting mapping in this type of application.
Volumetric Lights 1.0 (Nov 2018)
Initial tests on Volumetric Lights, a custom made software I have started developing and exploring in mid 2018. This software allows me to run 3D animations through a grid of lights installed in a physical space. This custom made software has also a visualisation tool, displayed in the screen-captures below. This is an indispensable feature that will help with the effects pre-programming and with discussing specific lighting solutions and overall look.
In the draft examples below, animations are drawn in a complex virtual lighting set up to control several parameters of each lighting fixture.
Interface for Live Control and Manipulation of Lighting Clusters (Oct 2018)
These are screenshots of an interface for a software I have developed to control clusters of lighting and other scenic effects.
This custom made software generates live video content, which is then mapped to lighting fixtures across the stage. producing unique effects with great dynamic range; from on-your-face digital type-of-look, all the way to subtle, minimal and organic.
Ear Training App (Aug 2018)
This is a screenshot for an app I have developed for personal use. This app generates and play notes in different major scale modes, changing root note every two rounds, or anytime desired.
The purpose of this app is ear training and scales recognition for musicians. This app can be connected to any digital work station (like Ableton Live, Logic etc..) or any physical MIDI instrument (like keyboards, synth and so on ...)
This interface was intentionally kept to minimal, and similar to a typewriter document, with letters typed with uneven pressure, and some ink traces. The main scope of this app is focussed around the sound that it generates, and having a specific or over-refined look became secondary.
Prototypes ideas for a CD cover (June 2018)
The content of these compositions is generated by the words used in emails exchanged by the band members. The more recurrent a word is, the larger the rectangle containing that word is.
Modern music bands workflow often revolves around digital platforms, like shared Dropbox folders, Google Drive spreadsheets, shared calendars, and often relies on emails for both creative and production purposes. These digital platforms can become vital tools, and an integral part of the band creative process.
Further refinements to this idea will include a code implementation to omit certain prepositions and articles, in order to let more significant words to come to the surface.
Depth Camera (June 2018)
As part of my research, I have started investigating different uses and potentiality of a depth camera.
The depth camera is a device that is able to read bodies or objects, in relation to the physical position they occupy, regardless of the amount or type of light used on the subject itself.
For this research I have designed an app that can receive depth data from the camera, and manipulate in real time several key parameters such as shifting the distance of the depth to be read, and altering the depth of field itself.
The result is an highly engaging and inspiring platform for the performer, offering a unique and captivating aesthetic.
Choreographers and Performers: Alan Schacher, WeiZen Ho
AFTRS Advanced Lighting Skills Course (Feb 2018)
These stills were taken during a 3 days intensive course at AFTRS on lighting for camera., a workshop/class conducted by director of photography Steve Arnold.
Each day we were lighting several different scenes, covering a wide range of styles and techniques. Taking turns, each person attending the workshop had the chance to lead the lighting set up, seeking feedback and inputs from the rest of the group.
The same actors and set were used for all the scenes, but different lighting set up were discussed and built for each filming exercise.
Prototyping Shadow Places (Aug 2017)
First round of tests for Shadow Places, a large outdoor installation exhibiting video and textile art works, performances and lighting installations.
Part of the scope of these outdoor tests was to find solutions on how to disguise the lighting source while achieving the desired look.
Black and White Photography (May 2017)
These photos are part of a larger collection of stills that were taken on film, then developed in a darkroom, experimenting on several different stages of the process.
Interface Prototype for Interactive Lighting Control (March 2017)
This is a working prototype of a software to control lighting. The lighting intensity, and other lighting parameters if needed, can be manipulated by simply moving the mouse over the screen, in an intuitive and immediate way. This system could be used to control an array of down lights for example, to easily follow the position of a dancer on stage. Note how the lighting intensity decreases the further away they are from the centre of the circle. This type of visualisation provide an highly readable and pleasant look, maximised to facilitate the pre-design stage and live show operation. The larger the circle, the brighter will be that particular light. The number in the centre of each sphere provides a visual feedback to inform the designer about the intensity of each individual light, scaled to a familiar 0 to 100 range.
Interactive Lighting Control (Jan 2017)
This is the initial draft for an interactive lighting control over a cluster of lights.
This draft was then fully developed and used in 2017 in different configurations, and served as prototyping, testing, and lighting research tool. This initial prototypes were subsequently fully developed and used for the design and plotting of Obscene Madame D and The NightLine in 2018.
Animated Textures (Jan 2017)
Digital Textures created with coding softwares. These textures are computer generated with softwares using lines of code to achieve the desired effect. Some of this textures are inspired by motion that exists in nature, like the motion of leaves, fluids, air, where other textures leverage and gravitate towards the digital realm.
Self Portraits (2017)
Some of the self portraits created while experimenting on live camera, creative coding and lighting manipulation.
Research Material (2016)
Exploring custom lighting textures & materials overlay
Research Material (2015)
Overlaying materials, custom gobos and textures