top of page


Membrane mimics an intelligent, flexible, sensitive membrane that changes its properties if stretched. It is a thin, skin-like latex film backlit by a projector, and its appearance changes based on the tension it feels when pushed inwards.

Concept and Design

Membrane possesses a tactile quality akin to skin. It serves as a bridge between humans and machines, mimicking the behavior and sensation of an organism. When pushed, it provides an interaction that feels alive and organic. My goal in bringing future machines to the present as Low-Resolution Beings is to cultivate this experience, prompting people to recognize the naturalness of machines.

The configuration involves using a Kinect sensor to obtain a depth map, which functions as an alpha mask overlaying a visual, whether an image or a video. This technique enables the visual to appear as if emerging from nothingness gradually. I achieved this prototype by positioning and meticulously aligning a projector behind the fabric.

Membrane 2, MIT Music and Theater Arts, W-97, 2024

Prototype 1

This project started as an elective class project at MIT. I took a class (Interaction Design and Projection for Live Performance) in Fall 2023 with Joshua Higgason, a technical instructor and teacher at MIT Music and Theater Arts Department. 

In the class, I learned to use Isadora, an interaction design flow-based programming software I used for the project. 

I used TouchDesigner, a visual programming language that allows users to create real-time interactive multimedia content.

I used the 3D map from the Kinect sensor and overlaid it with the visual I designed on Touchdesigner. After this, I calibrated the numbers in Isadora and gave it a specific range in space that only activated it when someone entered that range. 

Isadora Patch for Membrane, 2024.jpg

Isadora patch for Membrane, 2024

Site and Install Process

The Stata Center hosted the final installation of Membrane 1 as part of a group exhibition featuring other MIT students’ works.

 I borrowed a Kinect sensor from the DMI resources and used the built-in Lidar sensor to get a depth map. The camera on the Kinect takes a real-time three-dimensional map of its surroundings; hence, it knows when something is closer or farther away from it in space. I used this ability to track people’s distance from the sensor and realized there needed to be around 4 feet between the subject and the sensor. After calibrating the sensor, I placed the projector and calibrated it, too, so that the sensor and the projector could simultaneously take and project an image of the same size.

Once this was successful, I tested it using flexible fabric. After a few attempts and calibration, I was able to make it work. 

Install Process, Membrane 1, 2024

 Membrane 1, MIT, 2024

Prototype 2

The second prototype needed to be better in terms of fabrication and the visual that came up after touching the skin. I planned to use a skin-like latex sheet for this purpose. I sourced latex sheeting from Canal Rubber, a trusted rubber supplier in New York City.


I learned much about latex cutting, gluing, and fabricating through YouTube. I installed the second prototype at the Fresh Media show in 2024. It resembled skin and worked precisely as it was supposed to. I received productive feedback, which motivated me to delve deeper into designing more interactive, engaging machines.  

Process and Demo


bottom of page