MASK AS LOGIC

Reveal and Conceal in Brand Identity in the Generative AI Era

Mask as Logic explores how identity is constructed through what is revealed and what is concealed.

Traditional mask-making builds meaning through layers, with each decision shaping how the mask is understood by a specific audience. This project applies that same logic to brand identity in the age of AI.

Today, non-designers can instantly generate brand visuals. However, without a defined identity, these outputs often feel generic and interchangeable. The problem is not access to design, but rather the lack of structure to define meaning before generation.

This project introduces a system that translates human intuition into a structured identity that can be used consistently across AI tools.

Each year, millions of businesses start without access to professional design. Yet, identity determines whether they are noticed, trusted, or chosen. Although generative AI has made creating visuals fast and accessible, the problem has merely shifted rather than disappeared. AI produces outputs based on common patterns, resulting in brands that look polished but feel indistinguishable. While founders deeply understand their work, without a structured way to translate that understanding into design, they end up generating visuals that lack clear meaning before defining them.

Problem and objective

This thesis draws from the logic of masks, where identity is constructed through decisions about what to reveal and what to conceal. A mask is not decoration, but a system of controlled visibility, designed to communicate in a specific way to a specific audience. Brand identity operates similarly, as a structured set of signals rather than a collection of visuals. The challenge in the AI era is translating human intention, which is intuitive and emotional, into a form that machines can interpret without losing meaning. Maskit was designed as a translation layer to bridge this gap.

The MaskIt system works in two phases that transform intuition into structure. The first phase captures visual instinct through image-based selections, allowing users to define mood, color, typography, and spatial qualities without relying on design language. The second phase introduces questions about intention, audience, and values, refining and aligning the visual foundation with internal meaning.

System logic

All inputs are translated into parameters and processed through a logic system that maintains consistency while adapting tone and strategy, resulting in a structured brand kit that can be used across platforms and AI tools.

Testing with both designers and non-designers showed that while the system was intuitive and efficient, the outputs sometimes felt superficial and not fully aligned with the intended brand meaning. The process emphasized self-definition more than audience perception, and did not fully leverage AI’s ability to interpret inputs. In response, the system was refined to strengthen audience-focused questions, expand visual differentiation, and make the relationship between inputs and outputs more visible, shifting it from a tool that generates results to one that builds understanding.

Testing and iteration

Final Version

Phase 1: external visual landscape

Phase 2: internal identity

Outcome: brand kit generation

The next step is to integrate adaptive AI, allowing the system to respond dynamically to each input rather than relying on fixed questions. This project ultimately shifted my role as a designer from creating visual outcomes to constructing systems that produce them. In the context of generative AI, design becomes the logic that enables identity to move clearly between human intention and machine generation.

Next steps

Next
Next

Eyesee | Packaging