Perception-First Design™
Source document for the methodology. ~100 academic citations across cognitive psychology, HCI, behavioral economics, design theory, STS, and ADHD cognitive neuroscience. 15 years of applied work across e-commerce, SaaS, consumer electronics, and entertainment.
Stefan Kovalik · Aurochs · Version 3.6 · April 2026
The Bouncer
I figured out my design methodology in a nightclub.
I was head of security at a 300-capacity venue. Staff of 15. Every night at 1:30 AM, same problem: lights slam on, DJ announces we’re closed, and I’m barking at 300 drunk people to get out. They hated it. I hated it. Fights happened. Stragglers lingered. The whole thing was hostile.
So I tried something different. At 1:10, I walked table by table. “Hey, last call’s coming up. Get any drinks you want. Thanks for coming out tonight.” That’s it. No urgency. No authority voice. Just information, a suggestion, and respect.
At 1:30, when the DJ made the announcement and the lights went up, half the venue was already walking out the door. No barking. No fights. And bar sales went up 30% in those last 20 minutes.
I didn’t know what to call that. I just knew it worked, and I knew why it worked. The old way treated people like obstacles. My way treated them like people with their own goals who’d cooperate if you gave them the right information at the right time with the right tone.
That’s Perception-First Design before I had words for it. Everything I’ve done in the 15 years since; revenue 3x, Costco pallets and Walmart shelves and Disney licensing, improv theaters and smart home brands. It all comes from that same instinct. Design for how people actually perceive and process the world, not how you wish they did.
The Thesis
The design industry has spent 15 years saying the same thing: users have limited attention, so don’t waste it. Lower friction. Reduce steps. Don’t make them think.
My version is different: users don’t think. Not until you make them.
Nobody is thinking. They’re on autopilot. The same way you catch a ball without calculating trajectory, or walk without planning each step; that’s how people interact with websites, products, brands, everything. Autopilot. Energy conservation. The brain handles the vast majority of processing unconsciously. Kahneman (2011) calls this System 1: the fast, automatic mode that runs most of cognitive life. System 2, slow and deliberate and conscious, only activates when something demands attention.
Contemporary cognitive science calls this predictive processing (Clark, 2013; Friston, 2010). The brain is constantly generating predictions about what it expects to encounter. When reality matches prediction: autopilot. When reality violates prediction: attention fires, System 2 kicks in. That prediction error is what I call the activation point.
The design challenge isn’t “don’t waste their limited attention.” It’s that effortful attention is dormant by default. The real question is: when you do make them stop and actually think, when you generate a prediction error that activates them from autopilot, what do they think about? How do they feel? And what do they do next?
That’s what I design for. Not the funnel. The activation.
| The Industry Says | I Say | |
|---|---|---|
| Assumption | Attention is scarce; conserve it | Effortful attention is dormant; design the activation |
| Goal | Reduce friction to goals | Create the right moments that make someone stop, think, and act |
| Failure mode | User runs out of patience | User never engages; autopilots right past you |
| Metaphor | Budget of attention | Autopilot with activation points |
Steve Krug wrote Don’t Make Me Think. Great book. It gave me the vocabulary to talk to stakeholders early in my career. His scope was usability, and he stayed in it deliberately. But nobody’s thinking in the first place. Design for that bias, the bias of making someone stop and engage, instead of funneling them to an action with an affordance of attention they never gave you.
Why I See This
Two things about my brain that I didn’t choose but that turned out to be professional advantages.
I’m autistic. After my formal ASD diagnosis, my partner pointed out something I’d been missing: “Do you realize you’re not picking up on me not being interested at all in what you’re talking about, and you seem to be getting frustrated?” Around 2012, a friend I met at a Chicago meetup about emotions in video games suggested I try improv. I went in already analytical. I had impaired social instincts, not absent ones. I didn’t need to learn what a smile means. I had to learn what eye contact can mean, what a pause signals, what “said vs. unsaid” reveals. I learned it like you’d learn an instrument or ballet: theory, practice, application loops, over and over.
That means I can see and articulate the unconscious patterns that most designers apply intuitively but can’t explain. People interact with websites the same way they interact with other humans. We form first impressions on products the same way we do when someone walks into a room. It’s all unconscious, and we can’t help it.
I also have severe combined-type ADHD. Some experiences are literally painful for me because of how taxing they are on my attention and cognitive load. I’m acutely sensitive to friction, clutter, noise, confusion. The things that bad design inflicts on everyone, I feel at a higher volume. I design because I’m sensitive to these patterns. What most designers theorize about, I feel.
And the ADHD gave me something else: years of building mental prostheses. If it’s out of sight, it’s literally out of mind for me, so I keep post-it notes on my desk for today’s tasks. “Out of sight, out of mind” becomes: what MUST we show vs. what we want to show vs. what we could show. Then cut everything but what we must show.
Design is prosthesis for human cognitive functions and limitations. I know that because I’ve been building prostheses for myself my whole life. The curb cut effect: what I built for my own cognitive accessibility turned out to help everyone.
I’ve come to think of this as a curb-cut effect. The cognitive constraints I designed around — limited working memory, sensitivity to disfluency, impatience with unnecessary friction — are constraints every user has. Mine are just louder. A design that works for the constrained case tends to work better for everyone.
The five PFD layers draw from two largely distinct bodies of cognitive science — perception psychology and ADHD cognitive neuroscience — connected here for the first time. The citation lists share no duplicate entries.
The Conductor Model
I don’t build funnels. I conduct.
Think about a musician who’s reached unconscious mastery. They stop thinking about how things are done and focus entirely on how the music makes you feel. That’s the goal. Prime someone with a first impression, then conduct their thoughts and emotions through the experience. Design a mental waterfall.
There’s a flow of thought and perception: momentum, weight, cadence, tone, speed, context, priming, expectations. Like singing down a scale in front of a crowd and pausing before the last note. Everyone knows what comes next. That recognition, that pattern completion, is satisfying. That’s what a great CTA feels like: the only possible resolution to the experience that preceded it.
The User’s Mental Waterfall
The unconscious sequence of questions every visitor processes:
- What am I trying to accomplish here? → Intent
- Is there a clear path to that? → Navigation
- What’s the point of this site? → Value proposition
- For whom is this? → Do I belong here?
- What is it trying to tell me? → Messaging: ew or cool?
- What is it trying to get me to do? → Call to action
- How do I feel about doing it? → “Prove you’re worth my time” or “Tell me more!”
- Where’s the win-win? → Mutual value
- Does this give me what I want? → Delivery
The hero section is the activation point; the table visit at 1:10. If it answers questions 1 through 5 in the first second or two, the user is primed. Then you conduct them through 6 through 9 as they scroll. If the hero fails, it’s the lights slamming on.
My improv teacher gave me a formula: (Listen) × (Act + React). If Listen equals zero, if the user isn’t attending, then everything else multiplies by zero.
Methodology Siblings
PFD is a practitioner synthesis; the lineage isn’t novel. Three older traditions solve the same problem from different angles.
Kansei engineering (Nagamachi, 1995)
Developed at Hiroshima University in the 1970s, formalized in the mid-90s. Kansei = “psychological feeling, image of a new product.” The methodology maps consumer affective response onto measurable product attributes, then engineers toward the target feeling. This is the closest ancestor to PFD: both start with the pre-verbal response and work backward to design decisions. Kansei is PFD’s Japanese cousin. Anyone doing serious work in automotive, consumer electronics, or product design in Japan already knows it.
Gestalt psychology (Wertheimer, 1923; Metzger, 1936/2006)
The Berlin School’s empirical treatment of perception as organized wholes, not sums of parts. Proximity, similarity, closure, continuity, common fate, figure/ground. These aren’t decorative rules. They’re the first systematic description of how the visual system constructs meaning from raw input. PFD’s L1 (First Impression) and L2 (Processing Fluency) are applied Gestalt. The debt is direct.
Neuroaesthetics (Skov & Nadal, 2020; Leder, Belke, Oeberst & Augustin, 2004)
Modern empirical aesthetics, rooted in Vienna (Leder’s lab) and Copenhagen (Skov’s). Where Kansei engineers the feeling and Gestalt describes the perception, neuroaesthetics maps the neural substrate: amygdala, orbitofrontal cortex, reward pathways. Leder’s 5-stage aesthetic appreciation model (perception → implicit integration → explicit classification → cognitive mastery → evaluation) is a finer-grained L1 subroutine within PFD’s stack.
How PFD differs
PFD is a dependency stack. Upstream failures degrade everything downstream. Kansei maps feeling to attributes but is silent on ordering. Gestalt describes perception but not decision. Neuroaesthetics explains the brain but isn’t a working methodology for a Tuesday afternoon. What PFD adds is the ordering rule and the operational protocol. The sciences are shared; the stack logic is the practitioner contribution.
The Framework
Five layers as a dependency stack. Each depends on the layer below it. Fix them in order, because upstream failures degrade everything downstream. But think of it less as a stack and more as a composition. The foundation sets the key. Each layer adds another instrument. If the foundation is broken, the whole piece is noise.
In predictive processing terms (Clark, 2013; Hohwy, 2013): when a design matches the brain’s predictions at every layer, processing is fluid and unconscious. Autopilot. When a layer fails, it generates prediction error that demands conscious attention. Good design means each layer confirms predictions so the brain’s resources flow toward the decisions that matter, not toward deciphering the interface.
Cognitive Load Reduction
“Can’t perceive anything without the bandwidth to do so.”
Working memory holds about 3–5 chunks at a time, not the “7 plus or minus 2” you’ve seen quoted everywhere. Miller (1956) was describing the recurrence of the number 7 across disparate information-processing phenomena as a rhetorical observation, not proposing a single working memory limit. That nuance got lost. Cowan (2001, 2010) established the modern estimate: roughly 3–5 chunks when rehearsal and grouping strategies are prevented.
Sweller’s Cognitive Load Theory (1988) distinguishes intrinsic load (complexity inherent to the task) from extraneous load (complexity caused by poor design). You can’t change intrinsic load. But every unit of extraneous load you eliminate frees working memory for the decisions that actually matter.
This is L0 because without bandwidth, nothing else works. An overloaded visitor can’t form a first impression, can’t process your brand, can’t connect with your message, can’t find the trail. They’re just gone.
Dodson’s Interest-Based Nervous System model (IBNS) proposes a refinement: ADHD attention is regulated by interest, not importance. The PINCH model identifies five drivers that override cognitive load ceilings: Passion, Interest, Novelty, Competition, Humor. This expands L0 from “reduce cognitive load” to “reduce cognitive load OR increase intrinsic interest.” For ADHD users, and partially for everyone under low motivation, load reduction alone is insufficient when interest is absent. The constraint is not just bandwidth. It is bandwidth weighted by engagement.
In Practice
- Reduce choices at every step. Audit every dropdown, radio button, and “which one?” moment.
- Progressive disclosure: show what’s needed now, reveal the rest on demand.
- Smart defaults: pre-select what 80% of users choose.
- If you don’t need a form field to complete the transaction, delete it. If you need it later, ask later.
ClicSmart App: Started with complex user flows, narrowed down to 3 primary navigation tabs. Radical simplification.
First-Impression Architecture
“Your site has 50 milliseconds. That’s not a design constraint. That’s the physics of how your visitors’ brains work.”
Visual evaluation begins within 50ms, faster than conscious thought (Lindgaard et al., 2006; Tuch et al., 2012). Reinecke et al. (2013) at Google Research found effects at exposures as short as 17ms. The aesthetic-usability effect (Kurosu & Kashimura, 1995; replicated cross-culturally by Tractinsky, 1997) shows that attractive interfaces are perceived as more usable, even with identical functionality.
People form first impressions on products the same way they do when someone walks into a room. Heider & Simmel (1944) showed people attribute intentions and emotions to geometric shapes. Gray, Gray & Wegner (2007) formalized this as mind perception: Agency (capacity to think/act) and Experience (capacity to feel/sense). Users construct lightweight models of competence and care for websites; enough to pass or fail the same trust evaluation we run on humans.
The first impression IS the activation point. It’s the bouncer’s table visit. If it fails, Listen equals zero, and everything multiplies by zero.
In Practice
- The hero section is the thesis statement of the entire site. If the thesis is wrong, everything below it fails.
- Faces exploit neonatal face preference (Johnson et al., 1991) and adult face “pop-out” in visual search (Hershler & Hochstein, 2005).
- Visual quality must match price point. $20–40 tablets in wood cases? I created the perception of “smart home decor” and sold them for $100–200.
Processing Fluency
“If it’s easy to process, it feels true.”
Reber & Schwarz (1999) showed that identical statements in easy-to-read fonts are judged as more true. Alter & Oppenheimer (2009) showed fluency effects generalize across truth, confidence, liking, and trust. Dechêne et al. (2010) confirmed the illusory truth effect across a meta-analysis of 51 studies.
Consistency matters more than creativity. Lakoff & Johnson (1980) showed abstract thought is structured by metaphors rooted in bodily experience. Spence (2011) demonstrated cross-modal correspondences: people consistently match high-pitched sounds with small, bright objects. When sensory input is consistent and fluid across channels, the brain reads it as trustworthy.
Color consistency operates in perceptual space, not physical space. Near-miss color deviations (a brand blue 3% off) are disproportionately costly because they sit in the steep, high-sensitivity zone of the perceptual metric where prediction error is maximal (Bujack et al., 2022). Perceptually uniform spaces (OKLCH, CAM16-UCS) are L2 infrastructure.
Perception Bias Optimization
“It’s not for you. It’s for them.”
Users autopilot decisions, then construct rational explanations after the fact (Nisbett & Wilson, 1977). Kahneman (2011): System 1 makes the decision; System 2 constructs the explanation. Survey data captures the System 2 rationalization. Analytics captures the System 1 decision. The gap between them is where perception problems hide.
In predictive processing terms (Clark, 2013): design for what the brain predicts, not what users report. What users say in surveys is their System 2 reconstruction. What their brain predicted, and how the design matched or violated that prediction, is what actually drove behavior.
Ultimately what works in marketing isn’t “here’s what our product does.” It’s “here’s what having our product makes you feel.” The iPod wasn’t a better product than the Zune. It was about freedom of movement, self-expression, your library in your hands.
Decision Architecture
“They’re hunters looking for prey, and it’s my job to make a trail.”
Every design is a choice architecture. There is no neutral presentation of options. This builds on Thaler & Sunstein (2008), though a Bayesian re-analysis suggests nudge effects may be near-zero after correcting for publication bias (Maier et al., 2022). The structural principles (defaults matter, framing matters, fewer options at decision points reduce paralysis) remain well-supported.
Context qualifier: “Fewer options” applies strongest at high-stakes single decisions and unfamiliar purchase choices. It does NOT apply in expert tool palettes or workspace contexts, where option reduction actively harms productivity. The operative principle: reduce choices at decision points, not choices everywhere.
Barkley’s (1997) model adds another dimension for ADHD users: self-regulation failures occur at the point of performance, not at the point of planning. ADHD temporal processing collapses to “now and not now.” Future benefits have near-zero motivational weight. L4 must make the benefit immediate, not promised. Standard decision architecture assumes people weigh future consequences. For ADHD users, and to a lesser degree for everyone under cognitive load or time pressure, the trail must pay off at every step, not just at the destination.
The Diagnostic
How I actually diagnose a site. Not a checklist; a process.
Step 1: Feel (Pre-Verbal)
Arrive at the page and let the emotional response fire before Wernicke’s area translates it into words. “This makes me feel X.” The feeling comes before the language. System 1 before System 2. My ADHD sensitivity is an advantage here; I feel cognitive load and friction more acutely than most users. The emotional response IS the diagnostic instrument.
Step 2: Unpack
“Why does this feel X?” Now I’m looking for the systematic rule that’s broken: don’t overload, be specific, get to your point, signal-to-noise ratio, everything has a purpose, this should feel X. Does it?
Step 3: Diagnose
| Signal | Likely Layer |
|---|---|
| High bounce + low time-on-site | Layer 2: First impression |
| High time + low conversion | Layer 5: Decision architecture (or trust) |
| High add-to-cart + low checkout | Layer 1: Cognitive load (or price perception) |
| Good desktop + poor mobile | Layer 3: Processing fluency |
| Stated behavior ≠ actual behavior | Layer 4: Perception bias |
Step 4: Prescribe
Fix the lowest failing layer first. Always. Fixing decision architecture when the first impression is broken is like optimizing a trail that nobody enters.
The 5-Minute Perception Audit
1. Path Count (Cognitive Load)
Count every decision from landing to purchase. Compare to competitors. Higher count = losing to friction.
2. 5-Second Test (First Impression)
Show the homepage for 5 seconds. Can they state purpose, feel trust, credit card confidence?
3. Squint Test (Processing Fluency)
Blur your eyes. What stands out? If everything looks equally important, nothing is.
4. Analytics Gap (Perception Bias)
Compare what users say vs. what they do. The gap is where problems hide.
5. Trail Test (Decision Architecture)
Incognito mode. First-time visitor. Core offer in 3 clicks or search bar?
The Generative Protocol
The Diagnostic decodes failures. The Generative Protocol derives solutions. Evaluation and generation are different cognitive operations. Evaluating a proposal against PFD catches surface violations. Deriving a solution bottom-up from PFD’s constraints surfaces requirements that no existing proposal addresses.
- State the Design Problem in terms of user experience, not missing features
- Work Each Layer Bottom-Up. For each: state the constraint, state the violation, derive the MUST requirement, label R[n]
- Accumulate Requirements. R1–R5, all non-negotiable. Lower layers win conflicts.
- Derive the Solution. What satisfies R1 AND R2 AND R3 AND R4 AND R5?
- Test Against Proposals. Which requirements does each proposal fail?
- Identify the Gap. Requirements no proposal satisfies = where the non-obvious solution lives
The Gate
Run before shipping. Each layer is pass/fail. If any layer below L4 fails, do not ship.
| Layer | Gate Question | Pass Condition |
|---|---|---|
| L0 | Can a first-time user complete the primary task without getting stuck? | Zero “where do I click?” moments |
| L1 | Does the first impression activate the right emotion? | 5-second test: users can state purpose + feel trust |
| L2 | Is every element consistent with brand and price point? | Squint test passes. No rogue fonts, colors, spacing. |
| L3 | Does the design speak to what users respond to? | Copy leads with outcomes. Analytics aligns. |
| L4 | Does the trail lead to conversion without manipulation? | Core offer ≤ 3 clicks. Passes all three ethical tests. |
The Communication Hierarchy
Music, humor, and stories: three modalities for changing perception, ranked by power.
Music (Most Powerful)
Music communicates emotion and bypasses rational processing entirely. Koelsch (2014) showed music modulates subcortical emotion structures rapidly, potentially prior to conscious evaluation. Juslin & Västfjäll (2008) identified six mechanisms for music-evoked emotion, the majority bypassing conscious processing. In design: typography rhythm, spacing cadence, visual flow, animation timing, scroll momentum. The “music” of a page controls pacing and emotional state.
Humor (Second)
Jokes change how people feel about a thing by pointing out absurdity and giving permission to think without rigidity. A joke is: setup creates expectation, punch reveals a pattern that either aligns with your self-image or doesn’t, and the anxiety between amplifies the payoff. Laughter is the anxiety being released. Three comedy escalation patterns map to UX: linear repetition (brand consistency), sequential escalation (funnel narrative), exponential escalation (hero → proof → testimonial → CTA).
Stories (Third)
Green & Brock (2000) demonstrated “narrative transportation,” where absorption into a story produces belief changes. Two meta-analyses confirm (van Laer et al., 2014; Braddock & Dillard, 2016). The story isn’t the events. It’s how someone changes from A to B. The user is the protagonist. “I have a problem” → “This solves my problem.” The hero is the inciting incident. The CTA is the climax.
Websites as Relationships
Users evaluate websites the same way they evaluate people. Same unconscious process, same red and green flags.
| Relationship Red Flag | Website Equivalent | Mind Perception |
|---|---|---|
| Doesn’t take appearance seriously | Low-quality design, broken layouts | Low Agency (incompetent) |
| Lack of confidence | Hedging language, unclear value prop | Low Agency |
| Too verbose | Walls of text, burying the lead | Low Agency |
| Non-specific / vague | Generic claims, no proof | Low Agency + Low Experience |
| Too good to be true | Over-promising without social proof | Suspicious Agency |
| Acts like you’re lucky to be there | Self-centered messaging, no empathy | Low Experience |
The trust-killer that operates before anyone reads a word: disrespect. A site that doesn’t treat users with respect, their time, their intelligence, their needs, triggers the same rejection as a person who acts like you’re lucky to interact with them.
The Responsibility
If we’re engineers of thoughts and emotions, we need a Hippocratic oath.
The distinction: The designer controls the perception layer: presentation, UX, visual hierarchy, emotional messaging, activation points. The organization controls the value layer: product quality, pricing integrity, fulfillment, service. PFD’s ethical mandate is to align perception with reality.
When the product is good and the perception is bad (the Simply Smart Home case) PFD closes the gap upward, bringing perception up to match genuine value. That’s not manipulation. That’s correction. The violation occurs in the opposite direction: inflating perception above reality.
The Three-Part Ethical Test
- The Alignment Test: Does this design bring perception closer to reality, or further from it? (Berdichevsky & Neuenschwander, 1999)
- The Sincerity Test: If the user fully understood what this design choice does, would they feel served or exploited? Consumers don’t object to persuasion. They object to persuasion perceived as insincere (Friestad & Wright, 1994; Campbell & Kirmani, 2000).
- The Golden Rule: Would I, as the designer, consent to being influenced by this technique if I were the user?
Dark patterns (Brignull, 2010; Gray et al., 2018), nagging, sneaking, forced action, are the cheap laughs of design. They convert short-term and destroy trust. PFD is the expensive laugh: genuine emotional resonance that compounds.
Relationship to Value Sensitive Design
PFD’s ethical framework sits within the HCI tradition of Value Sensitive Design (Friedman, Kahn & Borning, 2002). Knobel & Bowker (2011) extended this to ask how values get embedded into designed systems, not just studied after the fact, but actively inscribed through design decisions that privilege certain stakeholders, workflows, and outcomes over others.
PFD’s perception layer is exactly such an inscription point: every visual hierarchy, every default, every activation point encodes a judgment about whose attention matters and what they should do with it. Where VSD asks broad questions about which values a system embeds, PFD’s ethical mandate is narrower and operational: does this specific design bring perception closer to reality, for this user, in this context?
What PFD Doesn’t See
Every framework has blind spots: things it makes visible and things it occludes. Knobel’s concept of ontic occlusion (how one representation of reality blocks another from being seen) applies to PFD itself. Naming these isn’t a weakness. It’s a quality criterion.
Structural power dynamics. PFD’s ethical test says “align perception with reality.” But it doesn’t interrogate who decides what “reality” is. When SSH’s product page undersold a good product, the “reality” was clear: the product was better than the perception. But what about a company whose product is mediocre and whose “reality” is genuinely ambiguous? The alignment test assumes a knowable ground truth. In practice, “reality” is contested, and the designer choosing which reality to align with is itself a power move that PFD doesn’t surface.
Individual vs. collective. PFD models one user’s cognitive journey through the mental waterfall. It doesn’t model what happens when thousands of users are conducted through the same waterfall simultaneously. Individual perception optimization can produce collective harms: engagement loops, attention extraction at scale, homogenization of choice. PFD’s unit of analysis is the user. The effects that emerge at population scale are outside its frame.
The fluency trap. Processing fluency is L2’s core mechanism: “if it’s easy to process, it feels true.” That mechanism is value-neutral. It works for honest brands and dishonest ones. A fluent lie is more persuasive than a disfluent truth. PFD’s ethical tests catch deliberate manipulation, but they don’t address the structural problem: by optimizing for fluency, PFD makes it harder for users to engage the critical evaluation that would catch deception. The very thing that makes PFD effective (reducing prediction error, keeping users in autopilot through the desired path) is the same thing that can prevent them from stopping to question. This is ontic occlusion baked into the mechanism.
The designer’s own perception biases. PFD’s diagnostic starts with “Feel”: the designer’s pre-verbal emotional response as the primary instrument. But that instrument is calibrated by the designer’s own culture, class, neurotype, and aesthetic history. I name my neurodivergence as an advantage (heightened sensitivity to friction). But heightened sensitivity to friction could also mean under-sensitivity to other things: social warmth cues, cultural signifiers outside my experience, accessibility needs I don’t personally feel. The diagnostic instrument has its own biases, and PFD doesn’t yet have a protocol for calibrating them.
The Results
Simply Smart Home
The problem: Template site. No unified branding. Feature-focused marketing. Premium price point with a presentation that said “knockoff.”
Every layer was failing: site didn’t answer basic questions (L0, cognitive load), no faces or warmth (L1), template at premium price (L2), feature specs instead of emotional connection (L3), homepage structured around company, not user goals (L4).
What I did: Unified brand system. Emotional messaging: “stay connected, even when you’re apart.” Hero sections with faces. Consistent collateral across every touchpoint. 4-year systematic brand infrastructure build.
- Revenue 3x within a year
- Disney licensing, Costco pallets, Walmart shelf space
- Competitor stole the tagline two years later
iO Theater
“Doesn’t look like a real theater.” Full redesign: dramatic photography, clear event navigation, integrated payment. Online ticket sales 50% → 75%. Same shows. Different perception.
VacuumSealersUnlimited
Embedded design partnership, 5+ years. Revenue 4x running on infrastructure I built and maintain. The long game. Perception compounds.
What’s Teachable
3 Hours with Business Owners
Hour 1: Business Design. Your business IS a design. Feedback loops between your vertical, users, and staff. I’d give them the tools to see their business as a system.
Hour 2: People as Research. Your staff and clients are your greatest resources. Design problem-solving is really just a lens of empathy. “For whom is this?” gives you directionality that frameworks can’t.
Hour 3: Tactics. The 5-Minute Perception Audit. The dependency stack. The red flags. Application, communication, process.
The Teachability Line
The lens (empathy, for-whom, design-as-problem-solving) is fully teachable. The pattern library (hundreds of projects of instant recognition) takes years. The sensitivity (neurodivergent tuning fork for bad design) may not be teachable. But the diagnostic process (Feel → Unpack → Diagnose → Prescribe) can be taught as a substitute.
The Improv Rules
My design practice runs on principles I learned from improv classes. These aren’t metaphors. They’re operational tools.
| Improv Rule | Design Application |
|---|---|
| Yes, and | Accept what exists and build on it. Brick by brick, not castle at a time. |
| (Listen) × (Act + React) | If you’re not listening, you multiply everything by zero. Same for users. |
| “If this is true, what else is true?” | One insight cascades. If this is the brand voice here, it should be everywhere. |
| “What comes before / after?” | User journey thinking. What preceded this page? What follows? |
| “It’s not for you, it’s for them” | Protect the design from stakeholder preferences. |
| Rule of eventually | “That’s horri-bly... good!” Subvert expectations for engagement. |
| Hold topics and return | Progressive disclosure. Introduce, develop elsewhere, callback at conversion. |
| “Familiar to you, novel to others” | Design for their first encounter, not your hundredth. |
The Dual Foundation
(“Foundation” here means evidentiary base: two non-overlapping citation pools. Not the L0 layer described above.)
PFD’s original 55 citations come from perception psychology: cognitive load theory, processing fluency, decision architecture, predictive processing. The Inverse Inflection research (2026) added 22 citations from ADHD cognitive neuroscience. Zero overlap between the two bodies. The v3.6 international expansion (April 2026) added 17 further citations across German, Austrian, Swiss, Japanese, Chinese, French, Dutch, Israeli, and Scandinavian research traditions. ~100 citations total, converging on the same design principles from distinct angles.
The Connection
Why do two separate literatures point to the same principles? Because they describe the same cognitive architecture from different angles. Perception psychology measures how the typical brain processes input: working memory limits, first-impression timing, fluency effects. ADHD neuroscience measures what happens when those same systems operate under tighter constraints: lower working memory ceiling, higher friction sensitivity, sharper temporal discounting. Same architecture, different parameters.
The five PFD layers map cleanly to ADHD neuroscience findings. Every layer has independent support from the ADHD literature, and the support is strongest at the layers that matter most: cognitive load (L0) and Processing Fluency (L2).
PFD-ADHD Layer Mapping
| PFD Layer | ADHD Neuroscience Support | Evidence Tier | Key Citations |
|---|---|---|---|
| Cognitive Load (L0) | WM deficits d=0.69–0.74. Context switching costs approach an hour for ADHD developers. Low perceptual load specifically benefits ADHD-symptomatic developers. | Tier 1–2 (Strong) | Meta-analyses; ICSE 2024; JetBrains 2023 |
| L1 | Reduced latent inhibition means more sensory intake: a badly designed first impression hits harder. Wall of Awful: bad first impressions deposit emotional barriers. Rejection Sensitive Dysphoria amplifies mild negative affect into disproportionate avoidance. | Tier 3 (Indirect) | Carson et al. 2003; Mahan 2020; Dodson |
| L2 | Low perceptual load benefits ADHD developers. Flat associative hierarchies amplify pattern-break sensitivity. Neurodivergent users perceive LLM outputs as “overly neurotypical”: what is fluent for NT may be disfluent for ND. | Tier 1–2 (Strongest extension) | JetBrains 2023; Mednick 1962; Carik et al. 2024 |
| L3 | The METR perception gap: developers thought they were 20% faster with AI but were 19% slower. Overconfidence amplification. The 9% competence penalty for visible AI assistance. | Tier 2–3 | METR 2025 |
| L4 | Barkley’s “point of performance”: self-regulation failures occur in the moment, not in planning. “Now and Not Now” temporal discounting means future benefits have near-zero motivational weight. | Tier 1 | Barkley 1997 |
The Curb-Cut Argument
PFD was not designed for ADHD. But because ADHD involves the most constrained version of the cognitive architecture PFD designs for (lower working memory ceiling, higher friction sensitivity, sharper temporal discounting), designs that work for ADHD users satisfy the tightest constraints. Everyone benefits.
The principle is the same one behind curb cuts, closed captions, and automatic doors. Curb cuts were designed for wheelchair users but turned out to help parents with strollers, travelers with luggage, delivery workers with carts. Designs that reduce cognitive load for ADHD users reduce cognitive load for all users. The constrained case reveals what the general case tolerates but shouldn’t.
Cowan (2010) established that working memory holds 3–5 chunks. ADHD brains hit the ceiling of that range faster and harder, making violations visible sooner. A page that overloads an ADHD user’s working memory is overloading everyone’s. The ADHD user just notices first.
The research supports this directly. Kasatskii et al. (JetBrains, 2023) found that high perceptual load degrades ADHD-symptomatic developers’ performance faster than neurotypical developers’. Mark et al. (2008) measured context-switching recovery at 23 minutes for typical workers. For ADHD developers, Mark et al. (2024) found the cost approaches an hour. Same friction, amplified. Design for the amplified case and the typical case improves too.
The Diagnostic Instrument
I’m autistic and have severe combined-type ADHD. The autism means I learned social cognition analytically: theory, practice, application loops. I can see and articulate unconscious patterns that most designers apply intuitively but can’t explain. The ADHD means cognitive load and friction are literally painful for me. What most designers theorize about, I feel.
That sensitivity is a diagnostic instrument. When I arrive at a page, the emotional response fires before language catches up. System 1 before System 2. My ADHD sensitivity means I feel cognitive load and disfluency more acutely than most users. The feeling IS the diagnostic. The autism means I can then unpack exactly why, tracing the violation back to the systematic rule that’s broken.
I’ve come to think of this as a curb-cut effect applied to cognition. The cognitive constraints I designed around (limited working memory, sensitivity to disfluency, impatience with unnecessary friction) are constraints every user has. Mine are just louder.
The L0 Refinement
Dodson’s Interest-Based Nervous System (IBNS) proposes that ADHD attention is regulated by interest, not importance. The PINCH model identifies five drivers that override cognitive load ceilings: Passion, Interest, Novelty, Competition, Humor.
This means L0’s “reduce load” is incomplete. For ADHD users (and partially for everyone under low motivation), the correct formulation is “reduce load OR increase interest.” This is not a contradiction of L0. It is the specification of when load reduction alone is insufficient. When interest is absent, no amount of cognitive streamlining produces engagement.
Cognitive Constraint Design
The Inverse Inflection research surfaced a unifying meta-framework: Cognitive Constraint Design. The principle: cognitive constraints are design material, not obstacles to work around.
Most design approaches treat cognitive limitations as problems. Reduce choices because people can’t handle too many. Simplify because users are overwhelmed. The framing is defensive: protect the user from their own brain.
CCD inverts this. The constraint is the design input. Working memory holds 3–5 chunks? That’s not a limitation to accommodate. That’s a specification to build for. The constraint tells you what to build, not what to avoid.
| Domain | Constraint | Design Response |
|---|---|---|
| PFD (interfaces) | Limited processing capacity | Reduce cognitive load, optimize fluency |
| ADHD-AI workflows | Executive dysfunction | Externalize structure at point of performance |
| Spatial workflow tools | Working memory limits | Spatial triggers externalize planning |
| AI writing assistants | Communication pattern drift | Corrections-as-training preserves authentic voice |
| Session memory | Context loss between sessions | Hierarchical persistence survives crashes |
Same principle, different surfaces. PFD applies Cognitive Constraint Design to interfaces. The five layers (L0 through L4) are a systematic treatment of perceptual and cognitive constraints as design specifications. Each layer names a constraint (bandwidth, first impression, fluency, bias, decision load) and prescribes a design response grounded in how that constraint actually operates.
CCD is the genus. PFD is a species. This distinction matters because it explains why PFD’s approach transfers across domains. The bouncer story, the improv rules, the diagnostic protocol: they all work because they’re applications of the same principle. Observe the constraint, design for it, test whether the design honors it.
The methodology page focuses on PFD as a design tool. This reference documents the broader intellectual framework. CCD is for researchers and practitioners who want to understand why PFD’s approach to interfaces also explains how to build session memory systems, voice preservation tools, and spatial orchestration canvases. The connecting thread is always: what constraint does the user face, and how do you design the system so the constraint becomes productive rather than destructive?
Practitioner Cognition
This section explains HOW PFD’s diagnostic produces insights that standard UX heuristics miss. It is not a different checklist. It is a different cognitive operation.
Bisociation
Koestler (1964) coined bisociation for the creative act of connecting two independent frames that don’t normally intersect. Standard association operates within a single frame. Bisociation jumps between frames to produce something neither frame contains alone.
PFD’s cross-domain structure (psychology + design + improv + neuroscience) is bisociative by architecture. The derivation protocol forces collision between constraint frames. When you work Layer 0 (cognitive load theory) and then Layer 3 (perception bias), the accumulated requirements R1 through R3 create a constraint space that neither field produces independently. The non-obvious solution lives in that collision.
Janusian Thinking
Rothenberg (1971) described Janusian thinking: the simultaneous conception of opposites as a creative process. PFD holds “reduce friction” AND “activate attention” simultaneously. The core thesis is literally Janusian: users don’t think, until you make them. Don’t make them think (L0). Make them think (activation point). Both are true. The resolution is not choosing between them but designing for the transition.
Deliberate and Spontaneous Modes
Dietrich (2004) mapped four creativity types to prefrontal function: deliberate/spontaneous crossed with cognitive/emotional. Creative insight involves both deliberate (prefrontal-guided, analytical) and spontaneous (defocused, associative) processing.
The PFD diagnostic protocol alternates between these modes. The Feel step is spontaneous: pre-verbal, emotional, System 1. The Unpack and Diagnose steps are deliberate: analytical, systematic, tracing violations back to specific layers. The protocol is not “first feel, then think.” It is “alternate between modes, and use each mode for what it does best.”
This is why the Feel step cannot be skipped. A pure checklist audit (deliberate mode only) catches surface violations but misses the emotional gestalt. A pure intuitive response (spontaneous mode only) identifies that something is wrong but cannot diagnose which layer is failing. PFD needs both.
Flat Associative Hierarchies
Mednick (1962) showed that creative people have flatter associative hierarchies: more distant conceptual connections are accessible. Where most people associate “table” with “chair” and stop, creative thinkers reach “table” to “periodic table” to “elemental structure” to “information architecture.”
Carson, Peterson & Higgins (2003) found eminent creative achievers were 7x more likely to have low latent inhibition: their brains let in more raw sensory and conceptual material. ADHD may produce flatter hierarchies via reduced latent inhibition through related mechanisms. The connection is indirect but convergent: ADHD involves reduced inhibitory filtering, creativity correlates with reduced inhibitory filtering, and PFD’s cross-domain derivation protocol rewards exactly the kind of distant associations that flat hierarchies produce.
The Communication Hierarchy
Three modalities for changing perception, ranked by power: Music > Humor > Stories.
Music communicates emotion and bypasses rational processing entirely (Koelsch, 2014; Juslin & Västfjäll, 2008; Huron, 2006). In design: typography rhythm, spacing cadence, visual flow, animation timing, scroll momentum. Humor changes how people feel about a thing by pointing out absurdity and giving permission to think without rigidity. Stories produce narrative transportation (Green & Brock, 2000; van Laer et al., 2014): absorption into a narrative produces belief changes.
The practitioner who understands all three modalities has more tools for changing perception than the practitioner who only understands information architecture. This is why PFD draws from improv, music theory, and narrative psychology alongside cognitive science. The communication hierarchy is not a bonus chapter. It is how the methodology gets delivered.
Evidence Tiers
Not all claims carry equal weight. This classification system governs what PFD states confidently, what it attributes carefully, what it frames as synthesis, and what it avoids. Every public claim should be traceable to one of these tiers.
Tier 1: Established Science
Well-replicated findings from peer-reviewed research, confirmed by multiple independent labs. PFD states these as facts.
| Finding | Source(s) | Why Established |
|---|---|---|
| Working memory deficits in ADHD (d=0.69–0.74) | Meta-analyses, Barkley model | Decades of convergent evidence, hundreds of studies |
| ADHD associated with stronger divergent thinking | White & Shah 2006/2011, Boot et al. | Multiple independent labs, peer-reviewed, conceptual replications |
| DMN hyperconnectivity in ADHD | Castellanos et al. | One of the most replicated neuroimaging findings |
| Creative thinking requires DMN + ECN + Salience cooperation | Beaty et al. (Three-Network Model) | Multiple fMRI studies, established in cognitive neuroscience |
| 7x rate of low latent inhibition in creative achievers | Carson, Peterson & Higgins 2003 | JPSP (top-tier), widely cited |
| Flat associative hierarchies produce remote associations | Mednick 1962 | 60+ years of citation, standard creativity psychology |
| DRD4/7R: opposite fitness outcomes by environment | Eisenberg 2008 | Well-designed field study, conceptually replicated |
| Cognition extends into reliably accessed external tools | Clark & Chalmers 1998 | One of the most cited papers in philosophy of mind |
| Context switching costs significant recovery time | Gloria Mark (multiple), ICSE 2024 | Well-established productivity research |
| Self-regulation, not attention, is the core ADHD deficit | Barkley model | Dominant ADHD theoretical framework, decades of research |
Public framing: “Research shows...” or state directly without hedging.
Tier 2: Strong Single Studies
Solid methodology, credible venues, but not yet independently replicated. PFD cites the specific source.
| Finding | Source | Limitation |
|---|---|---|
| ADHD is the only positive predictor of divergent thinking in engineering students | Taylor 2020 (n=60) | Single university, single lab |
| Emotional prompting improves LLM output ~11% | Li et al. 2023 (EmotionPrompt) | ArXiv preprint; 11% human eval is more defensible than 115% peak |
| 25% higher AI satisfaction among neurodivergent workers | UK Dept. for Business and Trade | Measures satisfaction, not performance |
| ND users perceive LLM outputs as “overly neurotypical” | Carik et al. 2024 | Peer-reviewed (ACM), specific study context |
| Low perceptual load benefits ADHD-symptomatic developers | JetBrains 2023 | Industry research, single study |
Public framing: “A study by [X] found...” or “Research from [X] suggests...” Always name the source.
Tier 3: Reasonable Inferences
Logical connections drawn between established findings. The foundations are real. The bridge is untested. PFD presents these as synthesis, not established research.
| Inference | Evidence Base | Untested Bridge |
|---|---|---|
| ADHD communication style is natural emotional prompting | EmotionPrompt + IBNS (Dodson) | Nobody has tested whether ADHD individuals produce better prompts |
| ADHD-AI error patterns are complementary | PNAS 2022 Bayesian framework | Zero studies measuring actual ADHD-AI error correlation |
| Centaur model generalizes beyond chess | Kasparov/ZackS 2005 | Chess has perfect information. Software development does not. |
| PFD disproportionately benefits ADHD users | WM deficits + L0/L2 research | PFD was not designed for ADHD. Connection is real but unmeasured. |
Public framing: “Our synthesis suggests...” or “The evidence points toward, though this hasn’t been directly tested...”
Tier 4: Speculative or Corporate
Claims that lack methodology, independent verification, or peer review. PFD either avoids these or flags them explicitly as unverified.
| Claim | Source | Problem |
|---|---|---|
| JPMorgan ND program: 90–140% higher productivity | Corporate self-report | No sample size, no comparison group, no methodology published |
| Deloitte “30% more productive” ND teams | Industry report | No peer review, no methodological detail |
| The “100+ sources” claim from popular ADHD-AI sites | Self-report | Counts blog posts alongside peer-reviewed papers |
Public framing: Do not use, or explicitly state: “Corporate reports claim...” Never present as research.
Accumulated Learnings
18 learnings from applying PFD across interfaces, spatial canvases, AI systems, pricing, ethics, and communication. These are not theoretical extensions. They emerged from practice and changed how the framework is applied.
| # | Learning | Domain | PFD Layer(s) |
|---|---|---|---|
| 1 | Generative vs. Evaluative | Core method | All |
| 2 | Access vs. Signal Mechanisms | Marketing / UX | L1–L4 |
| 3 | Pre-Send vs. Post-Response | AI / Communication | L1, L4 |
| 4 | Workspace vs. Product Separation | Product design | All |
| 5 | Recursive Validation | Meta-methodology | All |
| 6 | Infrastructure ≠ Activation | Product design | L1, L4 |
| 7 | Consistent Gap Identification | Core method | L2 |
| 8 | Epistemic Asymmetry as L4 Problem | AI systems | L3, L4 |
| 9 | Temporal/Session Continuity | Spatial tools | Gap (candidate L6) |
| 10 | Constraints Are Distributions | Accessibility | L0 |
| 11 | Visual Channel Utilization Audit | Data viz / Canvas | L2 |
| 12 | Route vs. Survey Knowledge | Spatial tools | L2, L4 |
| 13 | Keyboard Density Is L2, Not L4 | Expert workspaces | L2 |
| 14 | Ethics Are Operational, Not Structural | Ethics | Ethics |
| 15 | Experiential Self-Contradiction | Hero / landing | L2, L3, L4 |
| 16 | Near-Miss Color Asymmetry | Design systems | L2 |
| 17 | Iterative Regression Is Deeper Visibility | Forge / analysis | All |
| 18 | Backend Mechanics as Frontend Tiers | Pricing / SaaS | L1–L4 |
#1: Generative vs. Evaluative
When analysts were given solutions first and asked to evaluate through PFD, they missed requirements entirely. When forced to derive bottom-up, the gaps emerged. PFD only generates non-obvious solutions when used bottom-up with constraint accumulation. Top-down evaluation refines but cannot escape the originating frame of the solution being evaluated. The derivation protocol exists because evaluation is not generation.
#2: Access vs. Signal Mechanisms
PFD’s layers distinguish between “how do I access content?” and “how does the user know where content exists?” L1/L2 handle signaling. L3/L4 handle access. Conflating them produces solutions that solve access without solving discovery. A CTA that’s easy to click (access) is different from a page that makes the CTA feel like the obvious next step (signal).
#3: Pre-Send vs. Post-Response
Standard UX says “show your sources after the response” (attribution). PFD derivation says “show what the AI will see before the user sends” (legibility). Attribution is an L3 debugging tool. Pre-send legibility is an L1+L4 schema and trust tool. What you show BEFORE the ask is higher-leverage than what you show AFTER. This generalizes to communication and copy: setup and framing outweigh explanation and justification.
#4: Workspace vs. Product Separation
When analyzing a specific workspace through PFD, recommendations naturally split between workspace-level fixes (rearrange, rename, add zones) and product-level features the tool should provide. The separation between “user organizes content” and “tool provides perceptual infrastructure” is a critical design boundary. PFD should be applied at both levels, but the intervention target must be identified correctly.
#5: Recursive Validation
When PFD is used to analyze a workspace that is ABOUT PFD, three levels operate simultaneously: PFD as framework, PFD applied to the workspace, the workspace as proof-of-concept for the tool. When all three levels are legible, the workspace becomes its own best case study. This recursive structure is a strength, not a liability, but only when intentionally designed.
#6: Infrastructure ≠ Activation
Infrastructure that exists without activation is worse than no infrastructure. It creates the illusion of completeness while users experience the original problem. PFD correctly distinguishes between “the mechanism exists” and “the user can perceive and use the mechanism.” This is a Layer 4 violation masquerading as a Layer 1 problem. Having the right product without the right perception layer produces the same failure mode.
#7: Consistent Gap Identification
When the same requirement (R2: pre-attentive content signaling) surfaced independently across four different analytical frames, it was not a preference. It was a genuine perceptual constraint. When running multiple derivations on the same problem, note which requirements appear consistently. Those are load-bearing.
#8: Epistemic Asymmetry as L4 Problem
When an AI system gathers context via invisible mechanisms (graph traversal, priority weighting, depth limits), the resulting epistemic asymmetry (AI knows more than user about its own context) is a Layer 3 violation. Users’ System 1 response to opaque behavior is distrust. The intervention is making the mechanism visible pre-send, not explaining it post-response. This generalizes to any expert/client asymmetry: show your reasoning process, not just your conclusions.
#9: Temporal/Session Continuity
PFD addresses what the user perceives NOW but not how they re-enter a workspace after 48 hours. Six of nine independent analytical personas identified this gap. A workspace visited after two days is a different perceptual object than the same workspace visited five minutes ago. PFD’s five layers assume continuous interaction and need supplementation for interrupted/resumed sessions. Candidate for formal L6 after testing.
#10: Constraints Are Distributions, Not Constants
PFD states working memory holds “~3–5 chunks” as a hard constraint. This is a population mean. Neurodivergent users, users under cognitive load, users with fatigue all shift the distribution. The constraint is real but the threshold varies. PFD’s constraints define the shape of the curve, not a fixed cutoff. Good design provides calibration mechanisms for users at the tails. This strengthens PFD by acknowledging that perception science describes populations, not individuals.
#11: Visual Channel Utilization Audit
Bertin’s 7 visual variables (position, size, lightness, texture, color, orientation, shape) provide a systematic audit checklist for design. Most UIs use 2–3 channels. Under-utilized channels mean information that COULD be signaled pre-attentively is hidden behind interaction. The derivation protocol should include a channel audit: “How many visual channels does this design use? Which are available but unused?”
#12: Route vs. Survey Knowledge
Siegel & White’s spatial knowledge theory: users develop route knowledge (sequential, egocentric, fragile) before survey knowledge (allocentric, configurational, robust). Canvas tools should support both stages. New users need route affordances (guided paths, breadcrumbs, landmarks). Experienced users need survey affordances (minimap, overview, configurable views). “Appropriate density” depends on which knowledge type the user has developed.
#13: Keyboard Density Is L2, Not L4
When multiple analysts flagged “too many keyboard shortcuts” in an expert tool, the initial diagnosis was L4 choice overload. After applying the Scheibehenne (2010) context qualifier (choice overload is near-zero in expert workspaces), the problem reframed to L2: users cannot predict what a keypress will do across cognitive modes. The fix is a keyboard mode indicator, not fewer shortcuts. Don’t reduce expert affordances. Signal which mode they operate in.
#14: Ethics Are Operational, Not Structural
PFD’s three ethical tests (Alignment, Sincerity, Golden Rule) are operational ethics: executable by one designer in a single review session. They don’t constitute structural ethics in the Value Sensitive Design sense (systematic processes that produce good outcomes regardless of individual practitioner values). Long-term client relationship accountability enforces alignment better than any framework clause could. Operational tests handle decision-making. Structural incentives handle accountability. When PFD scales beyond solo practice, structural ethics needs to be baked into the process, not just carried by the practitioner.
#15: Experiential Self-Contradiction
When a hero section’s thesis claims speed and simplicity (“Your customers decide in milliseconds. I design for that.”) but the hero itself takes 10+ seconds to parse due to competing visual elements, it disproves its own claim experientially. This is an L3/L4 failure, not an L1 failure: the visitor’s lived experience contradicts the promise being made. The resolution preserves complexity in the background (competence signal) while simplifying the foreground (single message, single CTA). A diamond on crushed velvet. The complexity is the bed, not the message.
#16: Near-Miss Color Asymmetry
Human color perception is non-additive (Bujack et al., 2022, PNAS). A brand blue that’s 3% off-target imposes disproportionate processing cost compared to a completely different color, because small deviations sit in the steep, high-sensitivity zone of the perceptual metric where prediction error is maximal. Near-miss color deviations are the highest-cost L2 violations, more disruptive than using a completely different color. Design-token linting should weight near-misses as more severe than far-misses. Color structure is hardware (universal geometry). Color meaning is software (cultural, context-dependent).
#17: Iterative Regression Is Deeper Visibility
When iterative PFD analysis shows a score DROP around rounds 3–6, the cause is almost never that fixes made things worse. Resolving surface-level issues (nav overload, font proliferation) unmasks deeper problems that were previously invisible. The dominant signal was drowning out subtler violations. The regression reflects updated understanding, not degraded quality. Three terminal states for iterative analysis: converged (done), impasse (approach won’t converge with current premise), abandoned (user stopped). The impasse state is the highest-value signal most iterative tools lack.
#18: Backend Mechanics as Frontend Tiers
When internal system state (subscription tiers, feature flags, scarcity counters) is surfaced as visible UI complexity, it converts a backend concern into a frontend cognitive load violation. Founding member pricing with counter badges, tier comparisons, and crossed-out prices triggers defensive pattern-matching against growth hacks. The user’s 50ms L1 read is “this is a marketing tactic,” which directly contradicts the intended signal of gratitude. The system needs tiers. The user needs a yes/no decision. These are different design surfaces at different layers. A rare 4-layer simultaneous violation from a single UI element.
Framework Comparison
PFD exists in a landscape of design frameworks. Understanding where it overlaps and where it differs helps practitioners decide when to use it and when something else serves better.
| Framework | What It Does | Where PFD Differs |
|---|---|---|
| Laws of UX (Yablonski) | Collection of psychological principles for design | Descriptive catalog, not generative protocol. No dependency stack. No derivation method. Tells you what to know, not what to do first. |
| Fogg Behavior Model | Behavior = Motivation + Ability + Prompt | Single-event model. PFD addresses the perceptual pipeline BEFORE behavior triggers. Fogg assumes the user is already attending. |
| Don’t Make Me Think (Krug) | Reduce friction, make things obvious | PFD’s Cognitive Load layer (L0). Krug covers L0 deliberately and well. PFD adds four more layers on top, plus a generative protocol and ethics framework. |
| Emotional Design (Norman) | Visceral, behavioral, reflective processing levels | Complementary framing. Norman’s visceral ≈ PFD L1. PFD provides a dependency stack and generative protocol Norman doesn’t. |
| Nudge Theory (Thaler/Sunstein) | Choice architecture shapes behavior | PFD L4. But Maier et al. (2022) shows nudge effects near-zero after correcting for publication bias. PFD contextualizes when choice architecture works and when it doesn’t. |
| Cognitive Constraint Design | Meta-framework: constraints as design material | PFD is an application of CCD to interface design. CCD is the genus. PFD is a species. |
None of these frameworks are wrong. They are incomplete in different ways. Laws of UX catalogs principles without a generative method. Fogg models behavior but not perception. Krug covers L0 without the other four layers. Norman identifies processing levels without prescribing an order. Nudge Theory applies L4 without the dependency stack that determines when L4 interventions actually work.
PFD’s contribution is the dependency stack: five layers in a specific order, where each layer depends on the layer below it. Fix them in order, because upstream failures degrade everything downstream. This ordering principle is what separates PFD from frameworks that treat design principles as independent, parallel concerns. They are not parallel. They are stacked.
Make Me Think: Chapter Mapping
“Make Me Think: Perception-First Design for the Post-Usability Era” is a 12-part series that develops the framework through narrative, case studies, and citations. Each chapter maps to specific PFD layers and concepts.
| Ch. | Title | PFD Layer(s) | What It Adds |
|---|---|---|---|
| 1 | The Bouncer | Origin (all layers) | The nightclub story: how a security job became a design methodology |
| 2 | They’re Already Not Thinking | Core thesis | Predictive processing. Why most problems happen before conscious evaluation. |
| 3 | The Tuning Fork | Dual foundation | Autism as analytical social cognition. ADHD as friction sensitivity. |
| 4 | Cognitive Load | Cognitive Load (L0) | Miller vs. Cowan, Sweller’s CLT, progressive disclosure, smart defaults. |
| 5 | The 50-Millisecond Verdict | L1 | First impressions, the iO Theater case study, visual-price coherence. |
| 6 | The Feeling of Truth | L2 | Processing fluency, the McGurk Effect, consistency compounding trust. |
| 7 | The Gap | L3 | Perception bias. System 1/System 2 gap. Why a tagline change preceded revenue tripling. |
| 8 | The Trail | L4 | Decision architecture, path reduction, smart defaults. |
| 9 | Feel, Unpack, Diagnose, Prescribe | Diagnostic protocol | Emotional primacy, embodied cognition, why feelings arrive before language. |
| 10 | Music, Humor, Stories | Communication hierarchy | Three tools for changing perception, ranked by power. |
| 11 | The Oath | Ethics framework | Three operational tests, structural ethics, cognitive deskilling. |
| 12 | What I Don’t Know Yet | Limitations | WEIRD evidence base, the fluency trap, ontic occlusion, living methodology. |
The series is available at aurochs.agency/writing/make-me-think/. Each chapter stands alone. Read them in any order, though the dependency stack (Chapters 4–8) makes more sense sequentially.
Open Questions
These are genuinely unresolved. A living methodology names its edges.
Cultural generalizability (partially resolved, v3.6). The evidence base is predominantly WEIRD (Western, Educated, Industrialized, Rich, Democratic). The underlying cognitive architecture is expected to generalize: predictive processing, working memory limits, first-impression formation. But the parameters are culturally shaped: which visual patterns signal trust, what counts as “too much” cognitive load, how fluency maps to credibility.
Recent cross-cultural research resolves part of this. Masuda & Nisbett (2001) showed East Asian observers attend to background and relational context while Western observers focus on foreground and focal objects. Chua, Boland & Nisbett (2005) replicated this with eye-tracking: different scanning patterns on the same scenes. Goh et al. (2010) found cultural differences in face-processing cortex activation. Ji, Nisbett & Su (2001) showed cultural differences in prediction direction (linear vs non-linear reasoning about change).
What this resolves: the architecture is universal; the calibration is cultural. PFD’s 5-layer stack holds across populations. But L1 parameters (what triggers trust) and L3 parameters (what predictive priors load automatically) must be validated against target-market users before cross-regional rollout.
What it doesn’t resolve: L2 processing-fluency effects across non-Latin typographic traditions, L4 decision-architecture priors across markets with different default authority structures, and the relative dependency-stack ordering in high-context vs low-context cultures. These stay open.
Measurement. PFD lacks a validated measurement instrument. The 5-Minute Perception Audit is a practitioner heuristic: useful, fast, and directionally correct, but without inter-rater reliability data. A diagnostic tool where two PFD practitioners independently audit the same site and converge on the same layer diagnosis would strengthen the methodology from “expert intuition with a framework” to “teachable, repeatable diagnostic.” This is the gap between a clinical art and a clinical science.
Temporal dynamics. PFD describes the first encounter well: activation, first impression, mental waterfall. But user relationships with products evolve. At visit 100, the mental waterfall is different. Prediction errors that activated on first visit are now expected. Fluency that felt trustworthy can become invisible (users stop noticing). Does the 5-layer stack apply the same way to retention and long-term engagement as it does to acquisition? The current framework is weighted toward first contact.
Accessibility as justice vs. fluency. PFD frames accessibility through L0 (cognitive load reduction) and L2 (processing fluency). This produces good accessible design in practice. But disability justice scholars argue accessibility is a rights issue, not an optimization target. Framing it through “fluency” risks treating accessibility as something you optimize when the ROI justifies it, rather than something you guarantee because people deserve it. The framework’s utilitarian lens may not fully serve this.
AI-generated design. As AI tools generate more design artifacts, PFD’s emphasis on embodied sensitivity (the “neurodivergent tuning fork,” the pre-verbal Feel step) becomes harder to operationalize at scale. Can PFD’s diagnostic be automated? Can an AI run the Feel step? If not, the methodology has a ceiling on how far it scales beyond the individual practitioner. If yes, what’s lost when the diagnostic instrument is no longer a human body?
Appendix A: The Science
~100 citations across cognitive psychology, HCI, behavioral economics, design theory, STS, music cognition, narrative psychology, and ADHD cognitive neuroscience.
| Citation | Year | Finding | Layer |
|---|---|---|---|
| Heider & Simmel | 1944 | People attribute intentions/emotions to geometric shapes | L1 |
| Miller, G.A. | 1956 | “Magical Number Seven”: rhetorical, not literal | L0 |
| McGurk & MacDonald | 1976 | Visual lip movements integrate with auditory perception | L2 |
| Nisbett & Wilson | 1977 | People can’t accurately report why they made decisions | L3 |
| Lakoff & Johnson | 1980 | Abstract thought structured by bodily metaphors | L2 |
| Sweller | 1988 | Cognitive Load Theory (intrinsic vs. extraneous) | L0 |
| Johnson et al. | 1991 | Neonatal face preference | L1 |
| Sloboda | 1991 | Musical features trigger involuntary physical responses | Comm. |
| Friestad & Wright | 1994 | Persuasion Knowledge Model: users object to insincerity | Ethics |
| Kurosu & Kashimura | 1995 | Aesthetic-usability effect | L1 |
| Salthouse | 1996 | Processing speed declines continuously from early adulthood | L0 |
| Tractinsky | 1997 | Aesthetic-usability cross-cultural replication | L1 |
| Sweller, van Merriënboer & Paas | 1998 | Formalized CLT taxonomy | L0 |
| Berdichevsky & Neuenschwander | 1999 | Golden Rule of persuasion | Ethics |
| Reber & Schwarz | 1999 | Processing fluency → perceived truth | L2 |
| Campbell & Kirmani | 2000 | Sincerity as function of perceived ulterior motives | Ethics |
| Green & Brock | 2000 | Narrative transportation → belief change | Comm. |
| Tractinsky, Katz & Ikar | 2000 | “What is Beautiful is Usable” | L1 |
| Cowan | 2001 | Working memory revised to ~4 chunks | L0 |
| Winkielman & Cacioppo | 2001 | Fluency generates positive affect without awareness | L2, L3 |
| Friedman, Kahn & Borning | 2002 | Value Sensitive Design | Ethics |
| Johnson & Goldstein | 2003 | “Do defaults save lives?” | L4 |
| Norman | 2004 | Emotional Design (visceral, behavioral, reflective) | Bonus |
| Reber, Schwarz & Winkielman | 2004 | Processing fluency → aesthetic pleasure | L2 |
| Hershler & Hochstein | 2005 | Adult face “pop-out” in visual search | L1 |
| Huron | 2006 | ITPRA theory: prediction/reaction are pre-conscious | Comm. |
| Lindgaard et al. | 2006 | 50ms visual appeal judgments | L1 |
| Tractinsky, Cokhavi et al. | 2006 | Consistent aesthetic judgments at brief exposures | L1 |
| Gray, Gray & Wegner | 2007 | Two dimensions of mind perception: Agency + Experience | L1 |
| Juslin & Västfjäll | 2008 | Six mechanisms for music-evoked emotion | Comm. |
| Song & Schwarz | 2008 | Fluency and detection of misleading questions | L2 |
| Thaler & Sunstein | 2008 | Choice architecture / Nudge theory | L4 |
| Alter & Oppenheimer | 2009 | Fluency generalizes across truth, confidence, trust | L2 |
| Brignull | 2010 | Dark patterns taxonomy | Ethics |
| Cowan | 2010 | “The Magical Mystery Four”: WM refined to 3–5 chunks | L0 |
| Dechêne et al. | 2010 | Illusory truth effect meta-analysis (51 studies) | L2 |
| Friston | 2010 | Free-energy principle | Core |
| Henrich, Heine & Norenzayan | 2010 | WEIRD caveat for benchmark generalizability | Note |
| Scheibehenne, Greifeneder & Todd | 2010 | Choice overload effect near-zero in many conditions | L4 |
| Kahneman | 2011 | Thinking, Fast and Slow: System 1 / System 2 | Core |
| Knobel & Bowker | 2011 | “Values in Design”: values inscribed through design decisions (CACM) | Ethics |
| Owsley | 2011 | Age-related declines in visual processing | L0 |
| Spence | 2011 | Cross-modal correspondences | L2 |
| Tuch et al. | 2012 | Replication of 50ms visual appeal | L1 |
| Clark | 2013 | Predictive processing: brains match input to top-down predictions | Core |
| Hohwy | 2013 | The Predictive Mind | Core |
| Reinecke et al. | 2013 | Visual evaluation at 17ms (Google Research) | L1 |
| Koelsch | 2014 | Music modulates subcortical emotion structures (Nature Reviews Neuroscience) | Comm. |
| van Laer et al. | 2014 | Narrative transportation meta-analysis (132 effect sizes) | Comm. |
| Zak | 2015 | Narrative arcs trigger cortisol + oxytocin (mechanism undemonstrated) | Comm. |
| Braddock & Dillard | 2016 | Narrative persuasion meta-analysis | Comm. |
| Gray et al. | 2018 | Dark patterns taxonomy: nagging, obstruction, sneaking | Ethics |
| Sweller, van Merriënboer & Paas | 2019 | CLT reconceptualized | L0 |
| Mierop et al. | 2020 | Intranasal oxytocin: ~90% false positive rate | Note |
| Bujack et al. | 2022 | Perceptual color space is non-Riemannian (PNAS) | L2 |
| Maier et al. | 2022 | Nudge effects near-zero after publication bias correction | L4 |
| v3.6 International Expansion (April 2026) | |||
| Wertheimer, M. | 1923 | Gestalt foundations — perception as organized wholes | L1, L2 |
| Metzger, W. | 1936 / 2006 | Comprehensive Gestalt empirical treatment (English trans. 2006) | L1, L2 |
| Mori, M. / MacDorman & Kageki | 1970 / 2012 | Uncanny valley — near-human-but-not-quite triggers eeriness before conscious evaluation (trans. 2012) | L1 |
| Nagamachi, M. | 1995 | Kansei engineering — systematic methodology mapping consumer affective response to product attributes | Sibling |
| Masuda & Nisbett | 2001 | Cultural attention differences — East Asian observers attend to background/relational; Western to foreground/focal | L1, L3 |
| Ji, Nisbett & Su | 2001 | Cultural differences in prediction direction (linear vs non-linear reasoning about change) | L4 |
| Leder, Belke, Oeberst & Augustin | 2004 | 5-stage aesthetic appreciation model (perception → implicit integration → explicit classification → cognitive mastery → evaluation) | L1 |
| Chua, Boland & Nisbett | 2005 | Eye-tracking replication of cultural attention differences | L3 |
| Pessiglione et al. | 2007 | Subliminal incentive processing — reward cues below conscious threshold drive motor behavior | L1, L3 |
| Hassin, Bargh, Engell & McCulloch | 2009 | Implicit working memory — WM capacity spent on unconscious processing, not just conscious attention | L0 |
| Hertwig & Erev | 2009 | Description-experience gap — decisions from description underweight rare events | L4 |
| Goh, J.O.S. et al. | 2010 | Cultural differences in face-processing cortex activation (FFA) | L1 |
| Trope & Liberman | 2010 | Construal-level theory — psychological distance shapes concrete vs abstract linguistic processing | L4 |
| Forster, Leder & Ansorge | 2013 | Subjective felt fluency drives aesthetic liking; objective fluency alone insufficient | L2 |
| Hassin, R.R. | 2013 | “Yes It Can” — unconscious performs every high-level cognitive function previously assumed to require awareness | L3, Core |
| Seckler, Heinz, Forde, Tuch & Opwis | 2015 | Visual characteristics (color, typography, layout coherence) drive trust judgments directly | L3 |
| Skov & Nadal | 2020 | Neuroaesthetics as hedonic valuation — neural substrates of aesthetic experience | L1, Sibling |
| ADHD Cognitive Neuroscience | |||
| Barkley, R. | 1997 | Self-regulation, not attention, is the core ADHD deficit. Executive function failures occur at the point of performance. | L0, L4 |
| Castellanos, F.X. | 2002 | ADHD involves altered default mode network dynamics | L0 |
| Carson, Peterson & Higgins | 2003 | 7x rate of low latent inhibition in high-functioning creative achievers | L2 |
| Dodson, W. | 2005 | Interest-Based Nervous System: ADHD motivation responds to PINCH (Passion, Interest, Novelty, Competition, Hyperurgency) | L0, L3 |
| White, H. & Shah, P. | 2006 | ADHD associated with stronger divergent thinking on multiple measures | L2 |
| White, H. & Shah, P. | 2011 | ADHD linked to higher real-world creative achievement | L2 |
| Kounios & Beeman | 2009 | Pre-problem brain state predicts whether solution arrives via insight or analysis | L0 |
| Beaty et al. | 2014 | Three-network model of creative cognition (default + executive + salience) | L0 |
| Mahan, B. (Wall of Awful) | 2020 | Emotional barrier accumulated from repeated executive function failures | L4 |
| Taylor, J. | 2020 | Twice-exceptional individuals in professional settings | L0 |
| Eisenberg et al. | 2008 | DRD4/7R allele associated with better nutritional status in nomadic Ariaal pastoralists but worse in settled — opposite fitness outcomes by environment | L2/L0 |
| Koestler, A. | 1964 | Bisociation: the creative act of holding two independent, normally incompatible frames of reference simultaneously | L0 |
| Mednick, S.A. | 1962 | Flat associative hierarchies produce more remote associations; basis for the Remote Associates Test (RAT) | L2 |
| Dietrich, A. | 2004 | Four creativity types mapped to prefrontal function: deliberate/spontaneous x cognitive/emotional | L0 |
| Clark, A. & Chalmers, D. | 1998 | Extended Mind Thesis: cognitive processes extend beyond brain/body when external tools reliably couple with internal cognition | Core |
| Rothenberg, A. | 1971 | Janusian thinking: actively conceiving two or more opposite concepts simultaneously as a creative process | L0 |
| Kasparov, G. | 2017 | Centaur model: weak human + machine + better process outperforms strong computer alone (Deep Thinking, PublicAffairs) | L4 |
| Mark, G. et al. | 2008 | Cost of interrupted work: average 23 minutes 15 seconds recovery time after interruption (CHI 2008) | L0 |
| Steyvers, M. et al. | 2022 | Bayesian human-AI complementarity: emerges when human and AI error patterns are sufficiently independent, even when AI outperforms humans individually (PNAS) | L4 |
| Li, C. et al. | 2023 | EmotionPrompt: emotional stimuli in prompts improve LLM output; 10.9% average improvement on human evaluation, up to 115% on edge-case benchmarks (AAAI 2024) | L3 |
| Kasatskii, V. et al. (JetBrains) | 2023 | Perceptual load in IDE environments differentially affects developers with ADHD symptoms; low load shows speed benefits that differ from neurotypical developers | L2/L0 |
| Carik, B. et al. | 2024 | Neurodivergent users perceive LLM output as “overly neurotypical”; community-driven workarounds documented (ACM PACMHCI / CSCW2, peer-reviewed) | L2/L3 |
On methodology: This document was formalized with AI assistance (Claude, Anthropic). The underlying methodology, case studies, and practitioner insights are the author’s own work. AI was used for structuring, identifying relevant literature, and iterative refinement. Citations were separately audited for accuracy. The framework’s ideas predate the AI assistance; the formalization does not.
Appendix B: Glossary
| Perception barrier | An unconscious objection that prevents conversion. Not a product, price, or demand problem. |
| Fence-sitter | Someone with intent who doesn’t convert due to perception barriers. |
| Mental waterfall | The conducted sequence of thoughts and emotions from first impression through conversion. |
| Activation point | The moment a user switches from autopilot to active engagement. |
| Listen Multiplier | (Listen) × (Act + React). If the user isn’t listening, everything multiplies by zero. |
| Cognitive load | Total mental effort. Intrinsic (task) + extraneous (design). |
| Processing fluency | Subjective ease of processing. Higher fluency = more positive evaluations. |
| Perception bias | Systematic gap between what users say, what stakeholders believe, and what actually happens. |
| Decision architecture | The structure of choices as presented. Every structure is a nudge. |
| Empathy modeling | Embodying the traits and context of a target demographic for directionality. |
| Hardware layer | Shared cognitive architecture: attention, memory, autopilot, mind perception. Universal. |
| Software layer | Cultural conventions: style preferences, expectations, literacy. Varies by audience. |
| Mind perception | Users construct lightweight models of competence (Agency) and care (Experience) for websites. |
| Predictive processing | Brain generates predictions about input. Matched = autopilot. Violated = attention fires. |
| Mental prosthesis | Scaffolding for cognitive functions and limitations. Good design IS prosthesis for the brain. The curb cut effect. |
| Cognitive Constraint Design | A proposed meta-framework: the principle that cognitive constraints are design material, not obstacles. PFD is the first formal application to interface design. |
| Curb-Cut Effect | When designs optimized for constrained users improve experience for all users. Named after sidewalk curb cuts designed for wheelchair users that benefit everyone. |
| Dual Foundation | PFD’s evidence base draws from two largely distinct citation pools — perception psychology and ADHD cognitive neuroscience — with zero citation overlap. |
| IBNS (Interest-Based Nervous System) | Dodson’s clinical model describing ADHD motivation as driven by interest, novelty, competition, and urgency rather than importance or reward. |
| Prediction Error | The mismatch between what the brain expects and what it encounters. The fundamental mechanism underlying first impressions and trust judgments. |
| Convergent Discovery | When multiple independent practitioners arrive at the same insight without knowledge of each other. Ten or more ADHD-AI builders discovered the same collaboration patterns independently, confirming environmental fitness. |
| Point of Performance | Barkley’s (1997) principle: self-regulation failures in ADHD occur in the moment of action, not during planning. Design must provide support where the decision happens, not where the instruction was given. |
Use It
Now you have the framework. What comes next depends on what you need.
Learn it. The Make Me Think series walks through each layer with case studies, citations, and the stories behind the principles. This reference page is yours to come back to. Cite it, teach it, build on it.
Use it in your tools. The framework and AI skill are open source on GitHub. Free to use with attribution.
Apply it. If something isn’t converting and you want the methodology applied to a specific problem, that’s what I do. E-commerce, SaaS, consumer products. 15 years of it. Services has the details.
Ask me. Genuinely. If you read this far, I’ll read your email.