Enhancing iMessage for Reading Accessibility

This project explores how personalized features could be added to Apple iMessage to reduce reading friction for users experiencing permanent, temporary, or situational reading difficulties.
This concept was developed in late 2024, prior to Apple’s AI summary feature in iOS 18.

Type

UI / UX

Team

Independent

Role

Product Designer,
Researcher

UI/UX Designer, Fullstack Delevoper

Methods / Tools

Research, User interviews and observations, Persona Spectrum, Journey mapping, Paper prototyping, Usability testing, UI style guides, High-fidelity prototypes in ProtoPie

Timline

2–3 months
Fall 2024

Context

Reading difficulties aren’t always permanent. People may struggle due to dyslexia, illness, fatigue, multitasking, or language barriers.

This project focuses on identifying these challenges and exploring new iMessage features that can support different reading contexts—without changing the app’s familiar UI or interaction patterns.
This concept came together in late 2024, before Apple's AI summaries released in iOS 18(October, 2024).

Challenge

The main challenges were to:

  • Understand reading difficulties across permanent, temporary, and situational constraints

  • Identify pain points in current iMessage experiences using a persona spectrum

  • Reduce overwhelm from long or fast-paced conversations

  • Ensure new features feel optional and supportive, not disruptive

  • Design a solution that works across all types of user contexts

Methods

I combined qualitative research with structured frameworks to understand user needs from multiple angles:

  • Used a KWHL table to map existing knowledge, assumptions, and research gaps

  • Applied research triangulation to validate insights across interviews, observations, and secondary research

  • Identified behavioral patterns using Spradley’s matrix, which informed persona and journey creation

  • Translated insights into design decisions through personas and user journey maps

  • Tested early ideas using paper prototypes, then refined them into mid–high fidelity prototypes in Figma, followed by interactive prototypes in ProtoPie

Project Outcome

The final outcome is a conceptual iMessage prototype that adds new, optional features to support reading accessibility across different user contexts.

What was added:

  • A conversation summary feature to help users quickly understand long or fast-paced message threads

  • An emoji-supported text feature that introduces visual cues to aid interpretation and tone recognition

The solution is delivered as a high fidelity interactive prototype(ProtoPie), focusing on feature behaviour and interaction flow rather than visual redesign.

Process

Process

01

Research & Discovery

Steps

01

Define the Goal

Started with a How Might We (HMW) statement. The goal was to explore feature-level opportunities that reduce reading friction while keeping iMessage familiar and easy to use.

How Might We

Enhance The Texting Experience
For Users With Reading Difficulties?

02

Research Planning – KWHL

I used a KWHL table to structure my research early on.

  • Clarified what I already knew and what I needed to learn

  • Documented assumptions and new insights in an organized way

  • Identified gaps in the problem space and opportunities for redesign

03

Inclusive Design –
Understand Users Deeply

I approached reading difficulty as a human–technology mismatch rather than a personal condition, which helped frame three types of reading-related constraints.

  • Instead of creating traditional persona, I created Personas Spectrum represent three types of constraints: permanent (dyslexia), temporary (illness or fatigue), and situational (language barrier/multitasking).

  • Used journey maps and Spradley’s matrix to understand real texting behaviours and breakdown point

05

Research Methods:

Observed international students and non-native English speakers texting in English

  • Found that slang, idioms, and tone vary across cultures

  • Noted emojis as a universal visual language that helps users blend into conversations

Interviewed 10 iOS 17 iMessage users using open-ended questions:

  • Users preferred short messages when tired or sick

  • Many relied on emojis or voice messages instead of long text

  • Customization and assistive tools were valued, but not always trusted


Scholarly research helped validate and contextualize findings from primary research.

  • Users process messages faster when emotion-expressing emoticons are present

  • Dyslexic users show better comprehension with visual supports like infoicons

  • Emojis are widely understood across cultures and contexts

Research Insights:

Using Research Triangulation, I focused on overlapping patterns across all research methods.

01

Summarization Reduces Cognitive Load

Short summaries help users quickly understand long or fast-paced conversations and reduce reading overwhelm.

02

Emoji Context Aids Interpretation

Contextual emojis support tone and meaning, especially for users facing language barriers, fatigue, or dyslexia.

02

Design & Refinement

Steps

01

Journey Mapping

Mapped common texting scenarios to understand user goals, pain points, emotions, and moments of friction. This helped identify where summary and emoji support could be most helpful.

02

Ideation & Prototyping

Brainstormed multiple ways to integrate new features into the journey map. Rapid paper prototyping was used to explore feature placement, interaction flow, and early concepts.

03

Testing & Refinement

Conducted usability testing with paper prototypes to evaluate how users understood and interacted with the summary and emojify features.

  • Participants completed task-based scenarios using a think-aloud approach

  • Observed user behaviour, confusion points, and decision-making

  • Gathered feedback on usability, layout, and feature clarity

  • Design refinements were made based on these insights.

04

High-fidelity Design

Translated the refined concepts into high-fidelity wireframes, applying the iOS 17 design system to ensure visual consistency, interaction accuracy, and system alignment.

03

Development & Handoff

  • Built an interactive prototype in ProtoPie to simulate real interaction behaviours and system responses

  • Created clear development and interaction specifications to support handoff to a development team in a real-world scenario

(Final Solution)

|

Take Away

  • Accessibility isn’t a niche problem — reading difficulties can affect anyone, depending on context.

  • Research helped challenge my early assumptions and guided more grounded design decisions.

  • Optional features scale better than forced solutions and respect different user needs.

  • Small feature additions can create meaningful impact without changing a product’s core identity.

Reflection

This project pushed me to think beyond permanent disabilities and design for real, everyday constraints. Working within Apple’s design system taught me how to balance creativity with structure. Most importantly, it reinforced how thoughtful, research-led decisions can make familiar products feel more inclusive.

Future Improvements

  • Test the concepts with a broader and more diverse user group.

  • Refine how emoji support works and explore different levels of personalization.

  • Compare these ideas alongside existing AI-based solutions to understand overlap and gaps.

  • Measure long-term impact through quantitative testing, such as comprehension or task speed.