Safe
Detecting
Detected



A smartwatch app that detects AI voice-cloning scams in real time and helps families stay connected and protected through shared alerts.
Year
2025
My Role
UX & UI Designer
Lead UX Researcher
Target Users
Older adults and their families
facing evolving AI-driven phishing
scams.
Team
Project Manager
UX Researchers
Designers
Overview
My Contributions
Research & Design Lead
I led cross generational research, designed the core interaction flows, and created wearable prototypes to study how users respond to AI driven scam scenarios. I also shaped key communication moments so the experience stays intuitive for people with different levels of tech literacy.
Problem
As AI-generated phishing scams become harder to detect, older adults and their families struggle to recognize threats, share information, and stay protected across varying levels of tech literacy.
Context
The rising threat of scams in the U.S.
Americans over the age of 60 lost $3.4 billion to scams in 2023, an 11% increase from 2022.
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
💸
FBI's 2023 Elder Fraud Annual Report
Nearly 1 billion emails were exposed in 2021, affecting 1 in 5 internet users.
Target Users
Older adults with lower tech literacy and the family members who help them stay safe from evolving phishing scams.


Target Users
01
Older Adult

Linda
Age: 72
Lives alone, communicates with family mostly by phone.
Teach Confidence
Low
Pain points:
Easily pressured by urgency
Struggles to tell real calls from AI generated voices
Unsure how to report scams
Goals: They need simple, immediate guidance that helps them recognize suspicious calls without relying on high tech knowledge.
Target Users
02
Adult Child

Abby
Age: 32
Works full time, wants to watch over aging parents from afar.
Teach Confidence
High
Pain points:
Cannot always pick up the phone
Worries parents cannot tell real calls from scams
Limited visibility into what actually happened
Goals: They need clear visibility into potential scam incidents so they can support and protect their parents even from a distance.
Desired Outcomes
Building Safety, Awareness, and Action
01
Prevent and Protect Users:
Safeguard individuals from AI voice cloning phishing scams by detecting and alerting them to potential threats in real time.
02
Educate Through Data Sharing:
Share detailed data with family members to raise awareness and educate them about the latest phishing scams and tactics.
03
Report Scams with Evidence:
Provide a seamless way to report scams to authorities in real time, including critical evidence for swift action.
User Emotion Map
How emotions shift during a scam event
I created this emotion map to understand how phishing incidents affect each person involved and to visualize how Orbit shapes their emotional experience. The map illustrates two emotional trajectories, one for the user receiving the AI call and the other for the connected family members when they receive an alert, along with how their emotions shift as they use Orbit.

Solution
Orbit: real time protection against AI voice scams

Orbit, powered by Ircam Amplify’s AI Speech Detector Model, combats AI voice cloning scams by detecting voice anomalies, flagging suspicious activity, and alerting trusted contacts. It ensures digital security with customizable settings and incident reviews.
Design Process
Research & User Tests Summary
Understanding how different generations recognize, navigate, and respond to evolving phishing scams
I led our research by conducting interviews, social polling, and user tests with both younger adults and older adults age 65 plus. I also analyzed secondary sources and spoke with key stakeholders to understand how different generations detect and respond to phishing scams. These insights shaped our problem framing and guided our design decisions.
Conduct 4 interviews with younger generations
Conduct 6 interviews with adults age 65+
Understand how both older and younger adults navigate technology to detect phishing scams, addressing a growing concern in cybersecurity.
Understand current state of phishing scams, prevention methods, phishing tactics and possible future evolution
Conduct 3 semi-structured interviews
Social media polling
Desk Research: news articles, blogs, existing resources for phishing scam protection
Does the user understand our design concept and does it address major pain points?
How do younger generations interact with it?
Iterated design based on feedback
How do older generations interact with it?
Digital security professionals
Tech-savvy family members
Older adults age 65+
Method
Primary Research
Secondary Research
User Tests - Young
User Tests - Senior
Why?
Key Stakeholders:
Research Insights
Key insights surfaced through user interviews and structured synthesis
I led the research process by writing the research plan, recruiting and interviewing participants, and guiding my teammates through affinity mapping sessions. I synthesized our findings into clear insights that revealed cross generational pain points and shaped our design direction.
“They'll pretend that they are someone that you've known for years” -P3
“My mom has been defrauded three times.” -P2
“He said my computer would crash if I didn’t fix it.” -P1
“We have taken over all of her banking, and make payments on her behalf” -P2
“She [Participant’s mom] got upset each time that ‘these nice people’ could lie so well” -P2
Building Trust Through Familiarity
Scammers leverage the principle of “familiarity” to establish a sense of trust, making their scams appear more credible and convincing to the victim.
Creating Panic Through Urgency
Scammers exploit “urgency” to create a sense of panic, pressuring victims to act quickly without questioning the legitimacy of the situation.
Exploiting Low Tech Literacy
Scammers take advantage of low technological literacy to confuse and mislead victims, making it harder for them to recognize and avoid scams.
Design Challenge
Turning insights into a design challenge
Based on our research findings, I proposed a design challenge and aligned it with the team to guide our design principles and concept brainstorming.
How might we leverage trust within families, across varied levels of tech literacy, to create a shared learning experience that helps everyone recognize and avoid AI voice cloning phishing scams?
Ideation
Brainstorming Sketches
Each team member brainstormed around 35 ideas through quick sketches. We then shared and discussed them to identify promising directions for concept development.
Design Principles
Principles that guided the design decisions
I developed these design principles by translating our research insights into clear priorities for the team. I worked closely with my teammates to align on what users truly needed, and these principles became the foundation that guided our decision making and kept the design direction focused and consistent.
Create an inclusive environment
Establish a welcoming space where all family members to feel valued and supported.
Fostering interactions within families
Encourage meaningful connections between family members by designing opportunities for shared experiences and conversations.
Education that evolves in real time
Implement adaptable learning resources that respond to changing phishing scams, ensuring relevant and continuous growth.
Design & Iterations
User Testing Round 1
Evaluating early watch interactions and users’ responses to light based alerts
For our first test, I created a watch prototype with three light indications: green for safe, yellow for detecting, and red for AI detected. I designed a test plan and tested these signals with older adults to understand how intuitive the colors felt, how they interpreted each state when paired with phone screens, and how comfortable they were with wearing a watch for scam detection. These observations guided our next design iteration and shaped how Orbit communicates urgency and safety.

Real-Time AI Detection
Family Sharing & Incident Report




Real-World Scam Awareness

User Testing Round 1 - Process and Insights
Testing real interactions through a live Wizard of Oz setup
We used a Wizard of Oz approach to test Orbit with residents in a senior living community, allowing us to simulate real time alerts and responses. I led the entire test process, from facilitation to in the moment adjustments, and worked with my teammates afterward to analyze the results. Together we identified key insights about how older adults interpret signals, react under pressure, and move through the experience in a realistic context.


Users want to understand how to report scams to the proper authorities
Multiple people expressed the desire to have a way to report a potential scam to an appropriate government official, but were unsure about the best way to do so.
A key element of feeling protected against scams includes open communication and sharing information with family and friends.
All of them expressed that they would want to connect with family and close friends after they or someone they know experienced a scam, in order to check in and make sure everything is okay.
Wearing a smart watch helps users detect scams even when they aren’t looking at their phone.
The smart watch format for our wearable device made sense to our users, who said they’d prefer that over a ring or a phone application on its own.
Design Challenge
Balancing user autonomy with family safety
Users wanted control over when to share a suspicious call, while our earlier design pushed automatic reporting for safety. The challenge was to create a flow that respects user choice without reducing family awareness when it is needed.
By shifting from automatic reporting to user-controlled sharing, the new flow respects personal boundaries while still supporting family safety, increasing overall trust and adoption likelihood.
Before
After
Red alert triggered
System auto shares with family
User reviews call summary
User chooses what to do next
Share with family
Save only for myself
Smart Share
Report to authority
Auto prompt to report
Users feel pressured
Additional case details coming soon...










