Guardly

A Scam-detection and education feature embedded within LinkedIn

INTRODUCTION

Guardly is a UX-led product developed during a 48-hour designathon, exploring how moment-based interventions can support job seekers at the point where scams are most likely to occur.

MY ROLE

I kept everyone aligned under tight time constraints by prioritising work and coordinating parallel workstreams. I contribute to the UX research and problem-definition phase, including framing the core challenge, sourcing supporting data, and shaping user interview questions.

I also owned the product narrative and final presentation to the judging panel, defining the story and communicating Guardly’s value and feasibility.

INTRODUCTION

Guardly is a UX-led product developed during a 48-hour designathon, exploring how moment-based interventions can support job seekers at the point where scams are most likely to occur.

[Project Scope]

Team of 5

[Role]

UX Designer - Product Manger

[Timeline]

48 hours designathon

View prototype

PROBLEM DISCOVERY

Job Scams look real and they catch people at their moment vunerable

Job scams no longer look suspicious. They look like real opportunities — complete with polished job listings, interviews, and official-looking offer letters. The danger often appears late in the process, during payroll onboarding, when job seekers are asked to share personal identification and bank details.

This makes scams especially effective during the job search, a time filled with urgency, hope, and emotional pressure. Many people are confident they won’t fall for scams until they find themselves caught in one.

This sparked a question -

“How might we empower users to confidently spot and report job scams so they can protect their personal information from online predators?”
PROBLEM DISCOVERY

Job Scams look real and they catch people at their moment vunerable

Job scams no longer look suspicious. They look like real opportunities — complete with polished job listings, interviews, and official-looking offer letters. The danger often appears late in the process, during payroll onboarding, when job seekers are asked to share personal identification and bank details.

This makes scams especially effective during the job search, a time filled with urgency, hope, and emotional pressure. Many people are confident they won’t fall for scams until they find themselves caught in one. This sparked a question -

“How might we empower users to confidently spot and report job scams so they can protect their personal information from online predators?”
Our Solution - Guardly

A moment-based scam detection and education experience for job seekers

Timely warning and clear guidance right when it matters most

Stays out of the way — Once connected with Linkedin, it runs quietly in the background while users browse jobs and messages.

Steps in when risk is highest — surfacing calm warnings only when suspicious patterns appear.

Helps users make sense of the risk — breaking down what feels off so users can decide what to do next with confidence.

Empowers users to take action — Report suspected scams and always closing the loop with clear follow-up, so reporting never feels pointless.

The Process

STARTING POINT

With 48 hours on the clock, we did a deep dive into the world of online scams — a space that turned out to be far more complex, emotional, and human than we expected.

While the brief broadly explored online safety, early research revealed many forms of digital harm. We chose to narrow our focus to job scams, where financial loss and emotional distress are often inflicted during the urgency, hope, and pressure of the job search.

DISCOVERING THE PROBLEM SPACE

In Australia alone, $13.5 million was lost to job scams in 2025, often during moments of urgency, hope, and emotional vulnerability.

Our research revealed that most existing scam prevention relies on policy, broad education, or platform-level signals such as labels and warnings. These interventions often live outside the moment of decision — appearing in classrooms, public campaigns, or after harm has already occurred — rather than when someone is about to click, reply, or share personal details.

We also found that heavy-handed regulation or platform bans can trigger resistance or feelings of censorship, especially when users don’t understand why something is being blocked. Even when warnings are present, they are often passive and non-actionable, leaving users unsure of what to do next.

In fast, emotionally charged online environments, urgency, optimism, and cognitive overload make it difficult for people to slow down and assess risk — precisely at the moments when caution matters most.

TALKING TO OUR USERS

Interviews with 25 young graduates revealed a pattern of overconfidence and vulnerability during scam encounters.

To dig deeper into these patterns, we conducted semi-structured interviews with 25 young graduates and early-career participants. We synthesised responses using an affinity diagram to cluster our findings.





[01] Overconfidence increases vulnerability

Many users, particularly younger job seekers, believed they were unlikely to fall for scams because they “grew up with the internet.” Despite this confidence, all interview participants are scam victims or knew someone who had been scammed.

[02] Intervention is often too late

Most existing solutions focus on education, policy, or content labeling, which typically occur before or after a risky interaction — not at the moment a user is deciding whether to engage. Cognitive overload and emotional decision-making mean critical thinking is often bypassed

[03] Job Scams are often the most legitimate looking

Unlike obvious scams, job-related scams often mirror real recruitment processes. Interviews, offer letters, and professional language build trust. The scam typically reveals itself during onboarding or payroll setup, when sensitive information is requested

[04] Scams are ignored, not reported

Reporting scams often felt pointless — once a report was submitted, users received no feedback or visibility into what happened next. Over time, this lack of transparency discouraged action, allowing scams to be ignored and repeated rather than actively addressed.

COMPETITOR ANALYSIS

As our research progressed, we looked at existing scam protection tools such as Scam Shield and Truecaller. These products demonstrated how effective real-time intervention can be when scam detection is built directly into communication channels.

This made me think: “Why doesn’t this exist for online platforms - where scams are increasingly common?”

Scam Shield - Government backed
True Caller - Community based

Where it falls short

  • Reactive by design, protection often happens after exposure rather than before action

  • Only available for phone calls and SMS

  • Minimal learning and reflection, flags scams without helping users understand why or build long-term awareness

What it does well

  • Government-backed, signalling strong credibility for a safety product.

  • Grounded in real-world scam behaviour: Ability to check links, numbers and messages

  • Most protection happens automatically in the background, reducing the cognitive load required from users.

APPROACHING A SOLUTION

Creating just enough friction to stop scams, while building long-term awareness

Under tight time constraints, we ran a Crazy 8s ideation session to rapidly explore solution directions. Using dot-voting, the team prioritised ideas that best aligned with our research insights and design principles. This allowed us to converge quickly on a focused, feasible solution.




Opportunities I identified

Timely intervention is key

We should introduce friction at the moment of decision, not reactively.

Education through context

We should go beyond flagging scams by helping users understand why something is risky, building long-term scam awareness and confidence.

Expand to social media and online platforms

We should extend scam prevention into social media and online platforms where scams increasingly occur, but protections are limited.

REFRAMING THE HMW

Based on our research and interview, we reframed the design challengre to focus on when the scams occur, not just what they look like

To ground this reframe, we created personas and mapped the user journey to better empathize with our users.

Full research available on our Figjam board

FINAL SOLUTION

Job scams happens to more people than you think. Right now, no seamless tool helps job seekers verify opportunities as they browse. That gap leaves young Australians exposed — and it’s what our solution tackles.This made me think: “Why doesn’t this exist for online platforms - where scams are increasingly common?”

Trust building & Platform connection

Intentional friction at point of risk

Learning through real-world practice

Key learnings

How to make confident trade-offs without overthinking

Working fast (and sleep-deprived), I learned to synthesis data quickly, communicate decisively and prioritize under pressure

Teamwork really does make the dream work. With only 48 hours and a team of five, this designathon became a crash course in collaboration under pressure — and a reminder of how much I enjoy working collaboratively. This project reinforced my interest in designing systems that support users during emotionally vulnerable moments, where thoughtful UX decisions can make a real difference.

What I'd do with more time
Test with real job seeker - across different age group and level of digital literacy to validate the moment-based intervention

Friction without discouraging users - refine warning language and tone without causing fear or disengagement
Expand to advertisement scams - Extend the solution to address scams across border online platforms
COMPETITOR ANALYSIS

As our research progressed, we looked at existing scam protection tools such as Scam Shield and Truecaller. These products demonstrated how effective real-time intervention can be when scam detection is built directly into communication channels.

This made me think: “Why doesn’t this exist for online platforms - where scams are increasingly common?”

Scam Shield - Government backed

What it does well

  • Government-backed, signalling strong credibility for a safety product.

  • Grounded in real-world scam behaviour: Ability to check links, numbers and messages

  • Most protection happens automatically in the background, reducing the cognitive load required from users.

True Caller - Community based

Where it falls short

  • Reactive by design, protection often happens after exposure rather than before action

  • Only available for phone calls and SMS

  • Minimal learning and reflection, flags scams without helping users understand why or build long-term awareness

TALKING TO OUR USERS

Interviews with 25 young graduates revealed a pattern of overconfidence and vulnerability during scam encounters.

To dig deeper into these patterns, we conducted semi-structured interviews with 25 young graduates and early-career participants. We synthesised responses using an affinity diagram to cluster our findings.





[01] Overconfidence increases vulnerability

Many users, particularly younger job seekers, believed they were unlikely to fall for scams because they “grew up with the internet.” Despite this confidence, all interview participants are scam victims or knew someone who had been scammed.

[02] Intervention is often too late

Most existing solutions focus on education, policy, or content labeling, which typically occur before or after a risky interaction — not at the moment a user is deciding whether to engage. Cognitive overload and emotional decision-making mean critical thinking is often bypassed

[03] Job Scams are often the most legitimate looking

Unlike obvious scams, job-related scams often mirror real recruitment processes. Interviews, offer letters, and professional language build trust. The scam typically reveals itself during onboarding or payroll setup, when sensitive information is requested

[04] Scams are ignored, not reported

Reporting scams often felt pointless — once a report was submitted, users received no feedback or visibility into what happened next. Over time, this lack of transparency discouraged action, allowing scams to be ignored and repeated rather than actively addressed.

Opportunities I identified

Timely intervention is key

We should introduce friction at the moment of decision, not reactively.

Education through context

We should go beyond flagging scams by helping users understand why something is risky, building long-term scam awareness and confidence.

Expand to social media and online platforms

We should extend scam prevention into social media and online platforms where scams increasingly occur, but protections are limited.

APPROACHING A SOLUTION

Creating just enough friction to stop scams, while building long-term awareness

Under tight time constraints, we ran a Crazy 8s ideation session to rapidly explore solution directions. Using dot-voting, the team prioritised ideas that best aligned with our research insights and design principles. This allowed us to converge quickly on a focused, feasible solution.




FINAL SOLUTION

Job scams happens to more people than you think. Right now, no seamless tool helps job seekers verify opportunities as they browse. That gap leaves young Australians exposed — and it’s what our solution tackles.This made me think: “Why doesn’t this exist for online platforms - where scams are increasingly common?”

Trust building & Platform connection

Intentional friction at point of risk

Learning through real-world practice

REFRAMING THE HMW

Based on our research and interview, we reframed the design challengre to focus on when the scams occur, not just what they look like

To ground this reframe, we created personas and mapped the user journey to better empathize with our users.

Full research available on our Figjam board

Opportunities I identified

Timely intervention is key

We should introduce friction at the moment of decision, not reactively.

Education through context

We should go beyond flagging scams by helping users understand why something is risky, building long-term scam awareness and confidence.

Expand to social media and online platforms

We should extend scam prevention into social media and online platforms where scams increasingly occur, but protections are limited.

REFRAMING THE HMW

Based on our research and interview, we reframed the design challengre to focus on when the scams occur, not just what they look like

To ground this reframe, we created personas and mapped the user journey to better empathize with our users.

Full research available on our Figjam board

APPROACHING A SOLUTION

Creating just enough friction to stop scams, while building long-term awareness

Under tight time constraints, we ran a Crazy 8s ideation session to rapidly explore solution directions. Using dot-voting, the team prioritised ideas that best aligned with our research insights and design principles. This allowed us to converge quickly on a focused, feasible solution.




FINAL SOLUTION

Job scams happens to more people than you think. Right now, no seamless tool helps job seekers verify opportunities as they browse. That gap leaves young Australians exposed — and it’s what our solution tackles.This made me think: “Why doesn’t this exist for online platforms - where scams are increasingly common?”

Trust building & Platform connection

Intentional friction at point of risk

Learning through real-world practice

Key learnings

How to make confident trade-offs without overthinking

Working fast (and sleep-deprived), I learned to synthesis data quickly, communicate decisively and prioritize under pressure

Teamwork really does make the dream work. With only 48 hours and a team of five, this designathon became a crash course in collaboration under pressure — and a reminder of how much I enjoy working collaboratively. This project reinforced my interest in designing systems that support users during emotionally vulnerable moments, where thoughtful UX decisions can make a real difference.

What I'd do with more time
Test with real job seeker - across different age group and level of digital literacy to validate the moment-based intervention

Friction without discouraging users - refine warning language and tone without causing disengagement or panic
Expand to advertisement scams - Extend the solution to address scams across border online platforms
Guardly

A Scam-detection and education feature embedded within LinkedIn

INTRODUCTION

Guardly is a UX-led product developed during a 48-hour designathon, exploring how moment-based interventions can support job seekers at the point where scams are most likely to occur.

MY ROLE

I kept everyone aligned under tight time constraints by prioritising work and coordinating parallel workstreams. I contribute to the UX research and problem-definition phase, including framing the core challenge, sourcing supporting data, and shaping user interview questions.

I also owned the product narrative and final presentation to the judging panel, defining the story and communicating Guardly’s value and feasibility.

INTRODUCTION

Guardly is a UX-led product developed during a 48-hour designathon, exploring how moment-based interventions can support job seekers at the point where scams are most likely to occur.

[Project Scope]

Team of 5

[Role]

UX Designer - Product Manger

[Timeline]

48 hours designathon

View prototype

PROBLEM DISCOVERY

Job Scams look real and they catch people at their moment vunerable

Job scams no longer look suspicious. They look like real opportunities — complete with polished job listings, interviews, and official-looking offer letters. The danger often appears late in the process, during payroll onboarding, when job seekers are asked to share personal identification and bank details.

This makes scams especially effective during the job search, a time filled with urgency, hope, and emotional pressure. Many people are confident they won’t fall for scams until they find themselves caught in one.

This sparked a question -

“How might we empower users to confidently spot and report job scams so they can protect their personal information from online predators?”
PROBLEM DISCOVERY

Job Scams look real and they catch people at their moment vunerable

Job scams no longer look suspicious. They look like real opportunities — complete with polished job listings, interviews, and official-looking offer letters. The danger often appears late in the process, during payroll onboarding, when job seekers are asked to share personal identification and bank details.

This makes scams especially effective during the job search, a time filled with urgency, hope, and emotional pressure. Many people are confident they won’t fall for scams until they find themselves caught in one. This sparked a question -

“How might we empower users to confidently spot and report job scams so they can protect their personal information from online predators?”
Our Solution - Guardly

A moment-based scam detection and education experience for job seekers

Timely warning and clear guidance right when it matters most 

Stays out of the way — Once connected with Linkedin, it runs quietly in the background while users browse jobs and messages.

Steps in when risk is highest — surfacing calm warnings only when suspicious patterns appear.

Helps users make sense of the risk — breaking down what feels off so users can decide what to do next with confidence.

Empowers users to take action — Report suspected scams and always closing the loop with clear follow-up, so reporting never feels pointless.

The Process

STARTING POINT

With 48 hours on the clock, we did a deep dive into the world of online scams — a space that turned out to be far more complex, emotional, and human than we expected.

While the brief broadly explored online safety, early research revealed many forms of digital harm. We chose to narrow our focus to job scams, where financial loss and emotional distress are often inflicted during the urgency, hope, and pressure of the job search.

DISCOVERING THE PROBLEM SPACE

In Australia alone, $13.5 million was lost to job scams in 2025, often during moments of urgency, hope, and emotional vulnerability.

Our research revealed that most existing scam prevention relies on policy, broad education, or platform-level signals such as labels and warnings. These interventions often live outside the moment of decision — appearing in classrooms, public campaigns, or after harm has already occurred — rather than when someone is about to click, reply, or share personal details.

We also found that heavy-handed regulation or platform bans can trigger resistance or feelings of censorship, especially when users don’t understand why something is being blocked. Even when warnings are present, they are often passive and non-actionable, leaving users unsure of what to do next.

In fast, emotionally charged online environments, urgency, optimism, and cognitive overload make it difficult for people to slow down and assess risk — precisely at the moments when caution matters most.

TALKING TO OUR USERS

Interviews with 25 young graduates revealed a pattern of overconfidence and vulnerability during scam encounters.

To dig deeper into these patterns, we conducted semi-structured interviews with 25 young graduates and early-career participants. We synthesised responses using an affinity diagram to cluster our findings.





[01] Overconfidence increases vulnerability

Many users, particularly younger job seekers, believed they were unlikely to fall for scams because they “grew up with the internet.” Despite this confidence, all interview participants are scam victims or knew someone who had been scammed.

[02] Intervention is often too late

Most existing solutions focus on education, policy, or content labeling, which typically occur before or after a risky interaction — not at the moment a user is deciding whether to engage. Cognitive overload and emotional decision-making mean critical thinking is often bypassed

[03] Job Scams are often the most legitimate looking

Unlike obvious scams, job-related scams often mirror real recruitment processes. Interviews, offer letters, and professional language build trust. The scam typically reveals itself during onboarding or payroll setup, when sensitive information is requested

[04] Scams are ignored, not reported

Reporting scams often felt pointless — once a report was submitted, users received no feedback or visibility into what happened next. Over time, this lack of transparency discouraged action, allowing scams to be ignored and repeated rather than actively addressed.

COMPETITOR ANALYSIS

As our research progressed, we looked at existing scam protection tools such as Scam Shield and Truecaller. These products demonstrated how effective real-time intervention can be when scam detection is built directly into communication channels.

This made me think: “Why doesn’t this exist for online platforms - where scams are increasingly common?”

Scam Shield - Government backed
True Caller - Community based

Where it falls short

  • Reactive by design, protection often happens after exposure rather than before action

  • Only available for phone calls and SMS

  • Minimal learning and reflection, flags scams without helping users understand why or build long-term awareness

What it does well

  • Government-backed, signalling strong credibility for a safety product.

  • Grounded in real-world scam behaviour: Ability to check links, numbers and messages

  • Most protection happens automatically in the background, reducing the cognitive load required from users.

APPROACHING A SOLUTION

Creating just enough friction to stop scams, while building long-term awareness

Under tight time constraints, we ran a Crazy 8s ideation session to rapidly explore solution directions. Using dot-voting, the team prioritised ideas that best aligned with our research insights and design principles. This allowed us to converge quickly on a focused, feasible solution.




Opportunities I identified

Timely intervention is key

We should introduce friction at the moment of decision, not reactively.

Education through context

We should go beyond flagging scams by helping users understand why something is risky, building long-term scam awareness and confidence.

Expand to social media and online platforms

We should extend scam prevention into social media and online platforms where scams increasingly occur, but protections are limited.

REFRAMING THE HMW

Based on our research and interview, we reframed the design challengre to focus on when the scams occur, not just what they look like

To ground this reframe, we created personas and mapped the user journey to better empathize with our users.

Full research available on our Figjam board

FINAL SOLUTION

Job scams happens to more people than you think. Right now, no seamless tool helps job seekers verify opportunities as they browse. That gap leaves young Australians exposed — and it’s what our solution tackles.This made me think: “Why doesn’t this exist for online platforms - where scams are increasingly common?”

Trust building & Platform connection

Intentional friction at point of risk

Learning through real-world practice

Key learnings

How to make confident trade-offs without overthinking

Working fast (and sleep-deprived), I learned to synthesis data quickly, communicate decisively and prioritize under pressure

Teamwork really does make the dream work. With only 48 hours and a team of five, this designathon became a crash course in collaboration under pressure — and a reminder of how much I enjoy working collaboratively. This project reinforced my interest in designing systems that support users during emotionally vulnerable moments, where thoughtful UX decisions can make a real difference.

What I'd do with more time
Test with real job seeker - across different age group and level of digital literacy to validate the moment-based intervention

Friction without discouraging users - refine warning language and tone without causing fear or disengagement
Expand to advertisement scams - Extend the solution to address scams across border online platforms
COMPETITOR ANALYSIS

As our research progressed, we looked at existing scam protection tools such as Scam Shield and Truecaller. These products demonstrated how effective real-time intervention can be when scam detection is built directly into communication channels.

This made me think: “Why doesn’t this exist for online platforms - where scams are increasingly common?”

Scam Shield - Government backed

What it does well

  • Government-backed, signalling strong credibility for a safety product.

  • Grounded in real-world scam behaviour: Ability to check links, numbers and messages

  • Most protection happens automatically in the background, reducing the cognitive load required from users.

True Caller - Community based

Where it falls short

  • Reactive by design, protection often happens after exposure rather than before action

  • Only available for phone calls and SMS

  • Minimal learning and reflection, flags scams without helping users understand why or build long-term awareness

TALKING TO OUR USERS

Interviews with 25 young graduates revealed a pattern of overconfidence and vulnerability during scam encounters.

To dig deeper into these patterns, we conducted semi-structured interviews with 25 young graduates and early-career participants. We synthesised responses using an affinity diagram to cluster our findings.





[01] Overconfidence increases vulnerability

Many users, particularly younger job seekers, believed they were unlikely to fall for scams because they “grew up with the internet.” Despite this confidence, all interview participants are scam victims or knew someone who had been scammed.

[02] Intervention is often too late

Most existing solutions focus on education, policy, or content labeling, which typically occur before or after a risky interaction — not at the moment a user is deciding whether to engage. Cognitive overload and emotional decision-making mean critical thinking is often bypassed

[03] Job Scams are often the most legitimate looking

Unlike obvious scams, job-related scams often mirror real recruitment processes. Interviews, offer letters, and professional language build trust. The scam typically reveals itself during onboarding or payroll setup, when sensitive information is requested

[04] Scams are ignored, not reported

Reporting scams often felt pointless — once a report was submitted, users received no feedback or visibility into what happened next. Over time, this lack of transparency discouraged action, allowing scams to be ignored and repeated rather than actively addressed.

Opportunities I identified

Timely intervention is key

We should introduce friction at the moment of decision, not reactively.

Education through context

We should go beyond flagging scams by helping users understand why something is risky, building long-term scam awareness and confidence.

Expand to social media and online platforms

We should extend scam prevention into social media and online platforms where scams increasingly occur, but protections are limited.

APPROACHING A SOLUTION

Creating just enough friction to stop scams, while building long-term awareness

Under tight time constraints, we ran a Crazy 8s ideation session to rapidly explore solution directions. Using dot-voting, the team prioritised ideas that best aligned with our research insights and design principles. This allowed us to converge quickly on a focused, feasible solution.




FINAL SOLUTION

Job scams happens to more people than you think. Right now, no seamless tool helps job seekers verify opportunities as they browse. That gap leaves young Australians exposed — and it’s what our solution tackles.This made me think: “Why doesn’t this exist for online platforms - where scams are increasingly common?”

Trust building & Platform connection

Intentional friction at point of risk

Learning through real-world practice

REFRAMING THE HMW

Based on our research and interview, we reframed the design challengre to focus on when the scams occur, not just what they look like

To ground this reframe, we created personas and mapped the user journey to better empathize with our users.

Full research available on our Figjam board

Opportunities I identified

Timely intervention is key

We should introduce friction at the moment of decision, not reactively.

Education through context

We should go beyond flagging scams by helping users understand why something is risky, building long-term scam awareness and confidence.

Expand to social media and online platforms

We should extend scam prevention into social media and online platforms where scams increasingly occur, but protections are limited.

REFRAMING THE HMW

Based on our research and interview, we reframed the design challengre to focus on when the scams occur, not just what they look like

To ground this reframe, we created personas and mapped the user journey to better empathize with our users.

Full research available on our Figjam board

APPROACHING A SOLUTION

Creating just enough friction to stop scams, while building long-term awareness

Under tight time constraints, we ran a Crazy 8s ideation session to rapidly explore solution directions. Using dot-voting, the team prioritised ideas that best aligned with our research insights and design principles. This allowed us to converge quickly on a focused, feasible solution.




FINAL SOLUTION

Job scams happens to more people than you think. Right now, no seamless tool helps job seekers verify opportunities as they browse. That gap leaves young Australians exposed — and it’s what our solution tackles.This made me think: “Why doesn’t this exist for online platforms - where scams are increasingly common?”

Trust building & Platform connection

Intentional friction at point of risk

Learning through real-world practice

Key learnings

How to make confident trade-offs without overthinking

Working fast (and sleep-deprived), I learned to synthesis data quickly, communicate decisively and prioritize under pressure

Teamwork really does make the dream work. With only 48 hours and a team of five, this designathon became a crash course in collaboration under pressure — and a reminder of how much I enjoy working collaboratively. This project reinforced my interest in designing systems that support users during emotionally vulnerable moments, where thoughtful UX decisions can make a real difference.

What I'd do with more time
Test with real job seeker - across different age group and level of digital literacy to validate the moment-based intervention

Friction without discouraging users - refine warning language and tone without causing disengagement or panic
Expand to advertisement scams - Extend the solution to address scams across border online platforms
Guardly

A Scam-detection and education feature embedded within LinkedIn

INTRODUCTION

Guardly is a UX-led product developed during a 48-hour designathon, exploring how moment-based interventions can support job seekers at the point where scams are most likely to occur.

MY ROLE

I kept everyone aligned under tight time constraints by prioritising work and coordinating parallel workstreams. I contribute to the UX research and problem-definition phase, including framing the core challenge, sourcing supporting data, and shaping user interview questions.

I also owned the product narrative and final presentation to the judging panel, defining the story and communicating Guardly’s value and feasibility.

INTRODUCTION

Guardly is a UX-led product developed during a 48-hour designathon, exploring how moment-based interventions can support job seekers at the point where scams are most likely to occur.

[Project Scope]

Team of 5

[Role]

UX Designer - Product Manger

[Timeline]

48 hours designathon

View prototype

PROBLEM DISCOVERY

Job Scams look real and they catch people at their moment vunerable

Job scams no longer look suspicious. They look like real opportunities — complete with polished job listings, interviews, and official-looking offer letters. The danger often appears late in the process, during payroll onboarding, when job seekers are asked to share personal identification and bank details.

This makes scams especially effective during the job search, a time filled with urgency, hope, and emotional pressure. Many people are confident they won’t fall for scams until they find themselves caught in one.

This sparked a question -

“How might we empower users to confidently spot and report job scams so they can protect their personal information from online predators?”
PROBLEM DISCOVERY

Job Scams look real and they catch people at their moment vunerable

Job scams no longer look suspicious. They look like real opportunities — complete with polished job listings, interviews, and official-looking offer letters. The danger often appears late in the process, during payroll onboarding, when job seekers are asked to share personal identification and bank details.

This makes scams especially effective during the job search, a time filled with urgency, hope, and emotional pressure. Many people are confident they won’t fall for scams until they find themselves caught in one. This sparked a question -

“How might we empower users to confidently spot and report job scams so they can protect their personal information from online predators?”
Our Solution - Guardly

A moment-based scam detection and education experience for job seekers

Timely warning and clear guidance right when it matters most

Stays out of the way — Once connected with Linkedin, it runs quietly in the background while users browse jobs and messages.

Steps in when risk is highest — surfacing calm warnings only when suspicious patterns appear.

Helps users make sense of the risk — breaking down what feels off so users can decide what to do next with confidence.

Empowers users to take action — Report suspected scams and always closing the loop with clear follow-up, so reporting never feels pointless.

The Process

STARTING POINT

With 48 hours on the clock, we did a deep dive into the world of online scams — a space that turned out to be far more complex, emotional, and human than we expected.

While the brief broadly explored online safety, early research revealed many forms of digital harm. We chose to narrow our focus to job scams, where financial loss and emotional distress are often inflicted during the urgency, hope, and pressure of the job search.

DISCOVERING THE PROBLEM SPACE

In Australia alone, $13.5 million was lost to job scams in 2025, often during moments of urgency, hope, and emotional vulnerability.

Our research revealed that most existing scam prevention relies on policy, broad education, or platform-level signals such as labels and warnings. These interventions often live outside the moment of decision — appearing in classrooms, public campaigns, or after harm has already occurred — rather than when someone is about to click, reply, or share personal details.

We also found that heavy-handed regulation or platform bans can trigger resistance or feelings of censorship, especially when users don’t understand why something is being blocked. Even when warnings are present, they are often passive and non-actionable, leaving users unsure of what to do next.

In fast, emotionally charged online environments, urgency, optimism, and cognitive overload make it difficult for people to slow down and assess risk — precisely at the moments when caution matters most.

TALKING TO OUR USERS

Interviews with 25 young graduates revealed a pattern of overconfidence and vulnerability during scam encounters.

To dig deeper into these patterns, we conducted semi-structured interviews with 25 young graduates and early-career participants. We synthesised responses using an affinity diagram to cluster our findings.





[01] Overconfidence increases vulnerability

Many users, particularly younger job seekers, believed they were unlikely to fall for scams because they “grew up with the internet.” Despite this confidence, all interview participants are scam victims or knew someone who had been scammed.

[02] Intervention is often too late

Most existing solutions focus on education, policy, or content labeling, which typically occur before or after a risky interaction — not at the moment a user is deciding whether to engage. Cognitive overload and emotional decision-making mean critical thinking is often bypassed

[03] Job Scams are often the most legitimate looking

Unlike obvious scams, job-related scams often mirror real recruitment processes. Interviews, offer letters, and professional language build trust. The scam typically reveals itself during onboarding or payroll setup, when sensitive information is requested

[04] Scams are ignored, not reported

Reporting scams often felt pointless — once a report was submitted, users received no feedback or visibility into what happened next. Over time, this lack of transparency discouraged action, allowing scams to be ignored and repeated rather than actively addressed.

COMPETITOR ANALYSIS

As our research progressed, we looked at existing scam protection tools such as Scam Shield and Truecaller. These products demonstrated how effective real-time intervention can be when scam detection is built directly into communication channels.

This made me think: “Why doesn’t this exist for online platforms - where scams are increasingly common?”

Scam Shield - Government backed
True Caller - Community based

Where it falls short

  • Reactive by design, protection often happens after exposure rather than before action

  • Only available for phone calls and SMS

  • Minimal learning and reflection, flags scams without helping users understand why or build long-term awareness

What it does well

  • Government-backed, signalling strong credibility for a safety product.

  • Grounded in real-world scam behaviour: Ability to check links, numbers and messages

  • Most protection happens automatically in the background, reducing the cognitive load required from users.

APPROACHING A SOLUTION

Creating just enough friction to stop scams, while building long-term awareness

Under tight time constraints, we ran a Crazy 8s ideation session to rapidly explore solution directions. Using dot-voting, the team prioritised ideas that best aligned with our research insights and design principles. This allowed us to converge quickly on a focused, feasible solution.




Opportunities I identified

Timely intervention is key

We should introduce friction at the moment of decision, not reactively.

Education through context

We should go beyond flagging scams by helping users understand why something is risky, building long-term scam awareness and confidence.

Expand to social media and online platforms

We should extend scam prevention into social media and online platforms where scams increasingly occur, but protections are limited.

REFRAMING THE HMW

Based on our research and interview, we reframed the design challengre to focus on when the scams occur, not just what they look like

To ground this reframe, we created personas and mapped the user journey to better empathize with our users.

Full research available on our Figjam board

FINAL SOLUTION

Job scams happens to more people than you think. Right now, no seamless tool helps job seekers verify opportunities as they browse. That gap leaves young Australians exposed — and it’s what our solution tackles.This made me think: “Why doesn’t this exist for online platforms - where scams are increasingly common?”

Trust building & Platform connection

Intentional friction at point of risk

Learning through real-world practice

Key learnings

How to make confident trade-offs without overthinking

Working fast (and sleep-deprived), I learned to synthesis data quickly, communicate decisively and prioritize under pressure

Teamwork really does make the dream work. With only 48 hours and a team of five, this designathon became a crash course in collaboration under pressure — and a reminder of how much I enjoy working collaboratively. This project reinforced my interest in designing systems that support users during emotionally vulnerable moments, where thoughtful UX decisions can make a real difference.

What I'd do with more time
Test with real job seeker - across different age group and level of digital literacy to validate the moment-based intervention

Friction without discouraging users - refine warning language and tone without causing fear or disengagement
Expand to advertisement scams - Extend the solution to address scams across border online platforms
COMPETITOR ANALYSIS

As our research progressed, we looked at existing scam protection tools such as Scam Shield and Truecaller. These products demonstrated how effective real-time intervention can be when scam detection is built directly into communication channels.

This made me think: “Why doesn’t this exist for online platforms - where scams are increasingly common?”

Scam Shield - Government backed

What it does well

  • Government-backed, signalling strong credibility for a safety product.

  • Grounded in real-world scam behaviour: Ability to check links, numbers and messages

  • Most protection happens automatically in the background, reducing the cognitive load required from users.

True Caller - Community based

Where it falls short

  • Reactive by design, protection often happens after exposure rather than before action

  • Only available for phone calls and SMS

  • Minimal learning and reflection, flags scams without helping users understand why or build long-term awareness

TALKING TO OUR USERS

Interviews with 25 young graduates revealed a pattern of overconfidence and vulnerability during scam encounters.

To dig deeper into these patterns, we conducted semi-structured interviews with 25 young graduates and early-career participants. We synthesised responses using an affinity diagram to cluster our findings.





[01] Overconfidence increases vulnerability

Many users, particularly younger job seekers, believed they were unlikely to fall for scams because they “grew up with the internet.” Despite this confidence, all interview participants are scam victims or knew someone who had been scammed.

[02] Intervention is often too late

Most existing solutions focus on education, policy, or content labeling, which typically occur before or after a risky interaction — not at the moment a user is deciding whether to engage. Cognitive overload and emotional decision-making mean critical thinking is often bypassed

[03] Job Scams are often the most legitimate looking

Unlike obvious scams, job-related scams often mirror real recruitment processes. Interviews, offer letters, and professional language build trust. The scam typically reveals itself during onboarding or payroll setup, when sensitive information is requested

[04] Scams are ignored, not reported

Reporting scams often felt pointless — once a report was submitted, users received no feedback or visibility into what happened next. Over time, this lack of transparency discouraged action, allowing scams to be ignored and repeated rather than actively addressed.

Opportunities I identified

Timely intervention is key

We should introduce friction at the moment of decision, not reactively.

Education through context

We should go beyond flagging scams by helping users understand why something is risky, building long-term scam awareness and confidence.

Expand to social media and online platforms

We should extend scam prevention into social media and online platforms where scams increasingly occur, but protections are limited.

APPROACHING A SOLUTION

Creating just enough friction to stop scams, while building long-term awareness

Under tight time constraints, we ran a Crazy 8s ideation session to rapidly explore solution directions. Using dot-voting, the team prioritised ideas that best aligned with our research insights and design principles. This allowed us to converge quickly on a focused, feasible solution.




FINAL SOLUTION

Job scams happens to more people than you think. Right now, no seamless tool helps job seekers verify opportunities as they browse. That gap leaves young Australians exposed — and it’s what our solution tackles.This made me think: “Why doesn’t this exist for online platforms - where scams are increasingly common?”

Trust building & Platform connection

Intentional friction at point of risk

Learning through real-world practice

REFRAMING THE HMW

Based on our research and interview, we reframed the design challengre to focus on when the scams occur, not just what they look like

To ground this reframe, we created personas and mapped the user journey to better empathize with our users.

Full research available on our Figjam board

Opportunities I identified

Timely intervention is key

We should introduce friction at the moment of decision, not reactively.

Education through context

We should go beyond flagging scams by helping users understand why something is risky, building long-term scam awareness and confidence.

Expand to social media and online platforms

We should extend scam prevention into social media and online platforms where scams increasingly occur, but protections are limited.

REFRAMING THE HMW

Based on our research and interview, we reframed the design challengre to focus on when the scams occur, not just what they look like

To ground this reframe, we created personas and mapped the user journey to better empathize with our users.

Full research available on our Figjam board

APPROACHING A SOLUTION

Creating just enough friction to stop scams, while building long-term awareness

Under tight time constraints, we ran a Crazy 8s ideation session to rapidly explore solution directions. Using dot-voting, the team prioritised ideas that best aligned with our research insights and design principles. This allowed us to converge quickly on a focused, feasible solution.




FINAL SOLUTION

Job scams happens to more people than you think. Right now, no seamless tool helps job seekers verify opportunities as they browse. That gap leaves young Australians exposed — and it’s what our solution tackles.This made me think: “Why doesn’t this exist for online platforms - where scams are increasingly common?”

Trust building & Platform connection

Intentional friction at point of risk

Learning through real-world practice

Key learnings

How to make confident trade-offs without overthinking

Working fast (and sleep-deprived), I learned to synthesis data quickly, communicate decisively and prioritize under pressure

Teamwork really does make the dream work. With only 48 hours and a team of five, this designathon became a crash course in collaboration under pressure — and a reminder of how much I enjoy working collaboratively. This project reinforced my interest in designing systems that support users during emotionally vulnerable moments, where thoughtful UX decisions can make a real difference.

What I'd do with more time
Test with real job seeker - across different age group and level of digital literacy to validate the moment-based intervention

Friction without discouraging users - refine warning language and tone without causing disengagement or panic
Expand to advertisement scams - Extend the solution to address scams across border online platforms
Select this text to see the highlight effect