NexusGPT
Enhancing first-crack AI agent usability
with guided prompt design
Timeline
2 months (Dec '24 - Jan '25)
My role
Product designer
Team
1 PD, 1 PM, 1 DEV
Status
Under-development

Background
1/7
Problem
3/7
AI agents aren’t finished when they’re deployed.
Just like human employees, they grow through feedback.
As they interact with real users, they encounter edge cases, learn new tasks, and develop blind spots.
But improving an agent isn’t as simple as checking a box. It means updating its brain — the prompt.
Problem
3/7
Problem
3/7
In theory, Nexus supports this loop—but in practice, the system makes this incredibly difficult.
But don't just take my word for it. Here's Justine, a customer support manager, to share what her experience was really like.
Problem
3/7
Justine's experience
Agent is deployed
Week 1
Issues start cropping up
Week 2-3
Fixes attempted
Week 4
Real consequences
Week 5
The agent works well and receives positive feedback.
Confusing responses. Edge cases encountered.
Unnecessary escalations to human support.
Repetitive or robotic tone.
Justine edits the agent’s prompt.
She’s overwhelmed by a wall of text.
She makes a hesitant change.
The agent’s logic breaks the next day.
Justine loses trust and stops editing.
😄
🤔
😥
💀
Justine's feelings throughout
Justine wasn't alone. Countless managers faced this same uncertainty every day.
The gap
Justine's story is far from unique. It exposes the fundamental flaws in how Nexus handles prompt maintenance, making agent improvement unnecessarily complex and risky.
Problem
2/7
Problem
3/7
We studied how users maintained agents and where their understanding fell short.
Problem
3/7
Lack of prompt expertise
Most managers don’t understand prompt engineering, leading to guesswork and ineffective updates.
Feedback isn't actionable
Feedback from conversations and tasks isn’t mapped to the right parts of the prompt, forcing users to hunt for issues blindly.
Fear of breaking things
Even small edits can cause unexpected failures, making users hesitant to improve their agents.
Problem
3/7
Watching how managers maintained their agents made one thing obvious: improving agents in Nexus often feels like pure guesswork and frustration.
Problem
3/7
Problem
3/7
Unstructured, monolithic prompts
Low prompt literacy
Feedback disconnected from prompts
Risky edits
Problem
3/7
All this added friction increased the burden on the dev team and left users engaging with a feature that failed to solve the core problem
Problem
3/7
>90%
created agents were undeployable
An overwhelming majority of agents created through the form were essentially useless.
80%
of most prompts needed edits
The generated prompts needed heavy edits, showing the form wasn’t doing its job.
6
hours lost weekly
The dev team manually fixed agent prompts through client calls, which drained limited dev resources.
Insights
4/7
Problem
3/7
After talking to devs and watching how they manually improved prompts, it became clear: our flow asked the wrong things — or didn’t ask enough.
Problem
3/7
Focused on “what,” not “how”
Agent creation = hiring. But our flow missed the basics: job, personality, training, onboarding.
Training came after the fact
Knowledge wasn't part of setup. Agents launched unprepared and underinformed.
Key fields were optional or too basic
Only name and job title were required. Purpose and capabilities? Skipped or vague.
No preview, no feedback
Users couldn’t see the prompt or agent evolve. No chance to course-correct early.
No real personality building
Just one “writing style” input. No deeper traits, tone, or values — agents felt flat.
Lack of immersion = lack of ownership
It felt like “filling out a form,” not “creating an agent.”
Problem
3/7
And on top of all that, the UI didn’t help — it was cramped, flat, and lacked any real visual hierarchy. Nothing about it felt like you were building something meaningful.
Research
5/7
Problem
3/7
This wasn't a form problem. It was a framing problem.
To ground the idea, I studied how real teams onboard roles like the ones users gave their agents — support reps, consultants & sales associates.
FINDINGS
CONCLUSION
The original agent form only covered ~20% of a typical onboarding process.
It captured surface-level inputs — job title, responsibilities, writing style — but missed the deeper layers: tone, behaviour, tools, training, and context. Most agents weren’t usable on the first try.
We didn’t need a better form.
We needed a better frame — one that treats agent creation like onboarding a teammate.
Problem
3/7
Rebuilding the flow around the onboarding model
We modeled the flow after real-life hiring to ask better, more insightful questions — helping users define agents more clearly and end up with stronger, more usable first drafts.
Rebuilding the flow around the onboarding model
Problem
3/7

Designs
6/7
Problem
3/7
I turned previously optional identity fields into a required first step — helping users define who the agent is from the start, reducing ambiguity for both the user and the system.
This step mirrors how real teams start onboarding — by clarifying who the person is, what they do, and where they belong. It’s about grounding the agent in a role and context before anything else.
Problem
3/7

Info that was optional before is now front and center.
A live preview gives the agent a face — making the process feel more personal and real.
Establishing a solid identity nudges users to think of the agent as a clearly defined persona with intent — not just a text generator.
Lets users pick from leading LLMs depending on the task — whether it needs speed, reasoning, creativity, or long context.
Problem
3/7
Prompts were often vague about the agent’s role, so we introduced agent types to give structure — and dramatically improve prompt quality.
Just like a real role needs a job description, agents need purpose. Users also chose what type of work was to be done. This raised prompt quality immediately, because now Nexus knows what to expect.
Problem
3/7

Problem
3/7

Selection pushes users to be specific about intent
Choosing a type helps structure the prompt behind the scenes
Problem
3/7
To prevent hallucinations and silence, users can now add knowledge and fallback logic during setup.
That’s why I designed the Knowledge step — to define not just what the agent should know, but how to retrieve it and what to do when it doesn’t.
Problem
3/7
Add knowledge when it matters — skip when it doesn’t
Define how the agent handles gaps or uncertainty
Reduce hallucination by grounding what the agent knows
Problem
3/7
I reduced deployment friction by enabling users to equip agents with the right tools and skills during creation — so they’re functionally ready from creation
Previously, agents were created first, and only after they failed at certain tasks did users realize they needed to attach skills (like search, API access, Airtable, etc.). This made agents feel limited or broken by default — and it often wasn’t obvious what was missing until something didn’t work.
Problem
3/7

Reduce launch friction by ensuring the agent is functionally ready
Relevant skill suggestions remove the guesswork
Problem
3/7
To address agents lacking personality or brand alignment, I improved the personality step with better questions and a more comprehensive tone builder.
Users can upload brand copy → we analyze tone → generate a matching voice. Or just pick a preset. Either way, agents sound way more on-brand.
Problem
3/7
Problem
3/7
To emphasize intentionality and immersion, I designed a live preview that updates in real time as users build their agent
We made the flow full-screen to feel more immersive. Inputs on the left, live agent preview on the right — so users can watch their agent come to life in real time. The snappy transitions are pretty cool too.
Problem
3/7
Problem
3/7

Conclusion
7/7
Problem
3/7
Still in dev, but early feedback is strong
No hard data yet, but users were way more confident during walkthroughs. We’ll measure prompt quality pre/post-launch to see how much this actually helps.
Rebuilding the flow around the onboarding model
Problem
3/7
Final thought: users don’t need power. they need clarity
We thought we were simplifying prompting. But the real win came from designing an experience that felt natural — something people already knew how to do.
Rebuilding the flow around the onboarding model
Return home
Back to work