Designing Trustworthy AI for Helplines for Automated Data Entry

by | February 24, 2026 | Articles

Acknowledgements: The Aselo Team, Annalise Irby, Nick Hurlburt, Dee Luo

Key Points

  • AI summaries of helpline conversations are making counsellor workflows faster, reducing admin work and cutting repetitive documentation.
  • Counsellors and supervisors find the summaries useful, because they are accurate, structured, and consistent.
  • Classification experiments show that understanding counsellor workflows matters more than model accuracy alone.
  • Globally-relevant AI advances can be accomplished without compromising trustworthy data practices.

Every day, helpline counsellors handle emotionally heavy conversations, sometimes lasting hours. Each conversation must be documented: writing summaries, selecting issue categories, filling out form fields, and preparing reports for supervisors or case managers.

Counsellors’ documentation, which we call helpline data entry, is critical. It protects callers, supports data quality and supervision, and surfaces trends across regions and countries. It is also time-consuming and mentally demanding, leaving counsellors less time for actual counselling.

We built Aselo, our open source contact center for these helplines to add text channels to traditionally voice helplines and to improve productivity of the counselors. One of the main areas for productivity improvements is reducing the time counselors spent on helpline data entry.

With generative AI tools, we saw an opportunity to do more: could we design AI tools that make counsellors’ work easier, faster, and more consistent, while preserving trust, privacy, and counsellor agency?

We focused on two features that lay the foundation for a fully automated Counsellor Data Entry Assistant.

  1. Summarization, which supports counsellors with structured, accurate summaries of a conversation.
  2. Issue categorization, which explores how AI can help streamline or automate the selection of presenting issues.
    When we began building AI into Aselo, our guiding principle was to amplify counsellors’ work without compromising trust, data protection, or human judgment.

In this post, I share three key learnings from building AI tools for helplines, spanning data stewardship, human-in-the-loop summarization, and the limits of automated classification over large label sets. They reflect the technical challenges, the human-centered design decisions, and the operational principles that shaped our approach to responsible AI in helpline services.

Example of the summarization interface in Aselo
Example of the summarization interface in Aselo (right size of image)

Learning 1: Protecting Sensitive Data Builds Trust and Enables Innovation

Rather than limiting innovation, privacy and protection policies create a foundation for trustworthy collaboration and safe experimentation.

Working with sensitive helpline data means earning trust through clear, practical safeguards. One of Aselo’s founding principles was that helplines own their data. True to that principle, this AI work only included helplines who explicitly opted in. We were transparent about how their data would be used and who would have access. We adopted policies from the beginning that center data protection and privacy:

  • Personally Identifiable Information (PII) redaction, which removes identifying details before any experimentation occurs.
  • Zero data retention, meaning no conversational data is stored by models or model providers.
  • Aselo’s already secure data infrastructure underpins all work.

With these safeguards in place, we could prototype AI data entry features such as summarization and issue categorization on small datasets. This approach gave partners confidence to collaborate, knowing that no conversational data would be stored or repurposed by third party service providers. OpenAI supported zero data retention for their models, enabling experimentation without retaining any data.

High-performing entity recognition tools, combined with trust built from careful data stewardship, enabled experimentation that preserved privacy. This approach protected sensitive data but also established the foundation for trustworthy AI tools that augment counsellor workflows responsibly.

Learning 2: AI Helps Write Summaries Faster While Preserving Counsellors’ Agency

LLM-assisted summaries can speed up data entry and support human judgment along with consistent documentation.

Following Aselo’s co-creation approach, we designed AI with counsellors in the loop, ensuring that automation supports helplines’ data entry needs.

AI Summaries Prototype

Before development, we ran a prototyping pilot that tested four types of summaries. These ranged from short, concise versions to longer, structured formats similar to formats used in training documents by helplines. We evaluated them using BLEU and ROUGE scores, LLM evaluators, and direct human ratings for accuracy and usefulness.

Longer and more structured summaries were preferred by human raters and by LLM-as-judge evaluators, showing early alignment between the two. They provided the clearest understanding of what happened during the contact. I also revisited the literature on evaluation metrics for natural language generation tasks like summarization, and spent time, for example, on Ehud Reiter’s blog reading about the current state of NLG evaluation as well as LLM-based evaluation methods (and shortcomings). From there, I developed helpline-specific indicators for quality and usefulness.

How Co-Design Shaped the Feature

We listened to counsellors and supervisors to understand how summaries fit within their workflow. They told us that they needed

  • summaries that are transparent and editable,
  • support for long and trauma-heavy conversations,
  • and reduction in time spend reviewing and correcting summaries.

Based on this feedback, we created a one-click summarization tool that produces an editable three-paragraph summary directly within Aselo. This feedback informed design choices such as keeping counsellors in the loop, ensuring transparency, and embedding AI features securely.

Early Impact in the Beta Test

We evaluated the feature with a mixed methods approach. We tracked accuracy ratings, edits to summaries, user engagement, supervisor feedback, and wrap-up times (time spent entering data after a call ends). We adapted the four-tiered AI evaluation system developed by The Agency Fund and the Center for Global Development to track technical and human-centered metrics.

What happened? Here is what we observed:

  • Counsellors rated AI-supported summaries as highly accurate and more standardized across their peers.
  • Supervisors noticed that the format, structure, and quality of summaries became more consistent across their workforce of counsellors.
  • Counsellors edited some of the AI-generated summaries, which showed active engagement, but often used them largely unchanged.
  • And most importantly, we saw more chats with short wrap-up times.

When counsellors marked a summary as Not Accurate, it was usually because the conversation was very brief, and a three-paragraph summary didn’t make sense. In response to this feedback, our next version of the feature has taken the length of the interaction into account to generate appropriate summaries.

These findings are early indicators that the feature is meaningfully augmenting counsellors’ work, maintaining human judgment and accuracy while improving efficiency.

Chart: Summary Accuracy Ratings During Test Window

Around 60% of AI-generated summaries were rated by Counsellors during the Beta Test window, and of these, most were rated Very accurate (81%).

Summary Accuracy Ratings During Test Window

Call wrap-up times chart

There was a higher proportion of chats with wrap-up times under 30 seconds when Counsellors used the AI feature, and a fewer proportion of chat wrap-ups that took 2-60 minutes (a small, but statistically significant, effect).

Call wrap-up times chart

Learning 3: Classification Is a Counsellor Workflow Challenge, Not Just a Model Challenge

Task definition shapes model performance, and human workflow shapes task definition.

Our initial focus on using generative AI technology was helpline issue categorization. This seemed like a straightforward classification problem: pick the top three issues that came up in a conversation from a list of topics. We thought this would translate to significant productivity improvements, but this proved more challenging than expected.

In practice, counsellors handle hundreds of live calls each week, and most use a small set of familiar categories. Remembering definitions and navigating long lists while keeping up with emerging issue trends is the real challenge.

Example of the issue categorization interface in Aselo
Example of the issue categorization interface in Aselo (right side of image)

What we tested

We ran zero-shot and few-shot experiments with GPT-4o-mini using fully redacted transcripts. We compared the results to a baseline that always selected the three most common labels. The model outperformed the baseline for two of three helplines, and performance improved further when we grouped labels into broader themes.

On paper, this looked like progress. Even simple AI-assisted classification improved accuracy while reducing cognitive load. Yet the results did not fully align with what counsellors needed. The technical improvements revealed that prediction alone was not enough to make the process truly useful.

What counsellors taught us was reflected in data

I had assumed that the issue classification task was cognitively demanding. For example, one helpline uses 163 possible categories, and counsellors choose up to three issue categories for each contact. But counsellors shared something that surprised us. They do not find the 163-category list overwhelming; they mostly use a small set of familiar categories.

The real difficulty comes from remembering the definitions of similar or overlapping labels and navigating long lists required for reporting.

The data confirmed this pattern. Across helplines, roughly two-thirds of labels came from the ten most frequently used categories, and the remaining categories (up to 150) formed a long tail and are rarely used.

A promising direction

Helplines explained that they sometimes create new issue categories after noticing patterns in free-text summaries during data reviews. This highlighted a promising direction for AI: supporting the evolution of category sets by recognizing issues emerging from new summaries. Our focus shifted from predicting issue categories to understanding how these categories evolve, and considering how to design tools that support that evolution. The next phase for categorization efforts pauses AI experimentation to focus on co-designing categorization approaches with helplines, grounded in counsellors’ actual workflows. AI isn’t the most essential tool for this challenge, but it did help us discover what is essential: in this case, strengthening data collection practices.

Looking Ahead: Counsellor AI Assistant that Scales

Our next phase involves scaling the Counsellor AI Assistant across the entire data entry and case management workflow. This includes exploring automated entry of form fields while continuing to strengthen secure data pipelines and test AI models hosted fully within our infrastructure.

What we have found is that the most valuable AI features are those that assist counsellors in their workflow, promote agency about which data is recorded, and maintain transparency at every step. By embedding safeguards, incorporating human feedback, and continually evaluating outcomes, we can scale AI data entry responsibly while preserving counsellor agency and the quality of their interactions with children and families.

This direction aligns with Aselo’s broader strategy to strengthen the helpline ecosystem across four goals: improving counsellor experience, enhancing data quality, expanding insight into caller’s needs, and increasing global reach. To us, responsible AI in helplines is not about replacing people. It is about amplifying their impact, ensuring that every call is well-documented, every counsellor is supported, and that the system is built to be effective for people to spend time supporting people in crisis.

We would like to especially thank Safe Online and The Agency Fund for their financial support of this project, and would like to thank OpenAI for agreeing to zero data retention and the provision of in-kind credits.

References are live-linked in the document. This article was written with AI assistance (GPT-5) to support drafting and synthesis. All ideas, projects, and experiences are my own.

0 Comments

Submit a Comment

Your email address will not be published. Required fields are marked *

Sinekhaya Nongabe

Sinekhaya Nongabe

Solutions Engineer
Sinekhaya is a Solutions Engineer on the Aselo project having joined the team in Feb 2023. With an appreciation for the transformative power of technology, he is passionate about optimising business processes and operations to deliver exceptional user experiences. His professional journey encompasses diverse roles in financial technology, banking, and higher education, leveraging his expertise gained since graduating with a B. Com in Information Systems from the University of Cape Town. Currently based in Cape Town, South Africa, Sinekhaya finds joy in the simple pleasures of life, often enjoying leisurely walks on the beach and indulging in African Fiction books.
Alejandro Rivera

Alejandro Rivera

Solutions Engineer
Alejandro is a software engineer with 7+ years of experience working in different positions related to software development. It wasn’t until he lived in rural Kenya that he realized the positive impact that technology can have on society. Alejandro has spent more than four years working for an NGO whose mission was to eradicate the hunger season across East Africa. He joined Tech Matters as a Solutions Engineer on the Aselo project to prove to himself once more that technology matters. In his spare time, he spends hours playing with his daughter, taking walks with his family, and growing a food forest.
Mythily Mudunuru

Mythily Mudunuru

Software Engineer
Mythily is a software engineer based in Toronto, Canada, who brings a unique blend of understanding user experience and technology expertise to her work. After earning her degree in Behavioral Science and Psychology from the University of Toronto, she spent nearly a decade as a therapist and consultant, working directly with clients to drive meaningful change. Her transition to software development was inspired by witnessing firsthand how technological solutions could scale impact. She joined Tech Matters to work on the Aselo project in Jan 2022. When she's not crafting code, you'll find her exploring Ontario's trails with her four-legged hiking companion, Ruby, discovering new destinations, or diving into whatever hobby has recently caught her curiosity.
Katy McKinney-Bock

Katy McKinney-Bock

Senior Data Scientist

Katy is a data scientist, linguist, and researcher with over a decade of experience in research development, data analysis, and machine learning / natural language processing. She is passionate about empowering people with data, and values knowledge creation via a multitude of methods, from recording narrative stories, to small team interviews, to mapping systems from administrative datasets, and in models trained over corpora of natural language or quantified via survey questions and in health assessments. She has worked with a range of collaborators, from non-profits providing human services, to a data incubator focused on ethics and societal impact, within an interdisciplinary research center at a medical university, and in higher education teaching and research, and takes a systems approach to engaging in research and technology development.
Humairaa Mohamed

Humairaa Mohamed

Partnerships Manager
Humairaa is a proud Africanist, intrapreneur, critical thinker, and problem solver with a high aptitude for logical reasoning, strategic thinking, and mathematical analysis. Fueled by a strong passion for humanity, nature, technology, and learning, she continuously seeks multidimensional challenge, growth, and improvement opportunities. Humairaa is inspired by the possibility of a world that delights in diversity and embraces complexity, and has chosen to dedicate her life and skills to committed teams and progressive technologies to deliver intelligent and sustainable impact.
Dee Luo

Dee Luo

Director of Product and Business Operations
Dee is a Product Manager interested in leveraging innovative technology to build empathy-driven solutions that empower individuals and their communities. She has experience building SaaS (Software-as-a-Service) business software after working at Yext, a New York-based technology company, and has worked on pro bono projects for international social enterprises as a past fellow of the MovingWorlds Institute. She joined Tech Matters in December 2020 to lead the product management for Aselo and is based in New York City.
Jana Kleitsch

Jana Kleitsch

User Experience Designer
Jana is a seasoned user experience designer and an occasional serial entrepreneur. After co-founding the first online wedding planning site, Wedding Channel, Jana spent almost a decade working on Amazon’s Recommendations and Bestsellers teams. Jana joined the Tech Matters team in March of 2020 as the UX Designer for the Aselo helpline platform. Jana also serves as a Startup Advisor at the University of Washington mentoring aspiring entrepreneurs.
Annalise Irby

Annalise Irby

Product Manager

Annalise is a Product Manager interested in co-creating mission-driven technology alongside communities and governments. Previously at Schmidt Futures, a tech philanthropy, she rotated as a product manager between tech nonprofits and startups including Recidiviz, a criminal justice reform nonprofit, and JustFix, a housing justice nonprofit. Before this, Annalise worked in software engineering at Lyft and Uber, and in conversational UX (user experience) design at IBM Research. She graduated from Georgia Institute of Technology with a B.S. in Computer Science, and is currently based in Brooklyn, NY. In her free time, you can find her making pottery, adventuring outdoors, or learning something new.
Steve Hand

Steve Hand

Software Engineer
Steve is a Software Engineer with 20-years’ experience leading projects in industries including e-commerce, healthcare, video games, road haulage, finance, and consumer electronics, in organizations ranging from a three-person start-up to large multinationals and the Welsh National Health Service. He came to software the long way around, after studying History & Economics at Manchester University. He joined Tech Matters to work on the Aselo project in December 2021 and is based in Galway, Ireland. In his free time, he helps out at local Autism charities but also consumes large amounts of bad TV.
Michele Gomich

Michele Gomich

Telecom Program Manager
Michele is a leading Senior Project Manager in Telecommunications, with over two decades of experience. Over the course of her career, Michele has managed a variety of Telecom Projects, specializing in implementing systems for Public Safety Entities, Large-Scale Municipalities, School Districts, National Accounts and Private Sector Corporations. Contact Centers, PBX Integrations, Cloud Base Telephony and Carrier Solutions are just a few of the platforms Michele has experience with. Michele has always prided herself on solving problems, and having the trust and respect of her peers and clients. With years of dedication in providing clients with cutting-edge Telecom Solutions, Michele is thrilled to put forth her efforts with Aselo and the Child Help Lines. It’s such an amazing and important cause. In her spare time, Michele can be found spending quality time with her husband and two children. Both her daughter and son keep the family super busy with their competitive gymnastics and musical talents. Michele was born and raised in New Jersey, and is now raising her family in the Upstate of South Carolina.
Nick Hurlburt

Nick Hurlburt

Executive Director

Nick has 17 years of experience joining two parallel paths: software engineering and international development. Nick began his career developing early, large-scale artificial intelligence software at Amazon.com. He later managed software teams at Sift, a Silicon Valley machine learning company that protects companies like Airbnb and Twitter from online fraud. Nick spent six years in conflict relief in Burma and South Sudan, capturing human rights, migration, and multi-sector program data using methods from hand-delivered paper to smartphones to satellite modems. He leads the technology development for the Aselo helpline project, joining Tech Matters in mid-2019.

Pin It on Pinterest