The Challenge: Hidden Self-Service Content

Most customer service teams have already invested heavily in FAQs, help centers, and product documentation. Yet customers still raise tickets for issues that have been answered many times before. The core problem is hidden self-service content: the right article exists, but customers can’t discover it quickly enough, so they default to contacting support.

Traditional approaches rely on manual content audits, basic keyword search, and navigation tweaks based on intuition. These methods don’t scale when you have thousands of articles, multiple languages, and constantly changing products. Search engines that match only exact keywords miss the fact that customers describe problems in their own language, not in your internal terminology.

The business impact is significant. Hidden content leads directly to avoidable ticket volume, higher support costs, and longer wait times for everyone. Agents are forced to answer the same simple questions again and again instead of focusing on complex cases or revenue-generating interactions. Over time, this erodes customer satisfaction and creates a competitive disadvantage against companies that offer truly effective self-service experiences.

This challenge is real, but it is absolutely solvable. Modern AI — and tools like Claude in particular — can read your entire knowledge base, ticket history and search logs, then surface where your self-service content is failing to connect with customer intent. At Reruption, we’ve seen how the right AI approach can turn a static help center into a living system that learns from every interaction. Below, you’ll find practical guidance on how to do this in your own customer service organisation.

Need a sparring partner for this challenge?

Let's have a no-obligation chat and brainstorm together.

Innovators at these companies trust us:

Our Assessment

A strategic assessment of the challenge and high-level tips how to tackle it.

From Reruption’s work implementing AI in customer service, we’ve seen that the fastest wins often come from fixing how customers find answers, not from rewriting everything from scratch. Claude is particularly strong here: it can process large knowledge bases, understand natural language queries, and analyse ticket logs to reveal where your self-service experience is breaking. The key is approaching it strategically, not just as another chatbot project.

Think in Customer Intent, Not Internal Categories

Most help centers are organised around how the company thinks about products and processes. Customers come with intents: “cancel my order”, “reset my password”, “my invoice is wrong”. The gap between those two views is at the heart of hidden self-service content. Strategically, you need to use Claude to model and prioritize customer intents, then reshape your self-service around them.

Start by feeding Claude anonymised search logs and ticket subjects/descriptions. Ask it to cluster and label common intents in the language customers actually use. At a leadership level, this becomes your new map: instead of thinking in terms of FAQ categories, you plan your AI-powered self-service roadmap around the top 50–100 intents and their impact on volume and cost.

Use Claude as a Discovery Engine, Not Just a Chatbot

Many organisations treat AI as a front-end chatbot that sits on top of the same broken navigation. That’s a missed opportunity. Strategically, Claude should first be your discovery engine: a system that reads everything — articles, macros, product docs, previous tickets — and tells you where content is missing, redundant, or badly structured.

Give Claude your entire content corpus plus a representative set of tickets. Ask it: for each high-volume intent, is there a clear, up-to-date, customer-friendly article? Where is content too long, too technical, or conflicting? Leadership can then make informed decisions about what to consolidate, what to rewrite, and where a conversational experience will add real deflection value.

Align Customer Service, Product, and Knowledge Management

Hidden self-service content is rarely just a tools problem; it’s an organisational one. Content ownership is often fragmented between support, product, and technical documentation. Before you scale Claude, you need a clear strategic model for who owns which parts of the knowledge base and how AI-generated insights will be actioned.

Set up a cross-functional working group where support leaders bring volume and pain-point data, product brings roadmap context, and knowledge managers bring content standards. Claude then becomes a shared asset: its analyses and drafts feed into a unified backlog of improvements, with clear SLAs on how quickly high-impact gaps will be addressed.

Design for Governance, Not One-Off Improvements

A one-time cleanup of your help center will help for a few months, then decay. Strategically, you want an ongoing knowledge governance loop driven by Claude: continuously monitoring where customers search, which articles they bounce from, and which intents still end up as tickets.

Define governance rules upfront: how often Claude should re-analyse logs, what thresholds trigger content reviews, who approves AI-generated article changes, and how you measure the impact on ticket deflection. This prevents "AI chaos" and builds trust that Claude is improving your self-service in a controlled way rather than rewriting your knowledge base overnight.

Manage Risk Around Accuracy, Compliance and Tone

When you let an AI system interact with customers or draft help content, strategic risk management is essential. You need policies around factual accuracy, data privacy, regulatory requirements, and brand voice. Claude’s strength in following instructions is a benefit here, but only if those instructions are designed thoughtfully.

At a strategic level, decide which topics are safe for fully automated responses and which always need a human in the loop. Define guardrails in Claude prompts and system messages (for example, not guessing legal or financial details) and align with legal and compliance teams early. This creates the confidence to scale AI in customer service without exposing the business to unnecessary risk.

Used strategically, Claude becomes the connective tissue between what your customers ask and the content you already have — or should have — to answer them. Instead of launching yet another generic chatbot, you can systematically expose and fix hidden self-service gaps, then power a conversational layer that actually deflects tickets. Reruption combines this AI depth with a Co-Preneur mindset, embedding with your customer service team to turn these ideas into working systems. If you’re ready to make your help center truly work for customers, we can help you design, test, and scale the right Claude-based approach.

Need help implementing these ideas?

Feel free to reach out to us with no obligation.

Real-World Case Studies

From Fintech to Apparel Retail: Learn how companies successfully use Claude.

Klarna

Fintech

Klarna, a leading fintech BNPL provider, faced enormous pressure from millions of customer service inquiries across multiple languages for its 150 million users worldwide. Queries spanned complex fintech issues like refunds, returns, order tracking, and payments, requiring high accuracy, regulatory compliance, and 24/7 availability. Traditional human agents couldn't scale efficiently, leading to long wait times averaging 11 minutes per resolution and rising costs. Additionally, providing personalized shopping advice at scale was challenging, as customers expected conversational, context-aware guidance across retail partners. Multilingual support was critical in markets like US, Europe, and beyond, but hiring multilingual agents was costly and slow. This bottleneck hindered growth and customer satisfaction in a competitive BNPL sector.

Lösung

Klarna partnered with OpenAI to deploy a generative AI chatbot powered by GPT-4, customized as a multilingual customer service assistant. The bot handles refunds, returns, order issues, and acts as a conversational shopping advisor, integrated seamlessly into Klarna's app and website. Key innovations included fine-tuning on Klarna's data, retrieval-augmented generation (RAG) for real-time policy access, and safeguards for fintech compliance. It supports dozens of languages, escalating complex cases to humans while learning from interactions. This AI-native approach enabled rapid scaling without proportional headcount growth.

Ergebnisse

  • 2/3 of all customer service chats handled by AI
  • 2.3 million conversations in first month alone
  • Resolution time: 11 minutes → 2 minutes (82% reduction)
  • CSAT: 4.4/5 (AI) vs. 4.2/5 (humans)
  • $40 million annual cost savings
  • Equivalent to 700 full-time human agents
  • 80%+ queries resolved without human intervention
Read case study →

Morgan Stanley

Banking

Financial advisors at Morgan Stanley struggled with rapid access to the firm's extensive proprietary research database, comprising over 350,000 documents spanning decades of institutional knowledge. Manual searches through this vast repository were time-intensive, often taking 30 minutes or more per query, hindering advisors' ability to deliver timely, personalized advice during client interactions . This bottleneck limited scalability in wealth management, where high-net-worth clients demand immediate, data-driven insights amid volatile markets. Additionally, the sheer volume of unstructured data—40 million words of research reports—made it challenging to synthesize relevant information quickly, risking suboptimal recommendations and reduced client satisfaction. Advisors needed a solution to democratize access to this 'goldmine' of intelligence without extensive training or technical expertise .

Lösung

Morgan Stanley partnered with OpenAI to develop AI @ Morgan Stanley Debrief, a GPT-4-powered generative AI chatbot tailored for wealth management advisors. The tool uses retrieval-augmented generation (RAG) to securely query the firm's proprietary research database, providing instant, context-aware responses grounded in verified sources . Implemented as a conversational assistant, Debrief allows advisors to ask natural-language questions like 'What are the risks of investing in AI stocks?' and receive synthesized answers with citations, eliminating manual digging. Rigorous AI evaluations and human oversight ensure accuracy, with custom fine-tuning to align with Morgan Stanley's institutional knowledge . This approach overcame data silos and enabled seamless integration into advisors' workflows.

Ergebnisse

  • 98% adoption rate among wealth management advisors
  • Access for nearly 50% of Morgan Stanley's total employees
  • Queries answered in seconds vs. 30+ minutes manually
  • Over 350,000 proprietary research documents indexed
  • 60% employee access at peers like JPMorgan for comparison
  • Significant productivity gains reported by CAO
Read case study →

Ford Motor Company

Manufacturing

In Ford's automotive manufacturing plants, vehicle body sanding and painting represented a major bottleneck. These labor-intensive tasks required workers to manually sand car bodies, a process prone to inconsistencies, fatigue, and ergonomic injuries due to repetitive motions over hours . Traditional robotic systems struggled with the variability in body panels, curvatures, and material differences, limiting full automation in legacy 'brownfield' facilities . Additionally, achieving consistent surface quality for painting was critical, as defects could lead to rework, delays, and increased costs. With rising demand for electric vehicles (EVs) and production scaling, Ford needed to modernize without massive CapEx or disrupting ongoing operations, while prioritizing workforce safety and upskilling . The challenge was to integrate scalable automation that collaborated with humans seamlessly.

Lösung

Ford addressed this by deploying AI-guided collaborative robots (cobots) equipped with machine vision and automation algorithms. In the body shop, six cobots use cameras and AI to scan car bodies in real-time, detecting surfaces, defects, and contours with high precision . These systems employ computer vision models for 3D mapping and path planning, allowing cobots to adapt dynamically without reprogramming . The solution emphasized a workforce-first brownfield strategy, starting with pilot deployments in Michigan plants. Cobots handle sanding autonomously while humans oversee quality, reducing injury risks. Partnerships with robotics firms and in-house AI development enabled low-code inspection tools for easy scaling .

Ergebnisse

  • Sanding time: 35 seconds per full car body (vs. hours manually)
  • Productivity boost: 4x faster assembly processes
  • Injury reduction: 70% fewer ergonomic strains in cobot zones
  • Consistency improvement: 95% defect-free surfaces post-sanding
  • Deployment scale: 6 cobots operational, expanding to 50+ units
  • ROI timeline: Payback in 12-18 months per plant
Read case study →

Netflix

Streaming Media

With over 17,000 titles and growing, Netflix faced the classic cold start problem and data sparsity in recommendations, where new users or obscure content lacked sufficient interaction data, leading to poor personalization and higher churn rates . Viewers often struggled to discover engaging content among thousands of options, resulting in prolonged browsing times and disengagement—estimated at up to 75% of session time wasted on searching rather than watching . This risked subscriber loss in a competitive streaming market, where retaining users costs far less than acquiring new ones. Scalability was another hurdle: handling 200M+ subscribers generating billions of daily interactions required processing petabytes of data in real-time, while evolving viewer tastes demanded adaptive models beyond traditional collaborative filtering limitations like the popularity bias favoring mainstream hits . Early systems post-Netflix Prize (2006-2009) improved accuracy but struggled with contextual factors like device, time, and mood .

Lösung

Netflix built a hybrid recommendation engine combining collaborative filtering (CF)—starting with FunkSVD and Probabilistic Matrix Factorization from the Netflix Prize—and advanced deep learning models for embeddings and predictions . They consolidated multiple use-case models into a single multi-task neural network, improving performance and maintainability while supporting search, home page, and row recommendations . Key innovations include contextual bandits for exploration-exploitation, A/B testing on thumbnails and metadata, and content-based features from computer vision/audio analysis to mitigate cold starts . Real-time inference on Kubernetes clusters processes 100s of millions of predictions per user session, personalized by viewing history, ratings, pauses, and even search queries . This evolved from 2009 Prize winners to transformer-based architectures by 2023 .

Ergebnisse

  • 80% of viewer hours from recommendations
  • $1B+ annual savings in subscriber retention
  • 75% reduction in content browsing time
  • 10% RMSE improvement from Netflix Prize CF techniques
  • 93% of views from personalized rows
  • Handles billions of daily interactions for 270M subscribers
Read case study →

Duolingo

EdTech

Duolingo, a leader in gamified language learning, faced key limitations in providing real-world conversational practice and in-depth feedback. While its bite-sized lessons built vocabulary and basics effectively, users craved immersive dialogues simulating everyday scenarios, which static exercises couldn't deliver . This gap hindered progression to fluency, as learners lacked opportunities for free-form speaking and nuanced grammar explanations without expensive human tutors. Additionally, content creation was a bottleneck. Human experts manually crafted lessons, slowing the rollout of new courses and languages amid rapid user growth. Scaling personalized experiences across 40+ languages demanded innovation to maintain engagement without proportional resource increases . These challenges risked user churn and limited monetization in a competitive EdTech market.

Lösung

Duolingo launched Duolingo Max in March 2023, a premium subscription powered by GPT-4, introducing Roleplay for dynamic conversations and Explain My Answer for contextual feedback . Roleplay simulates real-life interactions like ordering coffee or planning vacations with AI characters, adapting in real-time to user inputs. Explain My Answer provides detailed breakdowns of correct/incorrect responses, enhancing comprehension. Complementing this, Duolingo's Birdbrain LLM (fine-tuned on proprietary data) automates lesson generation, allowing experts to create content 10x faster . This hybrid human-AI approach ensured quality while scaling rapidly, integrated seamlessly into the app for all skill levels .

Ergebnisse

  • DAU Growth: +59% YoY to 34.1M (Q2 2024)
  • DAU Growth: +54% YoY to 31.4M (Q1 2024)
  • Revenue Growth: +41% YoY to $178.3M (Q2 2024)
  • Adjusted EBITDA Margin: 27.0% (Q2 2024)
  • Lesson Creation Speed: 10x faster with AI
  • User Self-Efficacy: Significant increase post-AI use (2025 study)
Read case study →

Best Practices

Successful implementations follow proven patterns. Have a look at our tactical advice to get started.

Use Claude to Map Search Gaps and Missed Deflection Opportunities

The first tactical step is to quantify where your self-service is failing. Export a few weeks or months of search queries from your help center and a sample of support tickets (subjects, descriptions, tags, and resolutions). Feed these into Claude in batches and ask it to identify themes where customers searched but didn’t click, or searched and still opened a ticket.

Prompt example for Claude:
You are an analyst helping improve a customer service knowledge base.
You will receive:
1) A list of customer search queries and whether they clicked any article
2) A list of related support tickets with subjects and resolution notes

Tasks:
- Cluster the search queries into 10–15 intent groups
- For each cluster, indicate:
  - How many searches had no clicks
  - How many tickets were opened for that intent
- Highlight the 5 clusters with the biggest deflection opportunity
- Suggest what self-service content is missing or hard to find

This gives you a data-driven map of where hidden content or navigation issues are driving avoidable volume, so you can prioritise the highest-impact fixes.

Restructure Long Articles into FAQ-Style, Searchable Answers

Many knowledge bases are dominated by long, technical articles that are hard to scan. Claude is excellent at transforming dense documents into concise, FAQ-style content that matches how customers actually ask questions. Start by exporting your most-viewed or most-referenced articles and passing them to Claude with clear restructuring instructions.

Prompt example for Claude:
You are a customer service documentation specialist.
Here is an article from our help center. Rewrite it as:
- A short summary in plain language (max 3 sentences)
- 5–10 FAQ questions and answers in the exact phrases a customer would use
- Each answer should be 2–4 short paragraphs, with clear steps
- Avoid internal jargon; use the customer's language from these example queries: [insert]

Keep all factual content unchanged. If anything is ambiguous, highlight it in a note.

Import the restructured content back into your knowledge system, using FAQ questions as titles, H2s, or search synonyms. This makes it much easier for search and AI chat to retrieve relevant snippets that match user intent.

Create an AI-Powered Help Center Guide with Retrieval-Augmented Generation

Beyond static search, you can use Claude to build an AI-powered help assistant that reads from your existing knowledge base using retrieval-augmented generation (RAG). The idea: when a customer asks a question, your system retrieves the most relevant articles and passes them to Claude, which then synthesises a precise answer and links to the sources.

System message example for Claude in a RAG setup:
You are a customer support assistant for [Company].
You can ONLY answer using the information provided in the context documents.
If the answer is not in the documents, say you don't know and suggest contacting support.
Always include links to the exact articles you used.
Use friendly, concise language, and avoid internal codes or jargon.

User question: [customer query]
Context documents: [top 3–5 relevant article excerpts]

On the implementation side, this typically requires: connecting your CMS/knowledge base to an embedding store, building a retrieval endpoint, and integrating Claude via API in your help widget or portal. The result is a guided, conversational experience that surfaces the right content at the right time.

Auto-Draft Missing or Outdated Articles from Ticket Histories

Where your analysis shows clear gaps, Claude can dramatically speed up content creation by generating first drafts directly from ticket histories and agent macros. Select a set of resolved tickets for a specific intent, including the final agent responses and any internal notes, and have Claude propose a customer-friendly article.

Prompt example for Claude:
You are creating a public help center article from real support tickets.
Input:
- 10–20 anonymised tickets about the same issue
- The final agent replies and resolution steps

Tasks:
- Infer the underlying customer problem and write a clear problem statement
- Describe the solution in step-by-step form, in language suitable for non-experts
- Add a short "Before you start" checklist if needed
- Add a section "When to contact support" for edge cases we cannot solve via self-service

Do NOT include any personal data or internal system names.

Have a knowledge manager or senior agent review and approve these drafts before publication. This can cut article creation time from hours to minutes while ensuring content accurately reflects how issues are actually resolved.

Support Agents with Real-Time Article Suggestions and Case Summaries

Even with better self-service, some customers will always contact support. You can still improve deflection and consistency by giving agents Claude-powered side tools that suggest relevant content and summarise cases in real time. For example, when a ticket arrives, Claude can read the conversation, propose likely intents, and surface the top three articles or macros for the agent.

Prompt example for Claude inside an agent assist tool:
You are an assistant for customer service agents.
Input:
- The full conversation between the agent and the customer
- A list of available help center articles with titles and short summaries

Tasks:
- Summarise the customer's issue in 2 sentences
- Suggest the 3 most relevant articles, with a one-line rationale each
- Propose a short, friendly reply that uses links to those articles

Respond in JSON with keys: summary, suggested_articles, draft_reply.

This keeps agents aligned with the latest self-service content, encourages consistent linking to help articles, and trains customers to look to the help center first next time.

Measure Deflection and Continuously Optimise with Claude

To close the loop, you need to track whether these changes actually reduce ticket volume. Define clear KPIs such as self-service success rate, proportion of searches leading to resolved sessions, and the percentage of intents handled without agent intervention. Use Claude regularly to analyse logs and propose experiments.

Prompt example for Claude:
You are a CX analyst.
Here is data for the last 30 days:
- Search queries and click behaviour
- Chatbot conversations and handover rates
- Ticket volume by intent

Tasks:
- Identify 5 knowledge base improvements that are likely to increase self-service success
- For each, estimate potential ticket reduction based on the data
- Propose an A/B test or small experiment to validate the impact

Expected outcomes from a well-implemented setup are realistic but meaningful: 15–30% reduction in repetitive tickets for targeted intents within 3–6 months, improved first-contact resolution, and shorter handling times as agents work with better content and summaries. The exact numbers will depend on your baseline, but with disciplined measurement and iterative optimisation, Claude can become a core engine for continuous deflection improvement.

Need implementation expertise now?

Let's talk about your ideas!

Frequently Asked Questions

Claude can process your entire knowledge base, search logs, and historical tickets to identify where customers look for answers but fail to find them. Practically, you upload or connect:

  • Help center articles, FAQs, and internal docs
  • Search queries plus click/bounce data
  • Ticket subjects, descriptions, tags, and resolutions

Claude then clusters customer intents, highlights topics with high search volume but poor self-service resolution, and suggests where content is missing, outdated, or badly structured. It can also draft or restructure articles so that they directly match the language your customers use, which significantly increases the chance that existing content will be found and used.

You typically need three core capabilities: a customer service lead who understands your main contact drivers, a knowledge manager or content owner, and basic engineering capacity to connect Claude to your systems (help center, ticketing, and possibly a vector database for retrieval). If you don’t have internal AI expertise, a partner like Reruption can handle the technical architecture, prompt design, and integration work.

From your side, the most important inputs are access to data (knowledge base exports, logs, tickets) and decision-making capacity to prioritise which intents to tackle first. You do not need a large in-house data science team to start; many organisations begin with a focused PoC and a small cross-functional squad.

For a focused scope (e.g. the top 10–20 repetitive intents), you can usually see first effects within 4–8 weeks. The typical timeline looks like this:

  • Week 1–2: Data extraction, analysis of search/ticket gaps with Claude, intent clustering
  • Week 3–4: Drafting and restructuring key articles, initial AI assistant or improved search configuration
  • Week 5–8: Go-live for a subset of traffic, measurement of self-service success, iterative tuning

Substantial, portfolio-wide deflection (15–30% on repetitive tickets) usually emerges over 3–6 months as you iterate across more intents, improve content quality, and refine your AI-powered self-service flows.

The direct Claude API usage costs are typically modest compared to support headcount costs. The main investments are in initial design, integration, and content work. ROI comes from reduced repetitive ticket volume, lower handling times, and improved customer satisfaction.

As a rough benchmark, if even 10–20% of your low-complexity tickets are deflected via better self-service and AI assistance, the savings in agent time usually pay back the project within months. Reruption helps you define clear metrics (e.g. cost per contact, deflection rate by intent) and set up measurement so you can quantify ROI rather than relying on gut feel.

Reruption works as a Co-Preneur, embedding with your customer service and IT teams to build real AI solutions instead of just providing slideware. We typically start with our AI PoC for 9,900€, where we define a concrete deflection use case, integrate Claude with a subset of your knowledge base and ticket data, and deliver a working prototype along with performance metrics and an implementation roadmap.

From there, we can support hands-on implementation: designing retrieval-augmented search or chat flows, setting up governance and prompts, restructuring content at scale, and integrating with your existing tools. The goal is not to optimise your current help center slightly, but to build the AI-first customer service capabilities that will replace it over time.

Contact Us!

0/10 min.

Contact Directly

Your Contact

Philipp M. W. Hoffmann

Founder & Partner

Address

Reruption GmbH

Falkertstraße 2

70176 Stuttgart

Social Media