The Challenge: Inefficient Audience Segmentation

Most marketing teams still rely on blunt tools for audience segmentation: a few demographic filters, basic interests, maybe one or two lifecycle rules. It feels structured, but in reality it’s little more than educated guesswork. As data volume and channel complexity explode, this approach simply cannot keep up with the way customers actually behave across touchpoints.

Traditional segmentation is typically built once per quarter in a spreadsheet, then pushed into the ad platforms and CRM. It ignores micro-behaviors, evolving intent, and cross-channel signals. Static rules like “visited product page twice” or “opened last 3 emails” miss deeper patterns in frequency, timing, product mix, and engagement decay. Even when data teams are involved, the process of writing custom SQL or models for each new hypothesis is slow, expensive, and frustrating for marketers who need agility.

The business impact is substantial. Blunt targeting drives up CAC as low-value users see too many ads, while high-potential cohorts never get the differentiated offers, creatives, or journeys they deserve. Budget gets spread evenly instead of being focused on high-LTV clusters. Personalization efforts underperform, experimentation slows down, and competitors with more sophisticated segmentation quietly win the most valuable segments in your market. Over time, the gap in ROI and marketing efficiency compounds.

This challenge is real, but it’s solvable. With the right use of AI-driven segmentation, you can uncover patterns that no human analyst would ever think to test, and operationalize them directly into your campaigns. At Reruption, we’ve seen how combining strong engineering with a practical marketing lens turns unusable data into actionable audiences. Below, you’ll find concrete guidance on how to use Gemini with your Google data stack to move beyond static rules and build segmentation that finally matches how your customers behave.

Need a sparring partner for this challenge?

Let's have a no-obligation chat and brainstorm together.

Innovators at these companies trust us:

Our Assessment

A strategic assessment of the challenge and high-level tips how to tackle it.

From Reruption’s perspective, the real opportunity isn’t just using Gemini as another content generator, but as an intelligence layer on top of your BigQuery and Google Ads data. With hands-on experience building AI solutions inside marketing teams, we’ve seen that the combination of Gemini + your existing data warehouse can transform inefficient, rule-based segmentation into dynamic, behavior-driven clusters that marketers can actually use without relying on a data team for every change.

Start from Business Value, Not from Algorithms

When using Gemini for audience segmentation, the starting point shouldn’t be clustering techniques or feature engineering – it should be marketing and commercial goals. Before you touch data, define which outcomes matter: lower CAC for performance campaigns, higher LTV in retention programs, or improved conversion rates on key funnels. This gives Gemini clear “north stars” when it helps you derive segments and write SQL queries.

Translate these goals into concrete questions: “Which behavioral patterns predict repeat purchase within 60 days?” or “Which users show intent but stall before adding to cart?” When we embed with teams, we force this discipline early. It keeps the AI work focused on value-creating clusters, not just mathematically interesting ones that never make it into campaigns.

Treat Gemini as a Data Co-Analyst for Marketers

Strategically, the power of Gemini + BigQuery is to reduce dependency on scarce data engineering resources. Marketers should be able to collaborate with Gemini like a co-analyst: describe the audience they care about in plain language and have Gemini draft the SQL or segmentation logic against your schema. That changes the team’s operating model from ticket-driven to exploratory and hypothesis-driven.

This requires an intentional shift in mindset. Instead of asking “What segments can the data team give us this quarter?”, marketers start asking “What segments do we want to test this week?” and use Gemini to iterate. You still need guardrails and review, but the strategic trajectory is towards self-service, not more centralization.

Invest in Data Readiness Before Scaling Personalization

No matter how good Gemini is, it can’t segment on signals you don’t collect or can’t trust. Strategically, you need a minimum level of data quality and structure in BigQuery: consistent user IDs or stitched identities, clean event tracking, and clear mapping between marketing events and business outcomes (purchases, leads, churn, etc.).

Instead of trying to fix everything, focus on the 3–5 most critical signals for segmentation: recency/frequency/monetary value, key intent events (e.g. product views, trials started), and engagement across main channels. Our experience shows that once these are reliable, Gemini can do meaningful work on top, and the team builds trust in the AI-generated segments.

Design Governance and Validation Around AI-Generated Segments

Using AI-generated audience clusters without governance is risky. Strategically, define how segments move from “experimental” to “production”. For example: initial clusters are tested on 5–10% of budget with clear success criteria; only after hitting predefined thresholds (conversion rate, ROAS, engagement) do they become standard audiences.

Make sure someone owns validation. A marketer and an analyst (or Gemini plus a human reviewer) should regularly sanity-check segment definitions, size, and overlap. This governance layer is where you mitigate the risk of overfitting, biased segments, or simply segments that are not actionable at scale.

Prepare Your Team for an AI-First Way of Working

Gemini will change how your marketing team works with data. Strategically, that means upskilling marketers to be comfortable with concepts like schemas, events, and simple SQL prompts – not to turn them into engineers, but to make them effective AI users. It also means aligning with data and IT on roles and responsibilities so Gemini doesn’t become a shadow-IT project.

We’ve seen that teams that embrace an AI-first lens – asking “If we rebuilt segmentation with AI from scratch, how would it work?” – move much faster than those trying to bolt Gemini onto legacy processes. Align expectations early: Gemini will not be perfect from day one, but it will drastically shorten the path from idea to live segment if the team is ready to iterate with it.

Used thoughtfully, Gemini can turn inefficient, rule-based segmentation into a living system of behavior-driven clusters that marketers can explore and deploy in days, not quarters. The key is to combine business-focused goals, data readiness, and clear governance so that AI-generated segments are both powerful and safe to use. At Reruption, we’ve built similar AI-first capabilities inside organisations and can help you scope, prototype, and operationalize a Gemini-powered segmentation engine that fits your stack and team. If you’re considering this step, a short conversation and a focused AI PoC can quickly show what’s realistically achievable in your environment.

Need help implementing these ideas?

Feel free to reach out to us with no obligation.

Real-World Case Studies

From Manufacturing to Manufacturing: Learn how companies successfully use Gemini.

Ford Motor Company

Manufacturing

In Ford's automotive manufacturing plants, vehicle body sanding and painting represented a major bottleneck. These labor-intensive tasks required workers to manually sand car bodies, a process prone to inconsistencies, fatigue, and ergonomic injuries due to repetitive motions over hours . Traditional robotic systems struggled with the variability in body panels, curvatures, and material differences, limiting full automation in legacy 'brownfield' facilities . Additionally, achieving consistent surface quality for painting was critical, as defects could lead to rework, delays, and increased costs. With rising demand for electric vehicles (EVs) and production scaling, Ford needed to modernize without massive CapEx or disrupting ongoing operations, while prioritizing workforce safety and upskilling . The challenge was to integrate scalable automation that collaborated with humans seamlessly.

Lösung

Ford addressed this by deploying AI-guided collaborative robots (cobots) equipped with machine vision and automation algorithms. In the body shop, six cobots use cameras and AI to scan car bodies in real-time, detecting surfaces, defects, and contours with high precision . These systems employ computer vision models for 3D mapping and path planning, allowing cobots to adapt dynamically without reprogramming . The solution emphasized a workforce-first brownfield strategy, starting with pilot deployments in Michigan plants. Cobots handle sanding autonomously while humans oversee quality, reducing injury risks. Partnerships with robotics firms and in-house AI development enabled low-code inspection tools for easy scaling .

Ergebnisse

  • Sanding time: 35 seconds per full car body (vs. hours manually)
  • Productivity boost: 4x faster assembly processes
  • Injury reduction: 70% fewer ergonomic strains in cobot zones
  • Consistency improvement: 95% defect-free surfaces post-sanding
  • Deployment scale: 6 cobots operational, expanding to 50+ units
  • ROI timeline: Payback in 12-18 months per plant
Read case study →

Rolls-Royce Holdings

Aerospace

Jet engines are highly complex, operating under extreme conditions with millions of components subject to wear. Airlines faced unexpected failures leading to costly groundings, with unplanned maintenance causing millions in daily losses per aircraft. Traditional scheduled maintenance was inefficient, often resulting in over-maintenance or missed issues, exacerbating downtime and fuel inefficiency. Rolls-Royce needed to predict failures proactively amid vast data from thousands of engines in flight. Challenges included integrating real-time IoT sensor data (hundreds per engine), handling terabytes of telemetry, and ensuring accuracy in predictions to avoid false alarms that could disrupt operations. The aerospace industry's stringent safety regulations added pressure to deliver reliable AI without compromising performance.

Lösung

Rolls-Royce developed the IntelligentEngine platform, combining digital twins—virtual replicas of physical engines—with machine learning models. Sensors stream live data to cloud-based systems, where ML algorithms analyze patterns to predict wear, anomalies, and optimal maintenance windows. Digital twins enable simulation of engine behavior pre- and post-flight, optimizing designs and schedules. Partnerships with Microsoft Azure IoT and Siemens enhanced data processing and VR modeling, scaling AI across Trent series engines like Trent 7000 and 1000. Ethical AI frameworks ensure data security and bias-free predictions.

Ergebnisse

  • 48% increase in time on wing before first removal
  • Doubled Trent 7000 engine time on wing
  • Reduced unplanned downtime by up to 30%
  • Improved fuel efficiency by 1-2% via optimized ops
  • Cut maintenance costs by 20-25% for operators
  • Processed terabytes of real-time data from 1000s of engines
Read case study →

Khan Academy

Education

Khan Academy faced the monumental task of providing personalized tutoring at scale to its 100 million+ annual users, many in under-resourced areas. Traditional online courses, while effective, lacked the interactive, one-on-one guidance of human tutors, leading to high dropout rates and uneven mastery. Teachers were overwhelmed with planning, grading, and differentiation for diverse classrooms. In 2023, as AI advanced, educators grappled with hallucinations and over-reliance risks in tools like ChatGPT, which often gave direct answers instead of fostering learning. Khan Academy needed an AI that promoted step-by-step reasoning without cheating, while ensuring equitable access as a nonprofit. Scaling safely across subjects and languages posed technical and ethical hurdles.

Lösung

Khan Academy developed Khanmigo, an AI-powered tutor and teaching assistant built on GPT-4, piloted in March 2023 for teachers and expanded to students. Unlike generic chatbots, Khanmigo uses custom prompts to guide learners Socratically—prompting questions, hints, and feedback without direct answers—across math, science, humanities, and more. The nonprofit approach emphasized safety guardrails, integration with Khan's content library, and iterative improvements via teacher feedback. Partnerships like Microsoft enabled free global access for teachers by 2024, now in 34+ languages. Ongoing updates, such as 2025 math computation enhancements, address accuracy challenges.

Ergebnisse

  • User Growth: 68,000 (2023-24 pilot) to 700,000+ (2024-25 school year)
  • Teacher Adoption: Free for teachers in most countries, millions using Khan Academy tools
  • Languages Supported: 34+ for Khanmigo
  • Engagement: Improved student persistence and mastery in pilots
  • Time Savings: Teachers save hours on lesson planning and prep
  • Scale: Integrated with 429+ free courses in 43 languages
Read case study →

Bank of America

Banking

Bank of America faced a high volume of routine customer inquiries, such as account balances, payments, and transaction histories, overwhelming traditional call centers and support channels. With millions of daily digital banking users, the bank struggled to provide 24/7 personalized financial advice at scale, leading to inefficiencies, longer wait times, and inconsistent service quality. Customers demanded proactive insights beyond basic queries, like spending patterns or financial recommendations, but human agents couldn't handle the sheer scale without escalating costs. Additionally, ensuring conversational naturalness in a regulated industry like banking posed challenges, including compliance with financial privacy laws, accurate interpretation of complex queries, and seamless integration into the mobile app without disrupting user experience. The bank needed to balance AI automation with human-like empathy to maintain trust and high satisfaction scores.

Lösung

Bank of America developed Erica, an in-house NLP-powered virtual assistant integrated directly into its mobile banking app, leveraging natural language processing and predictive analytics to handle queries conversationally. Erica acts as a gateway for self-service, processing routine tasks instantly while offering personalized insights, such as cash flow predictions or tailored advice, using client data securely. The solution evolved from a basic navigation tool to a sophisticated AI, incorporating generative AI elements for more natural interactions and escalating complex issues to human agents seamlessly. Built with a focus on in-house language models, it ensures control over data privacy and customization, driving enterprise-wide AI adoption while enhancing digital engagement.

Ergebnisse

  • 3+ billion total client interactions since 2018
  • Nearly 50 million unique users assisted
  • 58+ million interactions per month (2025)
  • 2 billion interactions reached by April 2024 (doubled from 1B in 18 months)
  • 42 million clients helped by 2024
  • 19% earnings spike linked to efficiency gains
Read case study →

NYU Langone Health

Healthcare

NYU Langone Health, a leading academic medical center, faced significant hurdles in leveraging the vast amounts of unstructured clinical notes generated daily across its network. Traditional clinical predictive models relied heavily on structured data like lab results and vitals, but these required complex ETL processes that were time-consuming and limited in scope. Unstructured notes, rich with nuanced physician insights, were underutilized due to challenges in natural language processing, hindering accurate predictions of critical outcomes such as in-hospital mortality, length of stay (LOS), readmissions, and operational events like insurance denials. Clinicians needed real-time, scalable tools to identify at-risk patients early, but existing models struggled with the volume and variability of EHR data—over 4 million notes spanning a decade. This gap led to reactive care, increased costs, and suboptimal patient outcomes, prompting the need for an innovative approach to transform raw text into actionable foresight.

Lösung

To address these challenges, NYU Langone's Division of Applied AI Technologies at the Center for Healthcare Innovation and Delivery Science developed NYUTron, a proprietary large language model (LLM) specifically trained on internal clinical notes. Unlike off-the-shelf models, NYUTron was fine-tuned on unstructured EHR text from millions of encounters, enabling it to serve as an all-purpose prediction engine for diverse tasks. The solution involved pre-training a 13-billion-parameter LLM on over 10 years of de-identified notes (approximately 4.8 million inpatient notes), followed by task-specific fine-tuning. This allowed seamless integration into clinical workflows, automating risk flagging directly from physician documentation without manual data structuring. Collaborative efforts, including AI 'Prompt-a-Thons,' accelerated adoption by engaging clinicians in model refinement.

Ergebnisse

  • AUROC: 0.961 for 48-hour mortality prediction (vs. 0.938 benchmark)
  • 92% accuracy in identifying high-risk patients from notes
  • LOS prediction AUROC: 0.891 (5.6% improvement over prior models)
  • Readmission prediction: AUROC 0.812, outperforming clinicians in some tasks
  • Operational predictions (e.g., insurance denial): AUROC up to 0.85
  • 24 clinical tasks with superior performance across mortality, LOS, and comorbidities
Read case study →

Best Practices

Successful implementations follow proven patterns. Have a look at our tactical advice to get started.

Connect Gemini to Your BigQuery Marketing Dataset with Clear Context

To get useful AI-driven segmentation, Gemini needs more than raw data – it needs context. Start by ensuring your key marketing tables (events, sessions, transactions, campaigns) are in BigQuery with consistent user identifiers. Then document the schema in plain language: what each table represents, primary keys, and how events map to key funnel steps.

When you interact with Gemini (e.g. via Vertex AI, the Gemini API, or a connected notebook environment), always include this context at the start of the conversation. Provide table names, column descriptions, and example queries that follow your internal conventions. This dramatically improves the quality and safety of the SQL Gemini will generate.

System / setup prompt for Gemini:
You are a senior marketing data analyst.
You write safe, BigQuery-compatible SQL for segmentation.

Schema context:
- Table: events (user_id, event_name, event_time, source, campaign_id, revenue)
- Table: users (user_id, signup_date, country, device_type)
- Table: campaigns (campaign_id, channel, objective, cost)

Always:
- Use fully qualified table names: `project.dataset.table`
- Filter out test data where country = 'TEST'
- Return SQL only, no explanation.

With this setup, every subsequent prompt from marketers will yield more reliable, production-ready queries for segmentation.

Use Gemini to Generate and Refine RFM-Style Segments

A pragmatic first use case is to have Gemini build RFM (Recency, Frequency, Monetary)-based clusters that go beyond simple thresholds. Ask Gemini to compute RFM scores, propose segment definitions, and write the SQL. Marketers can then iterate on these definitions using natural language instead of hand-coding every change.

Prompt to Gemini:
Write a BigQuery SQL query that:
1) Calculates for each user_id in `project.dataset.events` over the last 365 days:
   - recency_days (days since last revenue > 0 event)
   - frequency (count of revenue > 0 events)
   - monetary (sum of revenue)
2) Assigns R, F, M scores from 1 to 5 using quintiles (5 = best).
3) Creates a segment_label with values like 'champions', 'loyal', 'at_risk', 'new_high_potential'
   based on R, F, M combinations.

Once the base query works, use Gemini to adjust segment thresholds, add channel-specific logic (e.g. champions by paid search vs. organic), or create separate segments for different product lines. This quickly gives you a robust segmentation backbone that can be mapped to Google Ads or CRM audiences.

Identify High-Value Cohorts by Intent and Channel Mix

Go beyond pure purchase history by having Gemini surface high-intent, high-value cohorts based on behavior patterns and channel mix. For example, you might want users who engaged with a specific category, responded well to email but not to display, and showed recent intent signals.

Prompt to Gemini:
Based on the schema context, write SQL that:
1) Identifies users who in the last 30 days:
   - Viewed product pages in category 'premium'
   - Had at least 2 sessions from email and 0 sessions from paid display
   - Did not purchase yet
2) Calculates predicted value proxy:
   - Number of premium product views
   - Total time on site (if available)
3) Outputs a table `project.dataset.premium_intent_email_fans` with user_id and the above metrics.

With this in place, you can sync the resulting table as a custom audience to Google Ads and run tailored creative for these cohorts, such as premium-focused upsell campaigns or email-to-search retargeting flows.

Automate Segment-to-Campaign Mapping with Structured Prompts

Once you have AI-generated segments in BigQuery, use Gemini to help design personalized campaign strategies for each cluster. Provide segment profiles (size, value, key behaviors) and ask Gemini to propose channel mix, messaging angles, offers, and basic measurement plans.

Prompt to Gemini:
You are a performance marketing strategist.
Given the following segment description, propose a campaign strategy.

Segment: at_risk_high_value
- Size: 12,300 users
- Last purchase: 90–180 days ago
- Historically high AOV and margin
- Primary channels: paid search, email
- High engagement with how-to content

Please outline:
1) Core value proposition
2) Recommended channels and budget split
3) 3–4 message angles
4) Offer strategy (discounts, bundles, content)
5) Primary KPI and expected benchmarks.

This doesn’t replace human judgment, but it speeds up planning and ensures every new segment quickly gets an appropriate activation plan instead of sitting unused in a data table.

Embed Gemini into a Repeatable A/B Testing Loop

To avoid one-off experiments, build a loop where Gemini helps design, monitor, and interpret A/B tests for different segments. First, standardize how you log experiments (e.g. an experiments table in BigQuery with variant, segment, KPI metrics). Then use Gemini to generate queries that evaluate performance by segment and suggest next steps.

Prompt to Gemini:
We run experiments logged in `project.dataset.experiments` with:
- experiment_id, segment_label, variant, impressions, clicks, conversions, revenue

Write SQL to:
1) Compare conversion rate and revenue per user for variant A vs B
   within each segment_label.
2) Flag segments where variant B outperforms A with at least 95% confidence.
3) Return a summary table sorted by highest uplift in revenue per user.

Use the results in a follow-up prompt to Gemini to get a natural-language summary for marketers and recommendations on where to roll out winners or design follow-up tests.

Operationalize Segments into Google Ads and CRM Workflows

Finally, turn your Gemini-generated segments into operational audiences. Use scheduled queries in BigQuery to refresh segment tables daily or weekly, and connect them to Google Ads, DV360, or your CRM/CDP using built-in connectors or lightweight pipelines.

Define clear naming conventions (e.g. seg_rfm_champions_search_high, seg_intent_premium_email_loyal) and document which campaigns each segment is mapped to. Gemini can help you generate and maintain this mapping as a structured spec that both marketing and data teams understand.

Expected outcome: Teams that follow these practices typically see more focused spend (10–20% budget reallocated from low- to high-value segments), faster time from segmentation idea to live campaign (from weeks to days), and measurable lifts in conversion or ROAS once the first wave of AI-driven segments is deployed and iterated.

Need implementation expertise now?

Let's talk about your ideas!

Frequently Asked Questions

Gemini improves segmentation by analyzing full behavioral and transactional histories rather than a few static rules. Instead of “age 25–34, visited product page twice”, Gemini can help you build segments like “users who discovered us via paid search, engaged heavily with category X content, and show early signs of becoming high-LTV based on RFM and channel mix.”

Practically, Gemini generates and refines the BigQuery SQL that computes these patterns at scale, allowing marketers to explore new segments quickly without waiting for a data engineer each time. The result is more granular, behavior-driven clusters that map more directly to how users actually buy and engage.

You need three main ingredients: access to your marketing data in BigQuery, someone who understands your data model (data engineer or analyst), and marketers who are willing to work with Gemini as a co-analyst. You do not need a large data science team to get started, but you do need at least basic data engineering to ensure identity stitching and event tracking are reliable.

From a skills perspective, marketers should be comfortable formulating clear business questions and reading high-level query outputs, while technical staff set up the initial Gemini integration, security, and prompt templates. Reruption often helps teams bridge this gap in the first phase, then gradually hands over more responsibility to marketing as they get comfortable.

If your BigQuery data is already in decent shape, you can see first useful segments in a matter of days to a few weeks. A typical timeline looks like this: 1–2 weeks to align on goals, validate data, and set up Gemini with your schema; another 1–2 weeks to generate initial segments, connect them to Google Ads or CRM, and launch pilot campaigns.

Meaningful performance improvements (e.g. better ROAS, higher conversion for key cohorts) usually appear after 4–8 weeks, once you’ve iterated on the first set of segments and optimized campaigns. The important thing is to start small with clear KPIs and then expand as you see what works.

The direct technology cost of using Gemini with BigQuery is typically modest compared to media budgets: you pay for data storage/processing and Gemini API usage. The larger investment is in initial setup and change management: aligning data, configuring the integration, and adjusting how the marketing team works with segments.

In ROI terms, we encourage teams to frame success around specific levers: for example, reallocating 10–20% of spend towards high-LTV segments, improving conversion rates for at-risk but high-potential cohorts, or reducing time spent on manual segmentation by 50%+. Even conservative gains in these areas usually pay back the investment quickly, especially in paid media-heavy environments.

Reruption works as a Co-Preneur, embedding with your team to build real AI capabilities rather than just slideware. We typically start with a focused AI PoC (9.900€) where we validate that Gemini can meaningfully improve your segmentation on your actual data: define the use case, connect Gemini to BigQuery, generate first segments, and test them in a small campaign setup.

From there, we can support end-to-end implementation: data preparation, Gemini prompt and SQL templates, integration with Google Ads/CRM, and enablement for your marketing team. Our goal is to leave you with a working, AI-first segmentation engine and a team that knows how to use it, not a theoretical concept. If you want to explore whether this fits your situation, we can scope a PoC around one concrete segmentation problem and measure its impact together.

Contact Us!

0/10 min.

Contact Directly

Your Contact

Philipp M. W. Hoffmann

Founder & Partner

Address

Reruption GmbH

Falkertstraße 2

70176 Stuttgart

Social Media