The Challenge: Inefficient Audience Segmentation

Most marketing teams still rely on blunt tools for audience segmentation: a few demographic filters, basic interests, maybe one or two lifecycle rules. It feels structured, but in reality it’s little more than educated guesswork. As data volume and channel complexity explode, this approach simply cannot keep up with the way customers actually behave across touchpoints.

Traditional segmentation is typically built once per quarter in a spreadsheet, then pushed into the ad platforms and CRM. It ignores micro-behaviors, evolving intent, and cross-channel signals. Static rules like “visited product page twice” or “opened last 3 emails” miss deeper patterns in frequency, timing, product mix, and engagement decay. Even when data teams are involved, the process of writing custom SQL or models for each new hypothesis is slow, expensive, and frustrating for marketers who need agility.

The business impact is substantial. Blunt targeting drives up CAC as low-value users see too many ads, while high-potential cohorts never get the differentiated offers, creatives, or journeys they deserve. Budget gets spread evenly instead of being focused on high-LTV clusters. Personalization efforts underperform, experimentation slows down, and competitors with more sophisticated segmentation quietly win the most valuable segments in your market. Over time, the gap in ROI and marketing efficiency compounds.

This challenge is real, but it’s solvable. With the right use of AI-driven segmentation, you can uncover patterns that no human analyst would ever think to test, and operationalize them directly into your campaigns. At Reruption, we’ve seen how combining strong engineering with a practical marketing lens turns unusable data into actionable audiences. Below, you’ll find concrete guidance on how to use Gemini with your Google data stack to move beyond static rules and build segmentation that finally matches how your customers behave.

Need a sparring partner for this challenge?

Let's have a no-obligation chat and brainstorm together.

Innovators at these companies trust us:

Our Assessment

A strategic assessment of the challenge and high-level tips how to tackle it.

From Reruption’s perspective, the real opportunity isn’t just using Gemini as another content generator, but as an intelligence layer on top of your BigQuery and Google Ads data. With hands-on experience building AI solutions inside marketing teams, we’ve seen that the combination of Gemini + your existing data warehouse can transform inefficient, rule-based segmentation into dynamic, behavior-driven clusters that marketers can actually use without relying on a data team for every change.

Start from Business Value, Not from Algorithms

When using Gemini for audience segmentation, the starting point shouldn’t be clustering techniques or feature engineering – it should be marketing and commercial goals. Before you touch data, define which outcomes matter: lower CAC for performance campaigns, higher LTV in retention programs, or improved conversion rates on key funnels. This gives Gemini clear “north stars” when it helps you derive segments and write SQL queries.

Translate these goals into concrete questions: “Which behavioral patterns predict repeat purchase within 60 days?” or “Which users show intent but stall before adding to cart?” When we embed with teams, we force this discipline early. It keeps the AI work focused on value-creating clusters, not just mathematically interesting ones that never make it into campaigns.

Treat Gemini as a Data Co-Analyst for Marketers

Strategically, the power of Gemini + BigQuery is to reduce dependency on scarce data engineering resources. Marketers should be able to collaborate with Gemini like a co-analyst: describe the audience they care about in plain language and have Gemini draft the SQL or segmentation logic against your schema. That changes the team’s operating model from ticket-driven to exploratory and hypothesis-driven.

This requires an intentional shift in mindset. Instead of asking “What segments can the data team give us this quarter?”, marketers start asking “What segments do we want to test this week?” and use Gemini to iterate. You still need guardrails and review, but the strategic trajectory is towards self-service, not more centralization.

Invest in Data Readiness Before Scaling Personalization

No matter how good Gemini is, it can’t segment on signals you don’t collect or can’t trust. Strategically, you need a minimum level of data quality and structure in BigQuery: consistent user IDs or stitched identities, clean event tracking, and clear mapping between marketing events and business outcomes (purchases, leads, churn, etc.).

Instead of trying to fix everything, focus on the 3–5 most critical signals for segmentation: recency/frequency/monetary value, key intent events (e.g. product views, trials started), and engagement across main channels. Our experience shows that once these are reliable, Gemini can do meaningful work on top, and the team builds trust in the AI-generated segments.

Design Governance and Validation Around AI-Generated Segments

Using AI-generated audience clusters without governance is risky. Strategically, define how segments move from “experimental” to “production”. For example: initial clusters are tested on 5–10% of budget with clear success criteria; only after hitting predefined thresholds (conversion rate, ROAS, engagement) do they become standard audiences.

Make sure someone owns validation. A marketer and an analyst (or Gemini plus a human reviewer) should regularly sanity-check segment definitions, size, and overlap. This governance layer is where you mitigate the risk of overfitting, biased segments, or simply segments that are not actionable at scale.

Prepare Your Team for an AI-First Way of Working

Gemini will change how your marketing team works with data. Strategically, that means upskilling marketers to be comfortable with concepts like schemas, events, and simple SQL prompts – not to turn them into engineers, but to make them effective AI users. It also means aligning with data and IT on roles and responsibilities so Gemini doesn’t become a shadow-IT project.

We’ve seen that teams that embrace an AI-first lens – asking “If we rebuilt segmentation with AI from scratch, how would it work?” – move much faster than those trying to bolt Gemini onto legacy processes. Align expectations early: Gemini will not be perfect from day one, but it will drastically shorten the path from idea to live segment if the team is ready to iterate with it.

Used thoughtfully, Gemini can turn inefficient, rule-based segmentation into a living system of behavior-driven clusters that marketers can explore and deploy in days, not quarters. The key is to combine business-focused goals, data readiness, and clear governance so that AI-generated segments are both powerful and safe to use. At Reruption, we’ve built similar AI-first capabilities inside organisations and can help you scope, prototype, and operationalize a Gemini-powered segmentation engine that fits your stack and team. If you’re considering this step, a short conversation and a focused AI PoC can quickly show what’s realistically achievable in your environment.

Need help implementing these ideas?

Feel free to reach out to us with no obligation.

Real-World Case Studies

From Banking to Fintech: Learn how companies successfully use Gemini.

bunq

Banking

As bunq experienced rapid growth as the second-largest neobank in Europe, scaling customer support became a critical challenge. With millions of users demanding personalized banking information on accounts, spending patterns, and financial advice on demand, the company faced pressure to deliver instant responses without proportionally expanding its human support teams, which would increase costs and slow operations. Traditional search functions in the app were insufficient for complex, contextual queries, leading to inefficiencies and user frustration. Additionally, ensuring data privacy and accuracy in a highly regulated fintech environment posed risks. bunq needed a solution that could handle nuanced conversations while complying with EU banking regulations, avoiding hallucinations common in early GenAI models, and integrating seamlessly without disrupting app performance. The goal was to offload routine inquiries, allowing human agents to focus on high-value issues.

Lösung

bunq addressed these challenges by developing Finn, a proprietary GenAI platform integrated directly into its mobile app, replacing the traditional search function with a conversational AI chatbot. After hiring over a dozen data specialists in the prior year, the team built Finn to query user-specific financial data securely, answer questions on balances, transactions, budgets, and even provide general advice while remembering conversation context across sessions. Launched as Europe's first AI-powered bank assistant in December 2023 following a beta, Finn evolved rapidly. By May 2024, it became fully conversational, enabling natural back-and-forth interactions. This retrieval-augmented generation (RAG) approach grounded responses in real-time user data, minimizing errors and enhancing personalization.

Ergebnisse

  • 100,000+ questions answered within months post-beta (end-2023)
  • 40% of user queries fully resolved autonomously by mid-2024
  • 35% of queries assisted, totaling 75% immediate support coverage
  • Hired 12+ data specialists pre-launch for data infrastructure
  • Second-largest neobank in Europe by user base (1M+ users)
Read case study →

Upstart

Banking

Traditional credit scoring relies heavily on FICO scores, which evaluate only a narrow set of factors like payment history and debt utilization, often rejecting creditworthy borrowers with thin credit files, non-traditional employment, or education histories that signal repayment ability. This results in up to 50% of potential applicants being denied despite low default risk, limiting lenders' ability to expand portfolios safely . Fintech lenders and banks faced the dual challenge of regulatory compliance under fair lending laws while seeking growth. Legacy models struggled with inaccurate risk prediction amid economic shifts, leading to higher defaults or conservative lending that missed opportunities in underserved markets . Upstart recognized that incorporating alternative data could unlock lending to millions previously excluded.

Lösung

Upstart developed an AI-powered lending platform using machine learning models that analyze over 1,600 variables, including education, job history, and bank transaction data, far beyond FICO's 20-30 inputs. Their gradient boosting algorithms predict default probability with higher precision, enabling safer approvals . The platform integrates via API with partner banks and credit unions, providing real-time decisions and fully automated underwriting for most loans. This shift from rule-based to data-driven scoring ensures fairness through explainable AI techniques like feature importance analysis . Implementation involved training models on billions of repayment events, continuously retraining to adapt to new data patterns .

Ergebnisse

  • 44% more loans approved vs. traditional models
  • 36% lower average interest rates for borrowers
  • 80% of loans fully automated
  • 73% fewer losses at equivalent approval rates
  • Adopted by 500+ banks and credit unions by 2024
  • 157% increase in approvals at same risk level
Read case study →

DBS Bank

Banking

DBS Bank, Southeast Asia's leading financial institution, grappled with scaling AI from experiments to production amid surging fraud threats, demands for hyper-personalized customer experiences, and operational inefficiencies in service support. Traditional fraud detection systems struggled to process up to 15,000 data points per customer in real-time, leading to missed threats and suboptimal risk scoring. Personalization efforts were hampered by siloed data and lack of scalable algorithms for millions of users across diverse markets. Additionally, customer service teams faced overwhelming query volumes, with manual processes slowing response times and increasing costs. Regulatory pressures in banking demanded responsible AI governance, while talent shortages and integration challenges hindered enterprise-wide adoption. DBS needed a robust framework to overcome data quality issues, model drift, and ethical concerns in generative AI deployment, ensuring trust and compliance in a competitive Southeast Asian landscape.

Lösung

DBS launched an enterprise-wide AI program with over 20 use cases, leveraging machine learning for advanced fraud risk models and personalization, complemented by generative AI for an internal support assistant. Fraud models integrated vast datasets for real-time anomaly detection, while personalization algorithms delivered hyper-targeted nudges and investment ideas via the digibank app. A human-AI synergy approach empowered service teams with a GenAI assistant handling routine queries, drawing from internal knowledge bases. DBS emphasized responsible AI through governance frameworks, upskilling 40,000+ employees, and phased rollout starting with pilots in 2021, scaling production by 2024. Partnerships with tech leaders and Harvard-backed strategy ensured ethical scaling across fraud, personalization, and operations.

Ergebnisse

  • 17% increase in savings from prevented fraud attempts
  • Over 100 customized algorithms for customer analyses
  • 250,000 monthly queries processed efficiently by GenAI assistant
  • 20+ enterprise-wide AI use cases deployed
  • Analyzes up to 15,000 data points per customer for fraud
  • Boosted productivity by 20% via AI adoption (CEO statement)
Read case study →

Pfizer

Healthcare

The COVID-19 pandemic created an unprecedented urgent need for new antiviral treatments, as traditional drug discovery timelines span 10-15 years with success rates below 10%. Pfizer faced immense pressure to identify potent, oral inhibitors targeting the SARS-CoV-2 3CL protease (Mpro), a key viral enzyme, while ensuring safety and efficacy in humans. Structure-based drug design (SBDD) required analyzing complex protein structures and generating millions of potential molecules, but conventional computational methods were too slow, consuming vast resources and time. Challenges included limited structural data early in the pandemic, high failure risks in hit identification, and the need to run processes in parallel amid global uncertainty. Pfizer's teams had to overcome data scarcity, integrate disparate datasets, and scale simulations without compromising accuracy, all while traditional wet-lab validation lagged behind.

Lösung

Pfizer deployed AI-driven pipelines leveraging machine learning (ML) for SBDD, using models to predict protein-ligand interactions and generate novel molecules via generative AI. Tools analyzed cryo-EM and X-ray structures of the SARS-CoV-2 protease, enabling virtual screening of billions of compounds and de novo design optimized for binding affinity, pharmacokinetics, and synthesizability. By integrating supercomputing with ML algorithms, Pfizer streamlined hit-to-lead optimization, running parallel simulations that identified PF-07321332 (nirmatrelvir) as the lead candidate. This lightspeed approach combined ML with human expertise, reducing iterative cycles and accelerating from target validation to preclinical nomination.

Ergebnisse

  • Drug candidate nomination: 4 months vs. typical 2-5 years
  • Computational chemistry processes reduced: 80-90%
  • Drug discovery timeline cut: From years to 30 days for key phases
  • Clinical trial success rate boost: Up to 12% (vs. industry ~5-10%)
  • Virtual screening scale: Billions of compounds screened rapidly
  • Paxlovid efficacy: 89% reduction in hospitalization/death
Read case study →

Cruise (GM)

Automotive

Developing a self-driving taxi service in dense urban environments posed immense challenges for Cruise. Complex scenarios like unpredictable pedestrians, erratic cyclists, construction zones, and adverse weather demanded near-perfect perception and decision-making in real-time. Safety was paramount, as any failure could result in accidents, regulatory scrutiny, or public backlash. Early testing revealed gaps in handling edge cases, such as emergency vehicles or occluded objects, requiring robust AI to exceed human driver performance. A pivotal safety incident in October 2023 amplified these issues: a Cruise vehicle struck a pedestrian pushed into its path by a hit-and-run driver, then dragged her while fleeing the scene, leading to suspension of operations nationwide. This exposed vulnerabilities in post-collision behavior, sensor fusion under chaos, and regulatory compliance. Scaling to commercial robotaxi fleets while achieving zero at-fault incidents proved elusive amid $10B+ investments from GM.

Lösung

Cruise addressed these with an integrated AI stack leveraging computer vision for perception and reinforcement learning for planning. Lidar, radar, and 30+ cameras fed into CNNs and transformers for object detection, semantic segmentation, and scene prediction, processing 360° views at high fidelity even in low light or rain. Reinforcement learning optimized trajectory planning and behavioral decisions, trained on millions of simulated miles to handle rare events. End-to-end neural networks refined motion forecasting, while simulation frameworks accelerated iteration without real-world risk. Post-incident, Cruise enhanced safety protocols, resuming supervised testing in 2024 with improved disengagement rates. GM's pivot integrated this tech into Super Cruise evolution for personal vehicles.

Ergebnisse

  • 1,000,000+ miles driven fully autonomously by 2023
  • 5 million driverless miles used for AI model training
  • $10B+ cumulative investment by GM in Cruise (2016-2024)
  • 30,000+ miles per intervention in early unsupervised tests
  • Operations suspended Oct 2023; resumed supervised May 2024
  • Zero commercial robotaxi revenue; pivoted Dec 2024
Read case study →

Best Practices

Successful implementations follow proven patterns. Have a look at our tactical advice to get started.

Connect Gemini to Your BigQuery Marketing Dataset with Clear Context

To get useful AI-driven segmentation, Gemini needs more than raw data – it needs context. Start by ensuring your key marketing tables (events, sessions, transactions, campaigns) are in BigQuery with consistent user identifiers. Then document the schema in plain language: what each table represents, primary keys, and how events map to key funnel steps.

When you interact with Gemini (e.g. via Vertex AI, the Gemini API, or a connected notebook environment), always include this context at the start of the conversation. Provide table names, column descriptions, and example queries that follow your internal conventions. This dramatically improves the quality and safety of the SQL Gemini will generate.

System / setup prompt for Gemini:
You are a senior marketing data analyst.
You write safe, BigQuery-compatible SQL for segmentation.

Schema context:
- Table: events (user_id, event_name, event_time, source, campaign_id, revenue)
- Table: users (user_id, signup_date, country, device_type)
- Table: campaigns (campaign_id, channel, objective, cost)

Always:
- Use fully qualified table names: `project.dataset.table`
- Filter out test data where country = 'TEST'
- Return SQL only, no explanation.

With this setup, every subsequent prompt from marketers will yield more reliable, production-ready queries for segmentation.

Use Gemini to Generate and Refine RFM-Style Segments

A pragmatic first use case is to have Gemini build RFM (Recency, Frequency, Monetary)-based clusters that go beyond simple thresholds. Ask Gemini to compute RFM scores, propose segment definitions, and write the SQL. Marketers can then iterate on these definitions using natural language instead of hand-coding every change.

Prompt to Gemini:
Write a BigQuery SQL query that:
1) Calculates for each user_id in `project.dataset.events` over the last 365 days:
   - recency_days (days since last revenue > 0 event)
   - frequency (count of revenue > 0 events)
   - monetary (sum of revenue)
2) Assigns R, F, M scores from 1 to 5 using quintiles (5 = best).
3) Creates a segment_label with values like 'champions', 'loyal', 'at_risk', 'new_high_potential'
   based on R, F, M combinations.

Once the base query works, use Gemini to adjust segment thresholds, add channel-specific logic (e.g. champions by paid search vs. organic), or create separate segments for different product lines. This quickly gives you a robust segmentation backbone that can be mapped to Google Ads or CRM audiences.

Identify High-Value Cohorts by Intent and Channel Mix

Go beyond pure purchase history by having Gemini surface high-intent, high-value cohorts based on behavior patterns and channel mix. For example, you might want users who engaged with a specific category, responded well to email but not to display, and showed recent intent signals.

Prompt to Gemini:
Based on the schema context, write SQL that:
1) Identifies users who in the last 30 days:
   - Viewed product pages in category 'premium'
   - Had at least 2 sessions from email and 0 sessions from paid display
   - Did not purchase yet
2) Calculates predicted value proxy:
   - Number of premium product views
   - Total time on site (if available)
3) Outputs a table `project.dataset.premium_intent_email_fans` with user_id and the above metrics.

With this in place, you can sync the resulting table as a custom audience to Google Ads and run tailored creative for these cohorts, such as premium-focused upsell campaigns or email-to-search retargeting flows.

Automate Segment-to-Campaign Mapping with Structured Prompts

Once you have AI-generated segments in BigQuery, use Gemini to help design personalized campaign strategies for each cluster. Provide segment profiles (size, value, key behaviors) and ask Gemini to propose channel mix, messaging angles, offers, and basic measurement plans.

Prompt to Gemini:
You are a performance marketing strategist.
Given the following segment description, propose a campaign strategy.

Segment: at_risk_high_value
- Size: 12,300 users
- Last purchase: 90–180 days ago
- Historically high AOV and margin
- Primary channels: paid search, email
- High engagement with how-to content

Please outline:
1) Core value proposition
2) Recommended channels and budget split
3) 3–4 message angles
4) Offer strategy (discounts, bundles, content)
5) Primary KPI and expected benchmarks.

This doesn’t replace human judgment, but it speeds up planning and ensures every new segment quickly gets an appropriate activation plan instead of sitting unused in a data table.

Embed Gemini into a Repeatable A/B Testing Loop

To avoid one-off experiments, build a loop where Gemini helps design, monitor, and interpret A/B tests for different segments. First, standardize how you log experiments (e.g. an experiments table in BigQuery with variant, segment, KPI metrics). Then use Gemini to generate queries that evaluate performance by segment and suggest next steps.

Prompt to Gemini:
We run experiments logged in `project.dataset.experiments` with:
- experiment_id, segment_label, variant, impressions, clicks, conversions, revenue

Write SQL to:
1) Compare conversion rate and revenue per user for variant A vs B
   within each segment_label.
2) Flag segments where variant B outperforms A with at least 95% confidence.
3) Return a summary table sorted by highest uplift in revenue per user.

Use the results in a follow-up prompt to Gemini to get a natural-language summary for marketers and recommendations on where to roll out winners or design follow-up tests.

Operationalize Segments into Google Ads and CRM Workflows

Finally, turn your Gemini-generated segments into operational audiences. Use scheduled queries in BigQuery to refresh segment tables daily or weekly, and connect them to Google Ads, DV360, or your CRM/CDP using built-in connectors or lightweight pipelines.

Define clear naming conventions (e.g. seg_rfm_champions_search_high, seg_intent_premium_email_loyal) and document which campaigns each segment is mapped to. Gemini can help you generate and maintain this mapping as a structured spec that both marketing and data teams understand.

Expected outcome: Teams that follow these practices typically see more focused spend (10–20% budget reallocated from low- to high-value segments), faster time from segmentation idea to live campaign (from weeks to days), and measurable lifts in conversion or ROAS once the first wave of AI-driven segments is deployed and iterated.

Need implementation expertise now?

Let's talk about your ideas!

Frequently Asked Questions

Gemini improves segmentation by analyzing full behavioral and transactional histories rather than a few static rules. Instead of “age 25–34, visited product page twice”, Gemini can help you build segments like “users who discovered us via paid search, engaged heavily with category X content, and show early signs of becoming high-LTV based on RFM and channel mix.”

Practically, Gemini generates and refines the BigQuery SQL that computes these patterns at scale, allowing marketers to explore new segments quickly without waiting for a data engineer each time. The result is more granular, behavior-driven clusters that map more directly to how users actually buy and engage.

You need three main ingredients: access to your marketing data in BigQuery, someone who understands your data model (data engineer or analyst), and marketers who are willing to work with Gemini as a co-analyst. You do not need a large data science team to get started, but you do need at least basic data engineering to ensure identity stitching and event tracking are reliable.

From a skills perspective, marketers should be comfortable formulating clear business questions and reading high-level query outputs, while technical staff set up the initial Gemini integration, security, and prompt templates. Reruption often helps teams bridge this gap in the first phase, then gradually hands over more responsibility to marketing as they get comfortable.

If your BigQuery data is already in decent shape, you can see first useful segments in a matter of days to a few weeks. A typical timeline looks like this: 1–2 weeks to align on goals, validate data, and set up Gemini with your schema; another 1–2 weeks to generate initial segments, connect them to Google Ads or CRM, and launch pilot campaigns.

Meaningful performance improvements (e.g. better ROAS, higher conversion for key cohorts) usually appear after 4–8 weeks, once you’ve iterated on the first set of segments and optimized campaigns. The important thing is to start small with clear KPIs and then expand as you see what works.

The direct technology cost of using Gemini with BigQuery is typically modest compared to media budgets: you pay for data storage/processing and Gemini API usage. The larger investment is in initial setup and change management: aligning data, configuring the integration, and adjusting how the marketing team works with segments.

In ROI terms, we encourage teams to frame success around specific levers: for example, reallocating 10–20% of spend towards high-LTV segments, improving conversion rates for at-risk but high-potential cohorts, or reducing time spent on manual segmentation by 50%+. Even conservative gains in these areas usually pay back the investment quickly, especially in paid media-heavy environments.

Reruption works as a Co-Preneur, embedding with your team to build real AI capabilities rather than just slideware. We typically start with a focused AI PoC (9.900€) where we validate that Gemini can meaningfully improve your segmentation on your actual data: define the use case, connect Gemini to BigQuery, generate first segments, and test them in a small campaign setup.

From there, we can support end-to-end implementation: data preparation, Gemini prompt and SQL templates, integration with Google Ads/CRM, and enablement for your marketing team. Our goal is to leave you with a working, AI-first segmentation engine and a team that knows how to use it, not a theoretical concept. If you want to explore whether this fits your situation, we can scope a PoC around one concrete segmentation problem and measure its impact together.

Contact Us!

0/10 min.

Contact Directly

Your Contact

Philipp M. W. Hoffmann

Founder & Partner

Address

Reruption GmbH

Falkertstraße 2

70176 Stuttgart

Social Media