Inspiration

Traditional A/B testing is slow, expensive, and often requires weeks of user recruitment. As a solo developer passionate about AI and growth hacking, I wanted to democratize experimentation for startups and marketers. What if teams could simulate entire campaigns, personas, and results before investing in real users? Generative AI made this possible. Inspired by tools like ChatGPT and the rise of synthetic data, I built a platform to turn product ideas into actionable insights in minutes, not months.

What it does

Gemini AI A/B Testing lets you:

  • πŸ§ͺ Design experiments: Turn a product description into A/B test hypotheses (e.g., email campaigns).
  • πŸ€– Simulate users: Generate personas (age, job, preferences) mirroring your target audience.
  • πŸ“Š Analyze AI-driven responses: Predict CTR, engagement, and conversion likelihood.
  • πŸ“„ Export reports: Get statistical summaries (p-values, confidence intervals) and recommendations.

No real users, or waiting required.

How we built it

  • Frontend: Streamlit for a simple, no-frills UI.
  • Backend: Python orchestrating Gemini API for text generation (emails, personas) and Agno agents for logic.
  • Workflow Automation: Prefect pipelines to chain tasks (generate emails β†’ simulate personas β†’ collect responses).
  • Data Handling: Pandas for CSV exports; custom prompts to ensure Gemini outputs structured data.
  • Evaluation: Statistical tests (chi-square, t-tests) to determine winning variants.

Challenges we ran into

  • Structured Outputs from Gemini: Getting consistent CSV-ready data from free-text AI responses.
  • Fix: Iterated on 50+ prompt templates with fallback regex validation.
  • Realistic Persona Simulation: Avoiding generic personas (e.g., "John, 30, likes tech").
  • Fix: Added granular traits (e.g., "Environmental activist, distrusts greenwashing").
  • Workflow Reliability: Prefect tasks failing mid-pipeline due to API timeouts.
  • Fix: Added retries and state checkpoints.

Accomplishments that we're proud of

  • πŸš€ Reduced A/B testing setup from days to 5 minutes.
  • πŸ”₯ Achieved 89% accuracy in simulated vs. real-user responses (validated with beta testers).
  • πŸ› οΈ Built an end-to-end platform solo, from ideation to deployment.
  • πŸ’‘ Featured on AI tool directories like Agno (Phidata), prefect!

What we learned

  • Generative AI’s Limits: Hallucinations require guardrails (e.g., forcing Gemini to pick from predefined categories).
  • Synthetic Data’s Power: Simulated users can uncover biases (e.g., younger personas preferring trendy messaging).
  • Prefect’s Strengths: Task orchestration is game-changing for reproducibility.

What's next for Gemini AI AB Testing, Simulate Campaigns, Personas, Reports

  • Multi-Channel Testing: Expand beyond emails to landing pages, ads, and push notifications.
  • Dynamic Personas: Let users upload customer data to fine-tune AI personas.
  • Collaboration Mode: Teams working on the same experiment with role-based access.
  • Live User Testing: Hybrid mode (AI + real users) for calibration.
  • Cloud Hosting: One-click deploy on AWS/Azure for enterprises.

Built With

Share this project:

Updates