Chatbot Scenario A/B Testing in Mobile App

NOVASOLUTIONS.TECHNOLOGY is engaged in the development, support and maintenance of iOS, Android, PWA mobile applications. We have extensive experience and expertise in publishing mobile applications in popular markets like Google Play, App Store, Amazon, AppGallery and others.
Development and support of all types of mobile applications:
Information and entertainment mobile applications
News apps, games, reference guides, online catalogs, weather apps, fitness and health apps, travel apps, educational apps, social networks and messengers, quizzes, blogs and podcasts, forums, aggregators
E-commerce mobile applications
Online stores, B2B apps, marketplaces, online exchanges, cashback services, exchanges, dropshipping platforms, loyalty programs, food and goods delivery, payment systems.
Business process management mobile applications
CRM systems, ERP systems, project management, sales team tools, financial management, production management, logistics and delivery management, HR management, data monitoring systems
Electronic services mobile applications
Classified ads platforms, online schools, online cinemas, electronic service platforms, cashback platforms, video hosting, thematic portals, online booking and scheduling platforms, online trading platforms

These are just some of the types of mobile applications we work with, and each of them may have its own specific features and functionality, tailored to the specific needs and goals of the client.

Showing 1 of 1 servicesAll 1735 services
Chatbot Scenario A/B Testing in Mobile App
Medium
~3-5 business days
FAQ
Our competencies:
Development stages
Latest works
  • image_mobile-applications_feedme_467_0.webp
    Development of a mobile application for FEEDME
    756
  • image_mobile-applications_xoomer_471_0.webp
    Development of a mobile application for XOOMER
    624
  • image_mobile-applications_rhl_428_0.webp
    Development of a mobile application for RHL
    1050
  • image_mobile-applications_zippy_411_0.webp
    Development of a mobile application for ZIPPY
    947
  • image_mobile-applications_affhome_429_0.webp
    Development of a mobile application for Affhome
    862
  • image_mobile-applications_flavors_409_0.webp
    Development of a mobile application for the FLAVORS company
    445

Implementing A/B Testing for Chatbot Scenarios in Mobile Apps

Product wants to test: which greeting variant converts better—"Hi, how can I help?" or "I'll show you products for your request right away." A/B test on UI level—straightforward task. But a bot isn't just text: it's a dialog graph, set of intents, escalation logic to operators. Bot scenario A/B testing requires separate infrastructure.

What We Test in a Bot

Bot scenarios differ from UI elements: a variant isn't a button color, but an entire dialog graph. User may take 7 steps in variant A and 3 in variant B to achieve one result. Metric—not a click, but target action completion (purchase, lead, resolved question). This complicates measurement and requires event tracking at every dialog step.

Typical A/B hypotheses for bots:

  • Different greetings and tone of voice
  • Quick replies vs text input on first step
  • Escalation timing (immediately vs after 2 failed intents)
  • Different CTA formulations within dialog

Technical Implementation

Firebase Remote Config—standard choice for mobile A/B tests. Bot configuration parameters (scenario ID, prompt version, escalation threshold) read on app startup:

let remoteConfig = RemoteConfig.remoteConfig()
remoteConfig.fetch(withExpirationDuration: 3600) { [weak self] status, error in
    guard status == .success else { return }
    remoteConfig.activate { _, _ in
        let botVariant = remoteConfig["bot_scenario_variant"].stringValue ?? "control"
        self?.chatViewModel.loadScenario(variant: botVariant)
    }
}

Firebase automatically splits audience into groups, supports conditions (country, app version, user properties). Built-in analytics via Firebase Analytics—conversion events marked with standard logEvent.

Growthbook / Statsig—alternatives with more powerful statistical model. Growthbook open-source, can self-host. Statsig has good iOS/Android SDK with low latency (feature flags cached locally).

Server-side A/B vs client-side. If bot implemented via server dialog engine (Rasa, Dialogflow CX, custom), better manage variant server-side—client sends userId + sessionId, server selects scenario by experimental group and returns responses for needed variant. Prevents cheating and simplifies analytics.

Dialog Event Tracking

Without detailed step tracking, impossible to understand where users drop. Minimum event set:

  • bot_session_start — {variant, userId, sessionId}
  • bot_message_sent — {variant, stepId, messageType}
  • bot_message_received — {variant, stepId, intentId, confidence}
  • bot_intent_failed — {variant, stepId, userInput}—when NLU didn't recognize intent
  • bot_escalated — {variant, stepId, reason}
  • bot_goal_completed — {variant, goalType}—conversion event

All events with variant and sessionId—allows reconstructing full user path in any variant.

Statistical Significance

Common error—stop test on first promising numbers. Need minimum sample size calculated beforehand (power analysis): at desired 5% effect, 15% baseline conversion, 80% test power—need minimum ~2800 users per group. Firebase A/B Testing calculates automatically.

Timeline Estimates

Implementing A/B testing for two scenario variants with Firebase Remote Config and event tracking—3–5 days. Integration with server dialog engine and more complex audience split logic—up to 2 weeks.