How is AI really influencing experimentation?
Has your C-suite started asking how you’ll use AI to improve experimentation? The fear of missing out on this trend is real. But rather than a flash in the pan, AI is changing how teams experiment, offering real advantages to those who implement it.
Understanding how Generative AI and Machine Learning can be used and what they are good at (and not) will help you make the most of this technology. For example, Large Language Models (LLM) aren’t good at answering discrete problems; ask them to order a list of names alphabetically, and they might make mistakes. However, LLMs are good at exploring broader open-ended issues or acting as a co-pilot on your team.
If facts and math aren’t suitable tasks for AI, what and where should we use them? To better understand this, we asked eight experimentation leaders how AI is changing their experimentation work.
How is AI changing A/B testing?
Less reliance on Devs to code test variants
Development resources are often scarce but necessary to run more complex A/B tests. Rather than letting a lack of resources slow you down, Johannes Mandl feels it’s a perfect task for your AI assistant;
Using AI to code tests can accelerate testing, leading to faster insights. It also frees up technical resources to work on other tasks, such as implementing the winning solutions into the codebase so you can reap the benefits quicker.
Faster user feedback analysis and problem exploration
Good conversion optimization ideas come from a deep understanding of your users. One of the best ways to achieve this is by reviewing qualitative user feedback. But it’s a big ask to review days worth of transcripts or user reviews. As the name suggests, Large Language Models are made for this type of job, as Iqbal Ali explains;
The other area is problem exploration. AI can be a valuable collaborator to user researchers, product managers, and others, helping avoid laziness and human biases in the problem exploration phase (common challenges I’ve observed in teams). AI can especially be effective in a workshop setting, where its contributions can be clearly demonstrated.
There are countless ways to prompt AI, and how you ask a question will influence the quality of the output. To explore problems, for example, you can use frameworks such as the Six Thinking Hats or The 5 Whys to delve into a topic. You can also ask AI to provide a range of problem exploration methods and then use those in follow-up prompts. Plus, we’ve written a guide to help you craft AI prompts.
Easier test anomaly detection
Once a test is live, you want to know if the results are within an acceptable range. It’s not just the A/B test results that must be monitored; other KPIs might indicate a problem. Often, these KPIs passively tick away in a neglected dashboard, silently witnessing something going awry. But this doesn't need to be the case. You can ask your always-on AI sidekick to alert you to anomalies, as Ellie Hughes explains:
In addition to simplifying anomaly detection, AI can help decision-making when an anomaly is found. By analyzing patterns and outliers, your AI analyst can provide deeper insights into the root causes, helping teams identify issues and understand the underlying factors driving them.
Moreover, AI’s ability to continuously learn from data means that its predictive accuracy improves over time, making future experiments more reliable.
Less time spent on project management
Not every testing team is fortunate enough to have dedicated project managers. Despite spending years honing your user behavior knowledge or understanding of statistical methods, a fair chunk of your time will need to be spent on project management.
Project management, while critical, is incredibly time-intensive and not a good use of specialized resources. If this is the case in your team, good news! AI can act as a project manager on your team, freeing you from tedious, busy work and helping you focus on high-value tasks. Eric Itzkowitz shares some examples;
Anjali Arora Mehra shares this sentiment, too, with some specific tasks you can recruit your AI co-pilot to help with;
However, Jonathan Shuster wants to see more evidence and human oversight;
The key to employing AI as your new project management is to stagger the approach. Start using AI for simple and routine processes. As you gain experience and AI continues to evolve, its role in managing complex projects could expand.
More (and wackier) test ideas
There’s no such thing as “the” solution in experimentation. Instead, there are often thousands of potential solutions that will have varying degrees of success. It’s why we test. But often, we’re quick to identify a solution we think will solve a problem without adequately other possibilities.
We might fall into the trap of doing what we’ve seen work before or for competitors or ideas that fit the status quo. However, this thinking falls foul of a number of biases and can hold companies back from moving beyond their local maxim. That’s why Iqbal Ali recommends consulting your AI assistant;
The key here is the volume and diversity of ideas. Try it for yourself: gather your team, present a problem, and ask them to list all the possible solutions. Then, try the same with your AI assistant. AI usually produces many more ideas, some of which might not have occurred to your team.
AI can also give you wackier ideas. While humans don’t like suggesting ‘silly’ ideas for fear of being ridiculed, AI doesn’t care. That’s why it deserves a place on your team.
Help analyzing test impact
Data interpretation is a tricky task, especially if you want to analyze multiple data sources to assess the impact of an experiment and establish a business decision. There’s a lot at stake, but your AI co-pilot can help interrogate data and point out things we might not notice on first inspection, as Eric Itzkowitz says;
Mike St Laurent is already putting AI to work on test analysis;
As of today, the applications that seem the most promising are in test analysis (which runs structured data sets through several statistical rules and pulls insights out) and test development (which can set up test files and write a significant portion of the code for certain types of experiments).
AI's role in analyzing test impact can extend beyond just interpreting data—it can also help forecast future outcomes based on past experiments. By leveraging advanced machine learning algorithms, your AI assistant can model potential scenarios and project the long-term effects of specific changes, enabling teams to prioritize tests better.
The need for new governance frameworks
The above use cases sound great, but you can’t rush in without some groundwork first.
Just as you’d have a contract between a company and a new hire, you’ll need to create a governance framework that explains where AI can be used and what’s above its pay grade (think anything connected to sensitive or personal user data).
A governance framework should be created by technical, data, experimentation, and legal professionals. Create a practical guide with real-world examples and write the framework in plain English. Ellie Hughes discusses this in greater detail;
What experimentation tasks should be off-limits to your AI assistant?
We’ve covered where your new assistant can help your experimentation team, but what and where are no-go areas? Outside of tasks that would be ethically or legally problematic, there are some aspects of A/B testing where AI struggles. Think of AI as a dedicated co-pilot; you wouldn't expect them to make important decisions or use their work without reviews. Your AI assistant is no different. However, certain tasks lead to more mistakes and hallucinations. Let’s find out what you should avoid.
Interpreting emotions
Humans are endlessly complex. If you’re ever in doubt of just how complex, try to read the emotions behind a text message. Is the smiley emoji happiness or passive-aggressive punctuation? It gets even more complicated when we recognize that people don’t always say what they think. Plus, there are cultural differences and contexts to consider.
Without understanding emotions, your coding skills or the quantity of insights won't help; reports can tell you what is wrong, but the heart will tell you why, and AI doesn't yet have a heart.
While AI can be trained to understand human emotion, they still struggle with some human favorites, such as sarcasm, irony, and tonal nuances. It’s also vital to understand that biases exist in AI models.
For example, one study found that emotional analysis technology assigns more negative emotions to black men’s faces than white men’s faces. AI must be trained on diverse datasets and built by diverse teams across gender, ethnicity, socioeconomic status, and views.
Neither humans nor AI are foolproof at interpreting emotions. The best solution is to use AI alongside diverse human teams to improve accuracy.
Final decision-making without human involvement
Many industry pundits have already suggested we treat AI tools like just another member of your team. Iqbal Ali elaborates;
Mike St Laurent also discusses the idea of human augmentation alongside your new AI coworker;
While AI can significantly enhance decision-making by providing data-driven insights and recommendations, it is crucial to maintain a balance between AI and human judgment.
Thinking of AI as a supportive tool rather than a decision-maker ensures that human expertise and contextual understanding complement AI's analytical capabilities.
This approach allows for more nuanced and well-rounded decisions, as humans can interpret AI-generated insights within the broader context of organizational goals, market dynamics, and ethical considerations.
The key to using AI is knowing when and where to use it
In conclusion, AI is a valuable tool in experimentation, but it's not a magic bullet. By understanding where AI excels, you can leverage it to enhance your testing program. This includes viewing AI as your new dedicated co-pilot who can assist with tasks such as coding test variants, analyzing vast amounts of user feedback, detecting anomalies, and automating tedious project management jobs.
However, AI has limitations; it struggles with interpreting emotions and shouldn’t be given final decision-making authority.
Integrating AI into your testing process requires thoughtful governance frameworks and a collaborative approach where AI supports human expertise rather than replacing it. The key is to use AI strategically, knowing when it adds value and when human intuition and oversight are irreplaceable.
Thanks to all of the experts who provided their insights for this article;
- Johannes Mandl, Senior CRO Manager at Better Collective
- Iqbal Ali, Freelance Experimentation Consultant
- Ellie Hughes, Head of Consulting at Eclipse Group
- Eric Itzkowitz, Director of Conversion Rate Optimization at FuturHealth
- Anjali Arora Mehra, Experimentation leader
- Jonathan Shuster, Digital Marketing Optimization Consultant
- Mike St Laurent, Managing Director, NA at Conversion
- Marcello Pasqualucci, Head of Web at Travelopia
If you want to set your new AI assistant to work, check out our guide to crafting AI prompts in experimentation.