Science of Justice
Our science, your art.
You've got the vision; we've got the data.
Is our science the right fit for your practice? Is the earth round? Let’s find out. We have created a unique suite of machine intelligence solutions that provide you with the best information in your legal cases. We explore insightful results through our proprietary algorithms with experts with decades of experience working with behavioral science issues or collaborating with legal advisors for successful case outcomes.
Science of Justice
7 Fatal Focus Group Analysis Mistakes
Use Left/Right to seek, Home/End to jump to start or end. Hold shift to jump forward or backward.
Ever walked out of a focus group riding high, only to realize later you were chasing a mirage? We dig into the seven hidden mistakes that quietly sabotage plaintiff focus groups and show how to replace seductive but shaky feedback with data you can actually use at trial.
We start where most strategies fail: recruitment. Convenience samples from Craigslist and generic online panels don’t mirror your jury pool and are now riddled with bots, farms, and professional survey takers. We break down purposive sampling, county-level quotas, and oversampling so your room reflects real demographics and decision styles. From there, we go inside the session to expose how groupthink, bandwagon effects, and courtesy bias manufacture artificial consensus, and why attorneys should never moderate their own groups. Neutral facilitators trained in psychology keep the conversation honest, probe dissent, and prevent subtle cues from steering the room.
Then we reframe the goal: focus groups diagnose; they don’t predict verdicts or damages. You’ll learn how to separate signal from noise, why three well-run groups capture most of the meaningful insight, and how to avoid overfitting to vivid anecdotes. We get practical with behavioral tools—facial coding that flags microexpressions at key moments and linguistic analysis that reveals who assigns agency, who leads with emotion, and which words backfire. Combined, these methods produce juror profiles that sharpen voir dire, refine openings and closings, and target discovery to pressure-test defense witnesses and language that alienates jurors.
If you’re ready to stop preparing for a phantom jury and start shaping strategy around how your venue truly thinks, this conversation is your blueprint. Subscribe, share with a colleague who runs focus groups, and leave a review telling us the one change you’ll make to your next mock—what will you fix first?
https://scienceofjustice.com/
The High Stakes Of Focus Groups
SPEAKER_01Aaron Powell All right. Our mission today is pretty precise. We're zoning in on a really critical part of high-stakes civil plaintiff law, the focus group.
SPEAKER_00Right.
SPEAKER_01Now, this process, it's supposed to give you that qualitative data, you know, deep insight into how a jury might see your case. Trevor Burrus, Jr.
SPEAKER_00Exactly. But too often.
SPEAKER_01Aaron Ross Powell But yeah, too often it becomes well, maybe the attorney's biggest strategic self-sabotage risk.
SPEAKER_00Trevor Burrus That's the danger.
SPEAKER_01Trevor Burrus So we're looking at seven critical errors. The ones that quietly undermine focus group analysis. We want to show you how to turn that qualitative stuff into a real actionable trial strategy.
Seven Silent Mistakes Framework
SPEAKER_00Aaron Powell And the stakes. They couldn't be higher, really, for the plaintiff bar. A poorly run focus group is honestly it's worse than doing nothing at all. Trevor Burrus, Jr.
SPEAKER_01Worse than nothing.
SPEAKER_00Trevor Burrus, Jr.: Absolutely. It doesn't just give you useless data. It actively creates maybe a false sense of security. Or maybe the opposite, it could lead you to tragically undervalue a really strong claim. If your focus groups aren't getting at those genuine juror perceptions, the real cognitive, the emotional drivers behind their decisions, well, you're just misjudging risk. You're setting yourself up for an ineffective presentation, and you might be walking away from the best settlement opportunities.
SPEAKER_01Aaron Powell Okay, so we've broken it down into seven major mistakes that silently sabotage things. Right. And like any structure, failure often starts right at the foundation. So we're going to spend this first part really digging into the fatal errors made during recruitment and sampling. That's mistakes one and two.
SPEAKER_00Aaron Ross Powell Exactly. Because if that input, if the sample itself is flawed, then everything else is built on sand.
SPEAKER_01Trevor Burrus, Jr.
SPEAKER_00Precisely. Everything that follows is well basically meaningless. We can't stress this enough. If the sample is bad, your discussion, your analysis, your whole courtroom strategy, it's all based on fundamentally flawed ideas. Aaron Powell Okay.
SPEAKER_01Mistake number one, this is the convenience trap. You know, treating an unrepresentative sample as if it holds some kind of universal truth about your specific trial venue. Right. When focus groups come from those easy, quick channels, the low effort sources, well, they just don't mirror the actual jurors you'll face. Not their demographics, their socioeconomic reality, or even the basic values in that jurisdiction.
SPEAKER_00Convenience sampling is really the opposite of robust research. By definition, it's putting speed and ease ahead of accuracy.
SPEAKER_01You're just getting anyone you can.
SPEAKER_00Anyone who will show up for a quick buck, basically. And that guarantees severe bias. Often you get this deep homogeneity in the group that's completely different from a real diverse jury pool.
SPEAKER_01And the danger.
SPEAKER_00The danger is when you confuse that biased feedback with what the real jury pool thinks or values. You end up maybe overestimating your case's strength or just as bad, dangerously undervaluing potential damages. The gap between a convenience panel and a real county jury can literally mean millions of dollars.
Why Craigslist And Panels Fail
SPEAKER_01Aaron Powell Let's get specific then, these poor sources. You often hear about platforms like, say, Craigslist. Why is using Craigslist, for example, so problematic for the kind of high-quality, nuanced research that complex civil cases really demand?
SPEAKER_00Aaron Powell Well, the population you typically find on Craigslist is inherently skewed, unless you're doing some really extraordinary filtering, which most aren't.
SPEAKER_01Skewed out.
SPEAKER_00Usually lower education, often lower income individuals. And their main motivation, it's the quick incentive, the money. Right. So because payment is the primary driver, they often show less commitment to the actual exercise. They're less likely to seriously engage with complex case details, tough narratives, their responses become really problematic because you just can't generalize them to the broader, often more affluent, more diverse jury pool you'll actually face in court. Aaron Powell Okay.
SPEAKER_01So Craigslist has its issues.
SPEAKER_00Right.
Bots, Farmers, And Cost Explosion
SPEAKER_01But that low quality problem seems, well, massively amplified when we look at online panels. You mentioned research pointing to a real crisis in quality control there.
SPEAKER_00Aaron Powell Oh, the data is quite alarming. We really need to grasp the scale of this degradation. There's confirmed data showing the proportion of low quality responses, you know, answers that are statistically way off, inconsistent, just nonsensical. It's jumped from around maybe 10.4 percent in past years to well, as high as 62 percent recently. Aaron Ross Powell, Jr.
SPEAKER_0162 percent.
SPEAKER_00Wow. And that's driven by two main things bots, which are just computer programs filling out surveys automatically.
SPEAKER_01Right. Automated responses.
SPEAKER_00And farmers, these are people often overseas using server farms to get around location restrictions and just churn through huge volumes of this work.
SPEAKER_01Aaron Powell Hold on. If the failure rate is as high as 62%, that means for every, say, ten responses you pay for, more than six of them are potentially unusable noise you have to discard immediately. So what does that do to the effective cost of getting just one good usable response?
Non‑Naivete And Panel Fatigue
SPEAKER_00Aaron Powell It makes the cost skyrocket. Precisely. If you budget, say,$10,000 for data collection and$52% of what you get back is junk, well, you didn't really spend$10,000 for your useful data. You effectively spent closer to$26,000 for the actual insights you could use. You're paying a massive hidden premium just to filter out the garbage.
SPEAKER_01Okay, so bots and farmers are one part, but you also mentioned issues with the actual human participants themselves. This idea of non-naivete.
SPEAKER_00Right. Non-naivete. It's a critical problem.
SPEAKER_01Can you break that down for us? What does that mean?
SPEAKER_00It means they're no longer naive participants. Estimates suggest that maybe 40% of all the responses on clickworker panels are actually generated by just 10% of the participants. Wow.
SPEAKER_01The small group doing most of the work.
SPEAKER_00Exactly. They spend so much time taking surveys that they become expert test takers. They get really savvy about what researchers are trying to study.
SPEAKER_01So they gain the system.
SPEAKER_00To some extent, yes. They even organize in online communities, sharing information about which studies pay well, which ones to avoid because they require too much effort. And this behavior leads to things like inattention, sometimes deliberate self-misrepresentation, and what researchers call socially desirable responding.
SPEAKER_01Meaning they tell you what they think you want to hear.
Purposeful Sampling With Quotas
SPEAKER_00Or what gets them through fastest. They're just trying to get paid quickly, not necessarily give you an honest, deep reflection of their actual views on liability or damages.
SPEAKER_01Okay, so if convenience sampling in these generic online panels are traps, what's the strategic answer? How do plaintiff attorneys build a sample that's actually representative?
SPEAKER_00Aaron Powell The only real solution is moving to what we call purpose of sampling.
SPEAKER_01Purpose of sampling, yeah.
SPEAKER_00Yes. This means rigorously matching your focus group pool to the specific demographics of your county jurisdiction. And you have to use current census data for that.
SPEAKER_01Aaron Powell And it's more than just age and race, right?
SPEAKER_00Oh, much more. That's to surface level. You have to drill down into strategic socioeconomic indicators. Things like median household income, maybe political affiliation proxies based on neighborhood voting patterns, education levels specific to that actual venue.
SPEAKER_01Aaron Powell Can you give an example?
SPEAKER_00Sure. Let's say you're handling a case in a highly educated, you know, white-collar suburb. Pulling your respondents from a pool where maybe 60% are unemployed, or just looking for side hustle income, that's completely useless. The mindset is totally different.
Oversampling To Protect Quality
SPEAKER_01Aaron Powell And practically speaking, this requires setting pretty rigid quotas during recruitment, doesn't it? How do you fix imbalances like the fact that online recruitment often skews heavily male?
SPEAKER_00Aaron Powell You absolutely have to use quotas to force representation. Online recruiting might naturally give you, say, a 70-30 male-female split. Right. But if the census data for your specific county says it's 55% female, well, you must cap female participation at 55% during recruitment. That forces the remaining slots to be filled by men, correcting that fundamental imbalance right from the start.
SPEAKER_01So it's a much more active, managed process.
SPEAKER_00It requires a really dedicated recruitment effort. It goes way beyond just posting an ad online. And another key limitation, specifically with click workers online, is that it generally only lets you filter by state, not the specific county. That makes it inherently unrepresentative of the highly localized population you actually care about.
SPEAKER_01Okay, so even with the best screening, given all these quality control risks, you still have to plan for people not showing up or finding some low quality participants slipping through. You mentioned oversampling. What's the practical rule there for these qualitative mock juries?
SPEAKER_00Yeah, the risk of a no-show or just a poor quality participant, someone who's not engaged, it's much higher in qualitative small group research compared to big surveys. Okay. So for a quantitative study, maybe 100 plus participants, oversampling about 10% is usually okay. But for qualitative focus groups, your typical eight to 12-person mock juries, you need to be much more aggressive.
SPEAKER_01How aggressive?
SPEAKER_00We generally recommend recruiting 150% of your target number.
SPEAKER_01150%. Yes.
Mistake 2: Ignoring Hidden Biases
SPEAKER_00So if you want a final group of 12 people for your mock jury, you should actually recruit 18 individuals. Why so many. That buffer is critical. It lets you immediately handle no shows without scrambling. It helps ensure you hit your diversity quotas. And importantly, it allows you to swiftly remove those individuals who might show up, but exhibit such extreme bias, the kind you definitely strike during actual voir dire without compromising your group size or composition. That extra effort upfront in recruitment is essential for maintaining quality and consistency in a small group setting.
SPEAKER_01Okay, so that leads directly into mistake number two, which follows right from sample quality, ignoring the biases that participants bring with them.
SPEAKER_00Exactly.
SPEAKER_01We're talking hidden biases, personal ideologies, maybe past experiences. These things fundamentally shape how people react in the focus group.
SPEAKER_00Absolutely. And if you just take their feedback at face value without trying to understand the baggage they carry into the room, you risk building a case narrative that sounds fantastic in the conference room, maybe gets nods from the focus group, but then just utterly collapses under pressure and scrutiny of a real courtroom.
SPEAKER_01Because it wasn't built on a solid understanding of juror psychology.
Psychographic Screening And Checks
SPEAKER_00Precisely. This means the screening process has to be way more rigorous than just checking demographic boxes. It really needs to be developed by professional behavioral scientists. Okay. And it has to be layered. It starts with, you know, basic qualification surveys, filtering out people who clearly don't meet essential criteria. But then during the online research phase itself, you have to use active scientific filtering methods to ensure people are actually engaged and suitable.
SPEAKER_01Aaron Powell Let's focus on those scientific measures. Beyond asking age or income, what does a proper scientifically sound screening measure actually look like for a complex civil case?
SPEAKER_00Well, it looks at psychographics, not just demographics. We're trying to measure underlying traits, things like authoritarianism levels, maybe anti-corporate sentiment, how much they rely on science versus intuition or their core beliefs about personal responsibility. These kinds of traits predict decision making far more reliably than just knowing their income bracket.
SPEAKER_01Aaron Powell And how do you measure those things accurately?
SPEAKER_00We embed internal validity checks within the questionnaire itself. So we don't just ask, say, what are your politics? We might include a question that seems simple, but requires a specific, maybe counterintuitive answer, like for this question, please select the option maybe.
SPEAKER_01An attention check.
SPEAKER_00Exactly. If the participant fails that basic instruction, it tells you they're either demonstrating poor effort, not reading carefully, or they might simply be a bot and they get filtered out immediately. These screening measures have to be reliable, meaning they give consistent results and valid, meaning they actually measure the psychological trait they claim to measure.
Mistake 3: The Positive Feedback Trap
SPEAKER_01You mentioned earlier the danger of using experienced participants. If I'm pulling from a platform where people do focus groups week after week, their feedback is compromised, yes. Because they're no longer representative of a truly naive juror, someone coming in fresh.
SPEAKER_00That's a critical flaw inherent in convenience sampling. If you rely heavily on those platforms, you end up with participants who are too experienced with the whole focus group process.
SPEAKER_01They know the drill.
SPEAKER_00They know the rhythm of the presentation, they anticipate the kinds of questions you'll ask. Sometimes they even recognize standard damage models or arguments. Their responses become, well, highly stylized. They don't generalize well to the genuinely inexperienced nature of real jurors who are walking into the courthouse completely cold, hearing these complex arguments for the very first time.
SPEAKER_01So the goal of screening isn't just demographics, it's finding someone who likely hasn't deeply considered, say, the nuances of product liability or medical negligence before that day.
SPEAKER_00That's a huge part of it. Yes. Finding that naive perspective is key.
SPEAKER_01So wrapping up these first two mistakes, the foundation of your strategy really hinges on this highly scientific quality control.
SPEAKER_00It does.
Groupthink, Bandwagon, Courtesy Bias
SPEAKER_01Making sure the people you're listening to are actually the right people, and that you have some understanding of the preexisting filters they're using to process your case.
SPEAKER_00Precisely. If your selection and screening are weak, your entire strategy risks being based on, well, unreliable fiction. You end up preparing your case against a phantom jury, not the real one.
SPEAKER_01Okay, so let's say we nail the recruitment. We get the right people in the room, virtual or physical. Now we move into the dynamic phase, the discussion itself. And this is where mistake number three often creeps in.
SPEAKER_00Right. The social setting starts to play a role.
SPEAKER_01And fundamental human psychology starts, potentially, skewing the data you receive. Mistake three is this overconfidence in what seems like positive feedback. We all suffer from confirmation bias, right? We tend to hear what we want to hear.
SPEAKER_00It's a very natural human tendency.
SPEAKER_01So if the participants seem to really love your opening argument, or they seem very sympathetic to your client, that affirmation can feel incredibly good. Intoxicating even.
Why Attorneys Shouldn’t Moderate
SPEAKER_00But it's deeply dangerous, extremely dangerous, because group dynamics are notorious for creating a kind of artificial consensus that can easily mask true dissent or underlying issues.
SPEAKER_01But you're saying positive feedback needs scrutiny.
SPEAKER_00All feedback does, but especially the positive kind. Favorable responses must always be rigorously stress tested. The risks here are well documented in social psychology.
SPEAKER_01Like what?
SPEAKER_00Well, you have groupthink. That's where participants start suppressing diverse or challenging viewpoints just to conform to what they perceive as the dominant opinion, especially if it gets established early on.
SPEAKER_01They don't want to rock the boat.
SPEAKER_00Exactly. And then there's just simple peer pressure. Individuals might subtly, or not so subtly, modify their genuine responses to align better with what they think the group wants, or maybe what they think the moderator wants to hear. This is especially true if you have one or two very loud, opinionated people in the room who take control early.
SPEAKER_01So the very informality we try to create, you know, to mimic a jury room. Hmm that can actually be the source of these biases. You mentioned specific ones like the bandwagon effect and courtesy bias. Aaron Powell Yes.
SPEAKER_00The bandwagon effect is pretty straightforward. People just start following the opinions of others, especially when dealing with complex legal stuff they don't fully grasp. They look around for cues on how to respond.
SPEAKER_01Safety and numbers.
SPEAKER_00Sort of. And the courtesy bias is equally insidious, maybe more so. This is where participants actively suppress their true, possibly negative opinions. They adopt a more acceptable or pleasant stance, often just to avoid conflict or seem agreeable.
SPEAKER_01Can you give an example of that?
Neutral Facilitation Techniques
SPEAKER_00Sure. Imagine the moderator asks, Did you find the defendant's expert witness convincing? The first person maybe hesitantly says, Yeah, mostly. Suddenly the next four people might find ways to agree, even if they had serious doubts. They might say, Yeah, he seemed knowledgeable or I could see his point, just to maintain that comfortable group dynamic. You lose the real friction, the genuine disagreement that drives actual jury deliberations.
SPEAKER_01Okay, this brings us squarely to the role of the moderator. This is often a point of debate. Many really experienced plaintiff attorneys feel they should run their own focus groups. They know the case best, right?
SPEAKER_00Aaron Powell They do know the case best.
SPEAKER_01Aaron Powell But you argue this introduces a massive, maybe unavoidable, facilitator bias.
SPEAKER_00Yeah.
SPEAKER_01How critical is that bias, really? If I'm a seasoned attorney, can I just be objective, suppress my reactions? Is hiring an external psychologist truly worth that extra cost?
Mistake 4: Confusing Signal And Noise
SPEAKER_00Aaron Powell It is absolutely critical. And frankly, no. You generally cannot suppress it effectively for your own case because the bias isn't always conscious. Attorneys are naturally experts in the in their case material, they're excellent advocates, often great cross-examiners. Right. But they are not neutral, objective moderators when their own money, their own client, their own reputation is on the line. It's psychologically almost impossible to completely turn off that personal and financial investment.
SPEAKER_01Even the subtle things matter.
SPEAKER_00Especially the subtle things. A slight frown when someone criticizes your damages model, a barely perceptible nod or smile when they validate your main theory of the case. Even just the way you phrase a follow-up question can inadvertently cue participants about your perspective, what you believe.
SPEAKER_01And participants pick up on that.
SPEAKER_00Oh, absolutely. They are highly attuned to those nonverbal cues. And knowing they're being paid, there's often an unconscious pressure to please the person running the show, the person who hired them.
SPEAKER_01So the solution really has to be objective external moderation.
How Many Groups Are Enough
SPEAKER_00Aaron Powell Precisely. To get unbiased data, you really need neutral moderators. Typically, these are people with expertise in human behavior, maybe advanced degrees in social science, psychology. Their skill isn't knowing the case law, it's guiding a discussion without imposing structure or revealing expectations.
SPEAKER_01How do they do that?
SPEAKER_00They're trained to use neutral language, maintain a poker face, actively probe for dissent rather than driving towards consensus. Their main job is to manage those tricky group dynamics, to ensure groupthink is challenged, and that all perspectives, even unpopular ones, get heard. This prevents the moderator's own hopes or expectations from unintentionally tainting the integrity of the data you collect.
SPEAKER_01Aaron Powell All right, let's move into the analysis phase now. We've got the data, hopefully moderated well. Mistake number four is failing to separate the genuine insights from what's basically just noise.
SPEAKER_00Aaron Powell Right. This is crucial.
SPEAKER_01Not every comment, not every opinion carries the same weight. We'll latch on to outlier comments maybe because they were really dramatic or memorable or just particularly well articulated. We risk basing our strategy on, well, clutter instead of significant reproducible trends.
Limits Of Generalization
SPEAKER_00Aaron Powell Trial strategy absolutely demands clarity and importantly replicability. The core mistake here is confusing the richness or depth of qualitative feedback with quantitative validity. Meaning just because one participant had a really powerful, memorable objection to your key witness doesn't automatically mean the entire jury pool in your county shares that specific objection. You might be reacting strongly to an anecdote, not a reliable trend.
SPEAKER_01Okay. Now there's some interesting research about the optimal number of focus groups you should even conduct. How do we figure out when we've gathered enough information to actually define a trend without just collecting more and more noise?
SPEAKER_00Yeah. That finding is really instructive for attorneys. The research suggests you generally only need about three focus groups to capture roughly 80% of the total necessary information about your case themes and potential juror reactions.
SPEAKER_01Only three.
Mistake 5: Treating Mocks As Predictions
SPEAKER_00Around three, yes, to get the core themes. After that third group, the returns start diminishing significantly. Conducting, say, five, six, seven groups. It usually adds only a tiny bit more new information while substantially increasing the margin for error, not to mention wasting huge amounts of time and budget.
SPEAKER_01So doing more isn't always better.
SPEAKER_00Often it's worse. If you run seven groups, you're probably just hearing the same core 80% over and over again. But you increase the risk of overinterpreting minor, maybe non-generalizable differences you see in that remaining 20% mistaking noise for signal.
SPEAKER_01And this ties right back to the sample representation issue. If you have a small sample, let's say three groups of 12 people each, so 36 people total, how far can you realistically generalize those conclusions to a whole county of maybe three million residents?
Real Versus Mock Jury Psychology
SPEAKER_00Aaron Powell You have to be incredibly cautious, extremely cautious. Generalized conclusions are very often undermined by that limited sample representation. A small sample size like that should never be treated like a poll of a thousand people, giving you statistical certainty.
SPEAKER_01So what does it give you?
SPEAKER_00It confirms themes, it validates language that resonates or backfires, it allows you to refine your arguments, identify potential landmines, but it doesn't give you statistical confidence in predicting a specific verdict outcome or dollar amount.
SPEAKER_01And the analysis needs structure.
SPEAKER_00Absolutely. The analysis itself must be highly structured to prevent the analyst's own influence. The facilitator's expectations or hypotheses shouldn't inadvertently seek confirmation. For example, you have to be careful about question framing, making sure questions don't unintentionally limit the range of possible answers, which would stifle the exploration of genuinely diverse perspectives.
Focus Groups As Diagnostics
SPEAKER_01Okay, that leads us perfectly into mistake number five. And this might be the most dangerous strategic illusion in this whole process. Assuming that focus groups actually predict verdicts or damage amounts.
SPEAKER_00Yes, this is a huge one.
SPEAKER_01Attorneys watch the mock jury deliberation, they see maybe a liability vote, they hear a damage number discussed, and they start thinking, okay, that's our projected outcome.
SPEAKER_00It's a profound and often costly over-interpretation. It can absolutely sink a case strategy.
SPEAKER_01Why is it so wrong? They look like jurors, they talk like jurors.
Mistake 6: Skipping Behavioral Data
SPEAKER_00They do, but they aren't operating under the same psychological or situational constraints as real jurors. That's the critical difference. Mock decisions reveal perceptions, they reveal how people process information, but they do not reliably predict outcomes. Okay. Interpreting those mock decisions as direct forecasts is a mistake that very frequently leads attorneys to misjudge their chase risk. And critically, it can cause them to tragically bypass crucial, often optimal settlement windows because they're overconfident from a focus group result.
SPEAKER_01Aaron Powell What are the key psychological differences then between a mock deliberation and a real one that make prediction basically impossible?
SPEAKER_00There are several major ones. First, real jurors receive formal judicial instructions on the law on concepts like negligence, burden of proof. Mock jurors typically don't or get a highly abbreviated version. Those instructions fundamentally change how facts are processed.
SPEAKER_01Makes sense.
SPEAKER_00Second, real jurors feel the social pressure of the formal courtroom setting, maybe sequestration, and the profound emotional weight of their civic duty. They know it's real. Mock jurors lack that same level of pressure and formality.
SPEAKER_01It's more academic for them.
SPEAKER_00Exactly. And third, perhaps most importantly, mock jurors lack the sense of absolute finality. A real jury knows their decision directly impacts real human lives, potentially millions of dollars. A mock juror knows they're participating in research, they're getting paid, and their decision, while they might take it seriously, doesn't have those same real world consequences. The stakes are vastly different.
Facial Coding For Moment Reactions
SPEAKER_01Okay, that's very clear. So if they don't reliably predict the verdict or the final dollar amount, what is their actual true purpose? Why do them?
SPEAKER_00Their true purpose is invaluable, but it's about strategic refinement. Focused groups provide incredibly useful data on perceptions, perceptions about the credibility of your attorneys, the authority and trustworthiness of your expert witnesses, the believability or impact of specific pieces of evidence, the effectiveness or potential pitfalls of your voir dire questions.
SPEAKER_01So it's diagnostic.
SPEAKER_00Precisely. It's a diagnostic tool. It allows for essential refinement of your case framing, your narrative, your overall strategy. Focus groups illuminate the path toward a favorable outcome by showing you where jurors are likely to get confused, where they might resist your arguments, or where they find the most resonance and assign accountability. They tell you how jurors in your venue are likely to think about your case, not definitively what they will ultimately decide.
SPEAKER_01Starting with mistake number six. Mistake six is overlooking the deeper psychographic and behavioral data. You know, surface-level insights like which side participants simply said they liked more, or just tallying up an initial gut reaction liability vote that's just not enough for sophisticated modern trial strategy.
SPEAKER_00Aaron Powell, yes.
SPEAKER_01Aaron Powell The real power, the real strategic edge from a focus group comes from identifying those underlying emotional triggers, the implicit biases, the things that really drive decision making, often below the level of conscious awareness. We need to get beyond just the words people say.
Building Juror Profiles
SPEAKER_00Moving from just getting data to extracting genuine insight. And for that, you really need advanced data techniques often developed by behavioral scientists to get at those true psychological insights.
SPEAKER_01Aaron Powell, Why can't we just rely on what people tell us?
SPEAKER_00Because jurors, like all people, often cannot accurately articulate why they reacted negatively to a piece of evidence or a particular argument. Their conscious feedback, what they tell you in the debrief, is frequently just a post hoc rationalization of a deeper gut level reaction. We need methods to capture that subconscious reaction as it happens.
SPEAKER_01Okay, so how do you access that deeper layer, that emotional cognitive information that participants might not even be aware of they're revealing?
SPEAKER_00Aaron Powell Two key techniques are incredibly powerful here. First is facial analysis.
SPEAKER_01Analyzing their faces.
Mistake 7: Failing To Integrate Findings
SPEAKER_00Exactly. Especially when conducting virtual focus groups, you can record the participant video feeds. Software, combined with trained human analysis, can then examine microexpressions, those fleeting facial signals of underlying emotion. Happiness, sadness, anger, fear, worry, surprise, disgust, contempt.
SPEAKER_01And this gives you objective data.
SPEAKER_00It gives you objective moment-by-moment data. It allows us to pinpoint exactly when during the presentation a juror reacted negatively, say, with disgust or contempt, or positively, maybe with sadness or surprise to a critical fact, a specific phrase you use, or even just a particular moment in your expert's testimony. Aaron Powell Wow.
SPEAKER_01So you're not just realizing later, oh, the group didn't like my expert. You might know the precise moment, the exact sentence he said, the specific grab he showed, maybe even the tone of voice he used that triggered widespread disgust or worry in the mock jurors.
SPEAKER_00Precisely.
SPEAKER_01How does that translate into an actionable strategy adjustment?
Driving Discovery And Depositions
SPEAKER_00It allows for extremely precise surgical adjustments to your case presentation. For example, if, say, eight out of ten mock jurors consistently displayed facial markers of disgust, the instant you introduce the defendant corporation's annual revenue figure, you know that fact presented that way is toxic. It needs to be handled completely differently, maybe contextualized much earlier, or perhaps even de-emphasized depending on the strategy. Or if they showed clear signs of worry or confusion when you explain the complex medical procedure involved, you know a specific portion needs simplification, clearer visuals, or maybe very strong supportive evidence immediately following it to alleviate that anxiety.
SPEAKER_01Okay, facial analysis is one. What's the second key technique you mentioned, linguistic analysis? That looks at the structure of their thought.
Time, Cost, And Opportunity
SPEAKER_00Yes, linguistic analysis. This involves analyzing the transcripts of their discussions and also their written responses. Using sophisticated computational methods, often derived from linguistics and psychology, we're not just looking at what words they use, but how they structured their language.
SPEAKER_01Give us a concrete example of how that works. What can language structure reveal?
SPEAKER_00It can reveal a lot about their underlying emotions, their motivations, even their cognitive processes information far beyond what they state explicitly. For instance, if a juror frequently uses passive voice when describing the defendant's actions, saying things like, well, the product was put on the shelf instead of the company put the product on the shelf.
SPEAKER_01Okay.
SPEAKER_00That pattern might indicate a subtle failure to assign accountability or agency to the defendant. Conversely, if they consistently use strong active voice, the defendant chose to ignore the safety warnings, it signals a much clearer assignment of responsibility and liability. Interesting.
SPEAKER_01What else?
Core Takeaways And Final Challenge
SPEAKER_00We can also analyze the frequency of certain types of words. For example, comparing the use of emotional words like fair, unfair, victim suffering, versus more rational or analytical words like data, standard, evidence, policy. This balance can give you a psychological map of their individual decision-making style. Are they primarily driven by emotion or logic in this context?
SPEAKER_01And these powerful behavioral analyses, the facial coding, the linguistic patterns, they're ultimately combined.
SPEAKER_00Yes, absolutely. They're combined to create the ideal juror profile.
SPEAKER_01Okay.
SPEAKER_00We merge these sophisticated behavioral data streams, the facial reaction, The linguistic markers, the emotional cues detected with everything else, their explicit comments in the public discussion, their private written questionnaire responses, their demographic and psychographic profiles from screening. This comprehensive data set allows us to build a really precise, actionable personality profile of the juror, most likely to be favorable and unfavorable for your specific case in your specific venue.
SPEAKER_01And that profile drives War Dyer.
SPEAKER_00It becomes the foundation for refining your War Dyer questions. It helps you develop case themes and narrative structures that, you know, based on validated psychological triggers, are most likely to resonate with the right kind of jurors you want on your panel.
SPEAKER_01Okay, that brings us to the final mistake. Mistake number seven. Isolating the focus group findings from your broader litigation strategy. You can do everything else right. Perfect sampling, expert moderation, deep behavioral analysis. But if that data just sits in a report on a shelf after the session is over, then even the best data loses almost all its strategic value.
SPEAKER_00Yes, focus groups cannot be treated as standalone academic exercises.
SPEAKER_01They have to permeate the whole process.
SPEAKER_00They absolutely must permeate and actively inform the entire litigation process for the plaintiff's case from the earliest stages right through to the closing argument. Data that isn't integrated is frankly just an expensive piece of paper.
SPEAKER_01Let's trace that. How should this information actually move through the case timeline to show that kind of total integration?
SPEAKER_00Aaron Powell Well, it really has to start early. The insights should immediately inform your discovery strategy and your depositions. Let's say your focus group clearly revealed that a key juror pain point is when the defense expert uses specific, maybe obfuscating industry jargon, or perhaps certain phrases the defense uses trigger high levels of juror contempt or skepticism.
SPEAKER_01Okay.
SPEAKER_00You use that knowledge in depositions, you build your deposition outline to specifically target and hammer those phrases. You force the defense expert to commit to using language that you now know is likely to alienate a real jury. You're essentially using the focus group findings to proactively engineer weaknesses in the defense's testimony before you even get to trial.
SPEAKER_01Wow. Okay. That really shifts the focus group from just being a preview of the trial to being an active weapon during discovery. Trevor Burrus, Jr.
SPEAKER_00Precisely. And beyond discovery, the data obviously refines your fact wording and framing. You ensure you're consistently using language that was validated by the focus group, language that resonates positively and avoids triggering known negative biases.
SPEAKER_01It shapes the narrative.
SPEAKER_00It dictates the development of your opening and closing arguments. You structure the entire narrative around the themes, the analogies, the stories that you saw land most powerfully and effectively with the mockers.
SPEAKER_01And jury selection.
SPEAKER_00And finally, yes, it's absolutely indispensable for jury selection. You use that ideal and anti-ideal juror personality profile derived from all that behavioral data to guide your questioning and ultimately to select the most favorable panel possible within the constraints you have.
SPEAKER_01This whole process, done right, highlights the immense time commitment and, frankly, the cost involved, especially if attorneys try to do it themselves without the right resources or expertise. When you talk about effective focus groups, you're really talking about massive time investments. Huge investments, yes. What does a proper focus group realistically demand in terms of, say, an attorney's own time?
SPEAKER_00It's a massive commitment, which is exactly why poorly implemented corner-cutting DIY groups often end up being financially devastating in the long run. Research and practical experience estimate that just conducting robust, purposive recruitment and proper screening takes a minimum of 30 days of lead time.
SPEAKER_0130 days just for recruitment.
SPEAKER_00Minimum for quality. And then a single effective focus group session typically demands somewhere between 60 to 80 hours of lead attorney time.
SPEAKER_0160 to 80 hours for one session.
SPEAKER_00Yes. That includes deep case preparation specifically for the focus group format, reviewing materials, observing the session itself, even if moderated externally, and then the critical phase of analyzing the data, integrating the findings. Plus, you have to factor in maybe 40 hours of dedicated staff time for logistics, participant management, transcription, video editing, et cetera, per session.
SPEAKER_01Wait a minute. If it's 60 to 80 hours of my time plus 40 hours of staff time for just one session, doing it yourself badly or without fully committing those resources essentially guarantees that a flawed DIY focus group costs you far more and lost billable hours and potentially flawed strategy than just hiring an external expert in the first place.
SPEAKER_00Almost certainly, yes. When you accurately calculate those hours as opportunity cost time you could have spent on other profitable casework, or even just preparing the core case itself, the true cost of an inefficient, poorly executed DIY group quickly dwarfs the fee for relying on specialized expertise.
SPEAKER_01So the investment is in efficiency and quality.
SPEAKER_00Exactly. Effective focus groups require significant dedicated time investment no matter who does them. But expert execution ensures that the substantial time and money you do invest yield data that is high quality, genuinely actionable, and ultimately worth it. It saves valuable case preparation time in the long run and maximizes your chances of understanding the real dynamics influencing recovery. Hashtag tech outro.
SPEAKER_01Okay, so we've really covered a lot of ground here, hitting the seven critical mistakes that can silently sabotage your case strategy. We went from those fundamental flaws right at the beginning, recruitment issues like the danger of convenience sources, the click worker bot crisis, the absolute necessity of purpose of sampling all the way through the psychological pitfalls that happened during the session. And in analysis, things like groupthink, confirmation bias, and that critical failure to distinguish the real strategic signal from just outlier noise.
SPEAKER_00And the overarching, I think, central takeaway from all this material is really clear. The single most damaging mistake a plaintiff attorney can make in this area is treating all the focus group data they gather as equally valid or equally predictive.
SPEAKER_01Aaron Powell Just taking it all at face value.
SPEAKER_00Exactly. Without that scientific rigor in the sampling, without objective skilled moderation, and without sophisticated behavioral interpretation that goes beyond the service, these sessions can actively mislead you. And that error has very real tangible consequences. It can cost clients millions in undervalued cases, or maybe even worse, steer you confidently towards a bad trial outcome when a perfectly good settlement window was actually open and available.
SPEAKER_01It really drives home the point that focus groups are meant to model the real world, right? To simulate that high stakes decision-making process of a jury.
SPEAKER_00That's the goal.
SPEAKER_01But if the very foundation, the sample you choose, the biases you fail to screen for or account for, if that foundation itself is fundamentally flawed, are you truly gaining insight into how your case will play out? Or are you maybe just gaining a false and dangerous sense of confidence based on interacting with a completely fictitious reality?
SPEAKER_00That's the question.
SPEAKER_01So maybe the final thought for you, the listener, is to critically review your own current focus group methodology right now, particularly the sampling. Ask yourself honestly are the people you're listening to really representative of the people who will actually hold your client's fate in their hands?
Podcasts we love
Check out these other fine podcasts recommended by us, not an algorithm.