
Online research has transformed how scientists, businesses, and policymakers study human behavior. But while collecting data online is faster and more efficient than ever before, it also introduces new challenges, especially around participant recruitment and data quality.
In our recent CloudResearch workshop, Senior Research Scientist Aaron Moss walked attendees through the fundamentals of launching online studies, choosing participant platforms, and protecting data quality. The session provided a practical introduction for new researchers and a refresher for experienced teams looking to improve their online studies.
Key Takeaways
- Online research is now the dominant method of behavioral data collection, with roughly 5 billion surveys completed online each year.
- Where you recruit participants matters. Market research panels and researcher-centric platforms serve different research goals.
- Data quality is the biggest challenge in online research, with some platforms seeing 30–40% problematic responses without safeguards.
- AI agents and organized click farms are emerging threats researchers must actively monitor.
- Simple safeguards (such as attention checks, benchmarking questions, and open-ended responses) can dramatically improve data quality from human participants.
Watch the Webinar
The Rise of Online Research
Fifteen years ago, online behavioral research was novel. Today, it’s the norm.
Across disciplines like psychology and political science to market research and public health, 60–70% of studies now collect data online. The reasons are clear: speed, participant diversity, and researcher flexibility.
Speed
Online platforms allow researchers to collect data in hours or days instead of weeks or months.
Diverse Participant Pools
Researchers can recruit participants from different ages, backgrounds, and geographic locations—far beyond traditional university subject pools.
Flexible Study Designs
Online tools enable complex designs that would be nearly impossible in a physical lab. For example, researchers recently used CloudResearch Connect to run a 28-day randomized controlled trial involving daily interactions with ChatGPT. The study recruited about 1,000 participants who logged nearly 32,000 conversations and over 1,500 hours of audio data.
Coordinating a study of that scale in person would have been next to impossible.
Choosing the Right Participant Platform
One of the most important decisions researchers make is where to recruit participants. The workshop highlighted two primary ecosystems.
Market Research Panels
Market research panels aggregate participants from across the internet into large marketplaces—often tens of millions of people worldwide.
Strengths
- Extremely large participant pools
- Rapid data collection
- Strong demographic targeting
- Access to niche audiences
Limitations
- Limited control over participant management
- Difficulty running long or complex studies
- Higher rates of problematic responses
Research suggests that 30–40% of responses from these sources may be low quality without proper safeguards.
Researcher-Centric Platforms
Platforms like CloudResearch Connect were built specifically for behavioral science research.
They typically have smaller but highly vetted participant pools and give researchers full control over their studies.
Strengths
- Higher baseline data quality
- Direct communication with participants
- Greater control over compensation and study design
- Ideal for complex, longitudinal, or high-engagement studies
Because participants are vetted during onboarding and monitored over time, problematic data rates are often closer to 5% or less.
Why Data Quality Is the Biggest Challenge
While online research offers enormous advantages, data quality remains the single biggest risk. Without safeguards, bad responses can distort results or invalidate studies entirely. Research has identified several common sources of problematic data that range from people yea-saying to people committing fraud in click farms. Now, there is also the threat of AI agents.
Practical Strategies to Improve Data Quality
To address data quality problems from human participants, researchers can draw upon a few proven techniques.
Use Attention Checks Strategically
Design questions that detect “yea-saying” behavior.
Examples include impossible or implausible events, such as asking participants whether they recently attended a concert at a stadium that no longer exists.
Benchmark Responses
Include questions with known population rates, such as owning a specific product or participating in a rare activity, to verify that responses align with expected patterns.
Include Open-Ended Validation
Open-ended questions help identify:
- Non-native language inconsistencies
- Copy-paste behavior
- AI-generated responses
Combined with other checks, these items provide valuable insight into participant engagement.
Getting Started with Online Research
For researchers new to online studies, the workshop recommended a simple path:
- Experience studies as a participant to understand the process.
- Design your survey using a tool like CloudResearch Engage or Qualtrics.
- Pilot your study with a small sample.
- Use attention checks and validation questions.
- Choose a recruitment platform aligned with your study goals.
Learn More
Many of the methods discussed in the workshop are explored in the upcoming textbook:
Research in the Cloud: An Introduction to Modern Methods in Behavioral ScienceThe book—developed by the CloudResearch team and forthcoming from Cambridge University Press—provides a comprehensive guide to conducting high-quality online research.
Check out our Events calendar to register for our next webinar!











