Is ChatGPT ESOMAR 37’s Friend or Foe?

Intro to the ESOMAR 37

ESOMAR is the major industry organization for market research around the globe. They have a very helpful questionnaire called the “37 questions to help buyers of online sample.” I prefer to call it the “ESOMAR 37.”

Many sample providers have chosen to produce a document that has their answers to all 37 questions and which ESOMAR posts on this website. Most of the reports are 15-20 page PDFs. The idea is that if you are shopping for online survey sample, you can read all the reports on the site or choose among the companies in which you are interested and understand the panels’ various strengths and weaknesses.

While it’s not difficult to read any individual company’s report, it is tedious to go through multiple reports and retain enough information to make a decision about who to work with. Furthermore, panels provide vague answers to some questions, making the whole experience feel a bit like mattress shopping. It would be great to have a smart friend help you quickly understand which companies do x, y, and z so you can move on with your busy day. This is where ChatGPT comes in.

The Big Question

Can you actually rely on ChatGPT to examine the contents of the ESOMAR 37 reports? The answer, as we’ll see, is “not really,” because you can get very different recommendations or answers to questions depending on how you phrase your prompt, which model you use, how you format the source files, and whether you toggle on the Deep Research tool. In short, you have to make a lot of decisions, and they’re all important, though if you play with ChatGPT enough, you can start to understand why certain panels were recommended in certain circumstances and you learn something about the panels along the way.

So, yeah, it’s not great to ask ChatGPT a question like “which panel has the best data quality?” However, you can use it to create a handy table that summarizes key features of each panel’s quality assurance protocol. I think this approach is more reliable, and I provide an example a little further down the page.

Methodology

I was able to get 52 ESOMAR 37 reports in a txt file format. This was a chore in itself, involving some python script that, surprise (!), ChatGPT wrote for me.

The following companies were not considered for the reasons provided.

Report Not Included Table
Report Not Included Reason Report Not Included
Data Diggers Broken link from ESOMAR Website
iMad Research File requested but not delivered
Online Market Intelligence No US presence
Probe Partners ESOMAR answers not found
SSRS ESOMAR answers not found

Here is the complete list of companies covered by the 52 reports.

1.      AFROEYE Research

2.      AYTM

3.      Bilendi

4.      Borderless Access

5.      DataSpring

6.      Dynata

7.      EMI Research Solutions

8.      Eno Research

9.      e-Research-Global

10.  ERT (Easy Reach Technologies)

11.  Cint

12.  Ipsos

13.  EMPanel

14.  Ronin International

15.  Frequent Research

16.  Full Circle Research

17.  GlobalSurvey

18.  GMO Research

19.  HBG (High Beam Global)

20.  IAPAC

21.  InfoAudienceResearch

22.  Insights Opinion

23.  Internet Research Bureau (IRB)

24.  Kaaya Research

25.  Kantar

26.  Logit Group

27.  M3 Global Research

28.  Macromill Southeast Asia

29.  Markelytics

30.  Netquest

31.  Nomadic Insights

32.  Norstat

33.  Ola Surveys/Survey Diem

34.  PMI Research

35.  Probity Research

36.  Prodege

37.  Prolific

38.  Pureprofile

39.  QRS Research

40.  Quest Mindshare

41.  Rep Data

42.  Rise2Research

43.  Robas Research

44.  Sample Junction

45.  Savanta

46.  SurveyMonkey

47.  Symmetric Research

48.  Syno International

49.  Talk Online Panel

50.  TNB Global Survey

51.  Toluna

52.  YouGov

Refining the List

I attempted to exclude companies that do not offer a US consumer panel. At first, I used the following prompt in ChatGPT 40.

“Which of these companies operate a US consumer survey panel?”

The answer included only 15 of the 52, which I know is not right. You can ask ChatGPT to try again and get different results, but I wasn’t confident in any of its answers to this question.

I scrapped the idea of filtering out companies that do not operate a consumer panel in the US, since I could always check that myself from the final list of companies I wanted to examine more closely.

Asking for a Recommendation

Next, I directed ChatGPT 40 to read my last blog post, which outlined the pros and cons of various identity verification methods. My goal was to teach ChatGPT which methods are better than others and from that, evaluate who does the best job ensuring this component of data quality.

"Here is a blog post I wrote about ways panel companies can prevent survey fraud: https://www.surveydiem.com/blog/how-to-prevent-survey-fraud Please read the blog post and consider the information as it relates to the best ways to validate panelists’ identities. Which companies do the best job ensuring that panelists are who they say they are?"

The five companies ChatGPT listed were M3, Kantar, Nomadic Insights, SurveyMonkey, and Prolific. I think this is a reasonable answer to the question, though it does exclude the Survey Diem panel, which I designed based around ID verification. Two interesting things may be happening here: a) my blog post didn’t take enough of a stand in promoting id verification as an essential process and b) my ESOMAR 37 report isn’t optimized for being included in an answer to this question.

Note: when I ran this analysis using pdf files instead of txt files, I got a different answer. Furthermore, when I analyzed the txt files using the Deep Research tool with ChatGPT 40, I was gratified to see Survey Diem as a top recommendation. The difference? The Deep Research tool used python occasionally for the analysis and reviewed other resources such as the Survey Diem website.

So yeah, I can craft an analysis that shows Survey Diem is the best, but the average sample buyer isn’t going to structure their search the same way.

Sticking to the “Facts”

I decided to take an alternative approach where I felt ChatGPT could perform better, which is to simply summarize what each company does for ID verification. Here’s the prompt I used:

“Summarize in a chart what each panel does for identity verification. Include columns specifically related to what the process is, when the verification happens, and whether it applies to all panel members. This information is often included in question 7 of the ESOMAR reports ("What form of validation do you use in recruitment to ensure that participants are real, unique, and are who they say they are? Describe this both in terms of the practical steps you take within your own organisation and the technologies you are using. Please try to be as specific and quantify as much as you can.")”

The prompt resulted in a table of just 12 panels out of the 52 examined. I asked ChatGPT why it only included 12 panels, and it ran the analysis again, this time resulting in 50 panels. I’m not sure why it didn’t do this the first time.

Here are the results. Thirteen panels are credited as having some form of ID verification beyond just email verification or phone validation. I am including the panel names because I want to give credit where credit is due. Kudos to these panels. I did not include the columns for when verification happens and to whom it applies because the model did not really make any distinctions among the panels. Regrettably, these 13 panels almost certainly make up less than 25 percent of survey panel revenue for the industry overall.

Verification Process Table
Panel Verification Process
GMO Research & AI AI fraud detection, email validation, CAPTCHA, identity verification via external databases
M3 Global Research Medical ID verification for healthcare panelists, CAPTCHA, digital fingerprinting
Macromill Southeast Asia SMS or email verification, ID card submission, CAPTCHA
Ola Surveys (Survey Diem) Driver's license or ID card scan, VPN detection, device fingerprinting
Prolific Identity verification using ID scans, AI fraud detection, CAPTCHA
RONIN International Digital fingerprinting, CAPTCHA, ID verification for high-value respondents
Sample Junction Device forensic API, Geo-IP checks, CAPTCHA, fraud scoring, unique ID validation
Savanta Imperium RegGuard fraud detection, relevant ID system, periodic quality checks
Symmetric Triple opt-in (email validation, follow-up survey, open-end verification), LinkedIn verification for B2B
EMpanel AI-based fraud prevention, CAPTCHA, manual identity verification
Markelytics Triple opt-in, CAPTCHA, behavioral fraud detection, address verification
Nomadic Insights Manual review of new registrations, Research Defender fraud detection, IP validation
Probity Research Digital fingerprinting, CAPTCHA, address verification, identity scoring

As for the rest, I thought I’d show what ChatGPT says the ESOMAR 37 reports say these companies do. Because ChatGPT could be wrong, I have removed the names of the panels. I don’t want to say panel x doesn’t do y and be wrong about it. (And no, the panels are not in alphabetical order in this table!)

Verification Process Table
Panel Verification Process
Panel 1Double opt-in verification, CAPTCHA, digital fingerprinting, phone number validation
Panel 2Geo-IP checks, CAPTCHA, unique ID assignment, profile reviews, T-Sign technology
Panel 3Email and phone verification, IP address tracking, behavior analysis
Panel 4Digital fingerprinting, behavioral analysis, email authentication, CAPTCHA
Panel 5IP tracking, browser fingerprinting, CAPTCHA, manual fraud checks
Panel 6Email validation, CAPTCHA, digital fingerprinting, behavioral screening
Panel 7Mobile phone verification, CAPTCHA, email authentication, behavioral tracking
Panel 8Two-step authentication, digital fingerprinting, behavioral analysis
Panel 9Biometric authentication, CAPTCHA, device fingerprinting, AI-driven identity verification
Panel 10Digital fingerprinting, IP validation, CAPTCHA, identity verification via email
Panel 11Double opt-in, CAPTCHA, behavioral tracking, digital fingerprinting
Panel 12Triple opt-in, CAPTCHA, deduplication system, anti-bot measures, location tracking
Panel 13Email verification, phone number validation, digital fingerprinting, CAPTCHA
Panel 14IP checks, digital fingerprinting, email verification, AI-based fraud detection
Panel 15Multi-step verification process, phone number confirmation, CAPTCHA
Panel 16Email verification, device fingerprinting, manual review of suspicious activity
Panel 17Two-factor authentication, device fingerprinting, automated fraud detection
Panel 18Email validation, CAPTCHA, duplicate checks, IP and geolocation tracking
Panel 19Mobile verification, CAPTCHA, IP address validation, duplicate email detection
Panel 20Email and mobile verification, CAPTCHA, AI-driven fraud checks
Panel 21Digital fingerprinting, text analytics, respondent-level tracking, hidden bot-detection text
Panel 22Business email validation for B2B panelists, CAPTCHA, behavioral analysis
Panel 23Mobile verification, CAPTCHA, digital fingerprinting, cross-checking for duplicate emails
Panel 24Machine learning fraud detection, email and phone verification, IP validation
Panel 25Two-step authentication, device fingerprinting, fraud detection via AI
Panel 26Double opt-in, email and phone verification, CAPTCHA, digital fingerprinting
Panel 27Device fingerprinting, CAPTCHA, email and phone verification, quality scoring
Panel 28Fraud detection algorithms, duplicate detection, CAPTCHA
Panel 29Triple-opt-in verification, biometric verification, blockchain data integrity, AI-powered screening
Panel 30Two-step verification, CAPTCHA, device fingerprinting, survey response analysis
Panel 31Identity verification through LinkedIn, CAPTCHA, behavioral tracking
Panel 32Email and phone verification, AI-based fraud detection, CAPTCHA
Panel 33Double opt-in, automatic de-duplication, quality score assignment, anomaly detection
Panel 34Mobile number verification, CAPTCHA, digital fingerprinting, manual quality checks
Panel 35AI-driven identity checks, CAPTCHA, IP tracking, manual fraud screening
Panel 36Geo-IP validation, CAPTCHA, duplicate detection, biometric analysis
Panel 37Double opt-in email confirmation, Geo-IP validation, anomaly detection, duplicate email/contact detection

Conclusion

ChatGPT can be good at summarizing content but sometimes needs corrective prompts (e.g., “Are you sure there are only 30 companies? I uploaded 52 files…) for basic errors and can struggle making judgments and recommendations, even with some prompting with a framework or value system. I worry about how a sample buyer might use an LLM to evaluate websites and other documents like the ESOMAR 37 reports. My recommendation to such shoppers is to stick to questions with verifiable answers and avoid making decisions solely on LLM recommendations or at least ask your questions a few different ways and see how results change. Know also that the model you use will affect results.

I find it fun to play around with ChatGPT and see what it can do. I would say at this point, if you’re using it as a shopping filter, just be a little skeptical of its recommendations. Happy researching and happy shopping.  

Previous
Previous

Can Fake IDs Fool Scanners? A Closer Look at Identity Verification in Survey Research

Next
Next

How to Prevent Survey Fraud