via University of Waterloo
Researchers from universities of Waterloo, Toronto, Yale, UPenn discuss AI and its application to their work
In an article published yesterday in the prestigious journal Science, leading researchers from the University of Waterloo, University of Toronto, Yale University and the University of Pennsylvania look at how AI (large language models or LLMs in particular) could change the nature of their work.
“What we wanted to explore in this article is how social science research practices can be adapted, even reinvented, to harness the power of AI,” said Igor Grossmann, professor of psychology at Waterloo.
Grossmann and colleagues note that large language models trained on vast amounts of text data are increasingly capable of simulating human-like responses and behaviours. This offers novel opportunities for testing theories and hypotheses about human behaviour at great scale and speed.
Traditionally, social sciences rely on a range of methods, including questionnaires, behavioral tests, observational studies, and experiments. A common goal in social science research is to obtain a generalized representation of characteristics of individuals, groups, cultures, and their dynamics. With the advent of advanced AI systems, the landscape of data collection in social sciences may shift.
“AI models can represent a vast array of human experiences and perspectives, possibly giving them a higher degree of freedom to generate diverse responses than conventional human participant methods, which can help to reduce generalizability concerns in research,” said Grossmann.
“LLMs might supplant human participants for data collection,” said UPenn psychology professor Philip Tetlock. “In fact, LLMs have already demonstrated their ability to generate realistic survey responses concerning consumer behaviour. Large language models will revolutionize human-based forecasting in the next 3 years. It won’t make sense for humans unassisted by AIs to venture probabilistic judgments in serious policy debates. I put an 90% chance on that. Of course, how humans react to all of that is another matter.”
While opinions on the feasibility of this application of advanced AI systems vary, studies using simulated participants could be used to generate novel hypotheses that could then be confirmed in human populations.
But the researchers warn of some of the possible pitfalls in this approach – including the fact that LLMs are often trained to exclude socio-cultural biases that exist for real-life humans. This means that sociologists using AI in this way couldn’t study those biases.
Professor Dawn Parker, a co-author on the article from the University of Waterloo, notes that researchers will need to establish guidelines for the governance of LLMs in research.
“Pragmatic concerns with data quality, fairness, and equity of access to the powerful AI systems will be substantial,” Parker said. “So, we must ensure that social science LLMs, like all scientific models, are open-source, meaning that their algorithms and ideally data are available to all to scrutinize, test, and modify. Only by maintaining transparency and replicability can we ensure that AI-assisted social science research truly contributes to our understanding of human experience.”
Original Article: AI could replace humans in social science research
More from: University of Waterloo | University of Toronto | Yale University | University of Pennsylvania
The Latest Updates from Bing News
Go deeper with Bing News on:
AI social science research
- Former OpenAI leader says safety has 'taken a backseat to shiny products' at the AI company
wrote in a series of posts on the social media platform X that he joined the San Francisco-based company because he thought it would be the best place to do AI research. "However, I have been ...
- How AI And Technology Can Impact Mental Health Care
AI has the potential to offer better, more accessible and personalized solutions for mental health care. However, these apps are not a substitute for professional care, and more work and research are ...
- AI In Life Science Analytics Market to Grow at 12.7% CAGR Through 2033
The AI in Life Science Analytics Market is projected to reach approximately USD 5.6 billion by 2033, growing from USD 1.7 billion in 2023, at a compound annual growth rate (CAGR) of 12.7% from 2024 to ...
- Detecting influence campaigns on X with AI and network science
In the age of generative-AI and large language models (LLMs), massive amounts of inauthentic content can be rapidly broadcasted on social media platforms. As a result, malicious actors are becoming ...
- As new tools flourish, AI ‘fingerprints’ on scientific papers could damage trust in vital research
Experts are warning that the “fingerprints” of generative artificial intelligence (GenAI) can be found in scientific papers, including peer-reviewed ones. View on euronews ...
Go deeper with Bing News on:
Social science LLMs
- How Does ChatGPT Think?
Researchers are striving to reverse-engineer artificial intelligence and scan the “brains” of LLMs to deduce the how any why of that they are doing ...
- Researchers find LLMs are easy to manipulate into giving harmful information
A team of AI researchers at AWS AI Labs, Amazon, has found that most, if not all, publicly available Large Language Models (LLMs) can be easily tricked into revealing dangerous or unethical ...
- Majority of Humans Fooled by GPT-4 in Turing Test, Scientists Find
OpenAI's GPT-4 is so lifelike, it can trick more than 50 percent of human test subjects into thinking they are talking to a person.
- 8 Ethical Challenges For Generative AI
The rapid advancement of generative AI technologies over the past year marks a significant leap in technological progress.
- Cohere Showcases How LLMs Are Transforming Retail Customer Experiences
May 17, 2024 — Cohere has published a new article discussing the transformative potential of large language models (LLMs) in the retail industry. In the post, David Stewart, Robin Gainer, and Sam ...