Scientists create AI agents to decode cause of religious conflict : The Tribune India

Join Whatsapp Channel

Scientists create AI agents to decode cause of religious conflict

LONDON: Using Artificial Intelligence, researchers at Oxford University have created psychologically realistic models of humans to test what causes religious conflict.

Scientists create AI agents to decode cause of religious conflict


London

Using Artificial Intelligence, researchers at Oxford University have created psychologically realistic models of humans to test what causes religious conflict.

The findings, published in The Journal for Artificial Societies and Social Stimulation, revealed that people are a peaceful species by nature. 

However, in a wide range of contexts they are willing to endorse violence — particularly when others go against the core beliefs which define their identity.

For the study, the researchers did not explicitly simulate violence, but instead focused on the conditions that enabled two specific periods of xenophobic social anxiety, that then escalated to extreme physical violence.

The researchers drew on the Northern Ireland Troubles spanning three decades from 1968 to 1998 and the 2002 Gujarat riots of India. 

Most people are generally familiar with AI that uses Machine Learning to automate human tasks like classifying something, such as tweets to be positive or negative etc., said one of the study authors Justin Lane from the University of Oxford.

"But our study uses something called multi-agent AI to create a psychologically realistic model of a human, for example — how do they think, and particularly how do we identify with groups? Why would someone identify as Christian, Jewish or Muslim etc. Essentially how do our personal beliefs align with how a group defines itself?"

 To create these psychologically realistic AI agents, the team used theories in cognitive psychology to mimic how a human being would naturally think and process information, to show how an individual's beliefs match up with a group situation.

They did this by looking at how humans process information against their own personal experiences. 

The researchers combined some AI models (mimicking people) that have had positive experiences with people from other faiths, and others that have had negative or neutral encounters. 

They did this to study the escalation and de-escalation of violence over time, and how it can, or cannot be managed.

To represent everyday society and how people of different faiths interact in the real world, they created a simulated environment and populated it with hundreds or thousands (or millions), of the human model agents. 

The only difference being that these 'people' all have slightly different variables  — age, ethnicity etc.

The findings revealed that the most common conditions that enable long periods of mutually escalating xenophobic tension occur when social hazards, such as outgroup members who deny the group's core beliefs or sacred values, overwhelm people to the point that they can no longer deal with them. 

It is only when people's core belief systems are challenged, or they feel that their commitment to their own beliefs is questioned, that anxiety and agitations occur, the study said. — IANS

Top News

Lok Sabha elections: Voting begins in 21 states for 102 seats in Phase 1

Lok Sabha elections 2024: Over 62 per cent voter turnout in Phase-1 amid sporadic violence Lok Sabha elections 2024: Over 62 per cent voter turnout in Phase-1 amid sporadic violence

Minor EVM glitches reported at some booths in Tamil Nadu, Ar...

Chhattisgarh: CRPF jawan on poll duty killed in accidental explosion of grenade launcher shell

Chhattisgarh: CRPF jawan on poll duty killed in accidental explosion of grenade launcher shell

The incident took place near Galgam village under Usoor poli...

Lok Sabha Election 2024: What do voting percentage and other trends signify?

Lok Sabha elections 2024: What do voting percentage and other trends signify

A high voter turnout is generally read as anti-incumbency ag...


Cities

View All