A University of Queensland study has shown large language models (LLMs) used in AI content moderation may be prone to subtle biases that undermine their neutrality. A team led by data scientist Professor Gianluca Demartini from UQ’s School of Electrical Engineering and Computer Science used persona prompting to test the tendency of AI chatbots to encode and reproduce political biases, and found significant behavioral shifts.
How AI bias can creep into online content moderation
More In Finance
-
Canada has some of the highest interchange fees in the world. Interchange fees are the fees businesses pay each time their customers pay by credit card. The average interchange fee in Canada is about 1.5 per cent of [...]
-
Main Street businesses that survived COVID-19 restrictions are now navigating a pandemic recovery where predicted changes in the retail industry have been accelerated by five to 10 years. The ability to adapt to these changes, coupled [...]
-
The big idea Consumers who see a product on sale being virtually touched are more engaged and willing to pay more than if the item is displayed on its own, according to a recent research paper [...]
-
Entrepreneurs, their associated startups and the subsequent growth of their companies have a vital impact on the health of our economy. In Canada, young adults have demonstrated a growing interest in entrepreneurship. Entrepreneurship has historically been narrowly [...]
-
Economics is broadly divided into macroeconomics and microeconomics. The big picture, macroeconomics, concentrates on the behavior of a national or a regional economy as a whole: the totals of goods and services, unemployment and prices. Then there’s a more [...]

