31 August 2022

English bias in computing: Images to the rescue

Machine Learning

Training sets for Machine Learning are normally created in English. This introduces cultural bias. A new image-grounded benchmark developed by researchers from the University of Copenhagen enables a more diverse approach.

Image based machine learning could widen the range of image recognition apps
Getty Images

So many languages; and yet English is allowed almost total domination when it comes to the AI technology Machine Learning (ML). If, for example, researchers are training a computer in grasping the content of a random text, the training samples will typically be in English.

“This introduces a significant unintentional cultural bias. Even after extensive training, the machine will never have been exposed to bull taming in India, to Chinese hot pot cooking, or to other phenomena which are familiar to millions of people, but just happen to lie outside the native English-speaking horizon,” says PhD researcher Emanuele Bugliarello, Department of Computer Science (DIKU), University of Copenhagen.

In a truly intercultural effort Emanuele Bugliarello and colleagues from a range of countries have created a new tool which encourages a more diverse approach. IGLUE (Image-Grounded Language Understanding Evaluation), as they have named the tool, is a benchmark which allows for scoring the efficiency of an ML solution in 20 languages (rather than English alone).

 

Their scientific article introducing IGLUE has been accepted for publication at the upcoming International Conference on Machine Learning (ICML), one of the top conferences in the field.

Volunteers provided culture specific images

How can a new benchmark make a difference?

“When ML research teams create new solutions, they are always highly competitive. If another group has succeeded in solving a given ML task with 98 per cent accuracy, you will try to get 99 per cent and so forth. This is what drives the field forward. But the downside is that if you don’t have a proper benchmark for a given feature, it will not be prioritized. This has been the case for multimodal ML, and IGLUE is our attempt to change the scene,” says Emanuele Bugliarello.

Basing training on images is standard in ML. However, the images are usually “labelled”, meaning bits of text will accompany each image, aiding the learning process of the machine. While the labels are normally in English, IGLUE covers 20 typologically diverse languages, spanning 11 language families, 9 scripts, and 3 geographical macro-areas.

A part of the images in IGLUE are culture specific. These images were obtained through a mail campaign. The researchers asked volunteers in geographically diverse countries  to provide images and texts in their natural language and preferably about things that were important in that country.

Overwhelmed by positive reactions

The current lack of multimodal ML does have practical implications, Emanuele Bugliarello explains:

“Let us say you have a food allergy, and you have an app which can tell you if the problematic ingredients are present in a meal. Finding yourself at a restaurant in China, you realize the menu is all in Chinese but has pictures. If your app is good, it may translate the picture into a recipe – but only if the machine was exposed to Chinese samples during training.”

In other words, non-English speakers get a poorer version of ML-based solutions:

“The performance of many top ML solutions will drop instantly, as they become exposed to data from non-English speaking countries. And notably, the ML solutions miss out on concepts and ideas that are not formed in Europe or North America. This is something which the ML research community needs to address,” Emanuele Bugliarello says.

Fortunately, many colleagues have seen the light, Emanuele Bugliarello notes:

“This all began a few years ago when we wrote a paper for the EMNLP conference (Empirical Methods in Natural Language Processing, a top conference in the field, ed.). We just wanted to point to an issue, but were soon overwhelmed with interest, and much to our surprise our contribution was selected as Best Long Paper. People clearly saw the problem, and we were encouraged to do more.”

May help the visually impaired

Sometimes the current success almost feels like a burden, Emanuele Bugliarello admits:

“As a public university, we do have limited resources. We cannot pursue all aspects of this huge task. Still, we can see that other groups are joining in. We can also feel interest from the large tech corporations. They are strongly engaged in ML and are beginning to realize how English bias can be a problem. Obviously, they are not happy to see the performance of their solutions dropping significantly when applied outside English-language contexts.”

Despite the positive developments, Emanuele Bugliarello does not allow himself to get carried away. Asked how close we are to achieving non-biased Machine Learning, he answers:

“Oh, we are very far away!”

Still, this is not only about cultural equality:

“The methodology behind IGLUE may find several applications. For example, we hope to improve solutions for visually impaired. Tools exist, which help visually impaired in following the plot of a movie or another type of visual communication. These tools are currently far from perfect, and I would very much like to be able to improve them. This is a bit further into the future, though,” Emanuele Bugliarello concludes.

 

The paper “IGLUE: A Benchmark for Transfer Learning across Modalities, Tasks, and Languages” was accepted in the “Proceedings of The Thirty-ninth International Conference on Machine Learning” (ICML). The conference took place in Baltimore, USA, July 17-23.

Contact

Emanuele Bugliarello
PhD Researcher
Natural Language Processing
Department of Computer Science
University of Copenhagen
Mail: emanuele@di.ku.dk
Mobil: +45 93 51 60 61

Kristian Bjørn-Hansen
Journalist
Faculty of Science
University of Copenhagen
Mail: kbh@science.ku.dk 
Phone: +45 93 51 60 02

Topics

More stories