Skip to content

The Cultural AI Lab's Repo for the Artificially Correct Hackathon

Notifications You must be signed in to change notification settings

valevo/artificially_correct_challenge

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

31 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Does Bias in Collections and Archives Survive Translation and Multilingualism?

The Cultural AI Lab's Repo for the Goethe Institute's Artificially Correct Hackathon.

CulturalAI Logo

https://www.cultural-ai.nl

Background

In the Cultural AI lab and in collaboration with the National Museum of World Cultures (NMvW), we have been developing SABIO (the SociAl BIas Observatory): a tool and visual interface for exploratory analysis of bias and patterns of bias in museum collections and heritage archives. SABIO emphasises that researching bias is a social act and therefore, instead of automatically detecting [and judging] potential instances of bias, it is designed to provide access to paths through cultural datasets that lead users to discover bias in ways they might not [be able to] anticipate. To this end, SABIO consists of an extensible suite of parameterised algorithms that measure variables likely to correlate with bias and (re-)orginasises the given collection according to these measurements and provides details about the origin individual values.

Description

So far, SABIO has been developed and tested exclusively on the NMvW's collection, a monolingual Dutch archive of around 800k objects with descriptions, titles and other meta-properties. At the same time, all of the design choices for SABIO have been made with the aim of maximal language independence and we would like to put this to the test. We therefore invite participants to experiment with and extend SABIO, or at least the algorithms it uses, in cross- and multilingual contexts. Potential tasks include but are by no means limited to:

  1. Findinging and analysing an originally multilingual collection in SABIO and identifying patterns of bias across the languages present and how they translate.
  2. Similarly, machine translating a monolingual translation and analysing with the of SABIO how biases are transmitted and whether they survive automatic translation.
  3. Conceptual research on how paths in search of bias might be different across the languages present in a collection, and implementations based upon conclusions.
  4. Designing extensions to SABIO itself or to its algorithms that are tailored to comparison and highlighting of bias across languages.

Considerations

  • Machine learning algorithms and their underlying datasets, such as the ones used by SABIO, require considerable computational resources and time; especially experimentation with such algorithms on large collections can be hindered by lack of resources
  • As of now, SABIO's data format is rather specific to the NMvW collection and thus requires specific properties to be present. We are working in the meantime on more general data specifications in order to fit other collections but this will likely restrict the choice of collection for the hackathon.
  • Collections are often subject to copyright restrictions or not prepared for digital processing which might make it difficult and time-consuming to gain access.
  • SABIO itself is a server-based application; even though it is entirely open-source, actually extending it or creating parallel versions requires server-side programming and resources for hosting.

Resources

We provide SABIO's search and visual interface to the NMvW collection (the collection is freely accessible via the NMvW's own interface, see the link above). We will also facilitate inputting collections selected by participants into SABIO for analysis, since SABIO resides is hosted from our own server. Additionally, we provide SABIO's backend, written in Python on our GitHub repository. This includes data preprocessing scripts for a collection (given in a specific format), a search interface and the scoring algorithms to reorganise a dataset in terms of bias.

Skills

Participants should have experience with either:

  • Basic statistics and machine learning, specifically with natural language applications & (Python) programming; e.g. constructing word embeddings or defining and computing semantic association scores.
  • Dealing with databases and digital archives, their intricacies and (pre-)processing them for machine learning; e.g. building a database cleaning pipeline and basic natural language preprocessing on large sets of text.
  • Performing research (historical, sociological and similar) on heritage collections, archives and the like, in particular with the use of search interfaces and visualisation tools.
  • Research into bias of diverse kinds, especially in contexts of cultural hertiage, and defining and analysing it in given aspects of society.

Optimally, the team would be made up of a mix of expertises in researching bias in cultural collections on the one hand and language processing on the other.

About

The Cultural AI Lab's Repo for the Artificially Correct Hackathon

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published