posted by organizer: l3s || 848 views || tracked by 3 users: [display]

BIAS 2019 : Special Issue on Social and Cultural Biases in Information, Algorithms, and Systems


When N/A
Where N/A
Submission Deadline Oct 31, 2018
Notification Due Dec 31, 2018
Categories    algorithms   bias   information   systems

Call For Papers

Special Issue on “Social and Cultural Biases in Information, Algorithms, and Systems”
Online Information Review (SSCI journal by Emerald Insight; 2016 Impact Factor: 1.534)

Computer algorithms and analytics play an increasing role in citizens’ lives, as they underlie the popular information services and “smart” technologies, which are rapidly being adopted across sectors of society, from transportation to education to healthcare . Algorithms allow the exploitation of rich and varied data sources, in order to support human decision-making and/or take direct actions; however, there are increasing concerns surrounding their transparency and accountability. There is growing recognition that even when designers and engineers have the best of intentions, systems relying on algorithmic processes can inadvertently result in serious consequences in the social world, such as biases in their outputs that can result in discrimination against individuals and/or groups of people. Recent cases in the news and media have highlighted the wider societal effects of data and algorithms, and have highlighted examples of gender, race and class biases in popular information access services.

It is important to note the complexity of the problem of social and cultural biases in algorithmic processes. For instance, recent research shows that word embeddings, a class of natural language processing techniques that enable machines to use human language in sensible ways, are quite effective at absorbing the accepted meaning of words (Caliskan et al., 2017). These algorithms also pick up on the human biases, such as gender stereotypes (e.g., associating male names with concepts related to career, and female names with home/family) and racial stereotypes (e.g., associating European-/African-American names with pleasant/unpleasant concepts) embedded in our language use. These biases are “accurate” in that they are comparable to those discovered when humans take the Implicit Association Test, a widely used measure in social psychology that reveals the subconscious associations between the mental representations of concepts in our memory (Greenwald et al., 1998).

The biases inherent in word embeddings provide a good illustration for the need to promote algorithmic transparency in information systems. Word embeddings are extensively used in services such as Web search engines and machine translation systems (e.g., Google Translate), which rely on the technique to interpret human language in real time. It may be infeasible to eradicate social biases from algorithms while preserving their power to interpret the world, particularly when this interpretation is based on historical and human-produced training data. In fact, another way of viewing such unconscious biases is as sources of ‘knowledge diversity’; what one thinks are the true facts of the world, and how one uses language to describe them, is very much dependent on local context, culture and intentions. An alternative approach would be to systematically trace and represent sources of ‘knowledge diversity’ in data sources and analytic procedures, rather than eliminate them (Giunchiglia et al., 2012). Such approaches would support accountability in algorithmic systems (e.g., a right to explanation of automated decisions, which to date has proven very challenging to implement). In addition, these approaches could facilitate the development of more “fair” algorithmic processes, which take into account a particular user’s context and extent of “informedness” (Koene et al., 2017).

The *purpose* of the special issue is to bring together researchers from different disciplines who are interested in analysing and tackling bias within their discipline, arising from the data, algorithms and methods they use. The theme of the special issue is social and cultural biases in information, algorithms, and systems, which includes, but is not limited to, the following areas:
- Bias in sources of data and information (e.g., datasets, data production, publications, visualisations, annotations, knowledge bases)
- Bias in categorisation and representation schemes (e.g., vocabularies, standards, etc.)
- Bias in algorithms (e.g., information retrieval, recommendation, classification, etc.)
- Bias in the broader context of information and social systems (e.g., social media, search engines, social networks, crowdsourcing, etc.)
- Considerations in evaluation (e.g., to identify and avoid bias, to create unbiased test and training collections, crowdsourcing, etc.)
- Interactions between individuals, technologies and data/information
- Considerations for data governance and policy

As the topic is highly interdisciplinary, we expect that this will be reflected by the submissions. We intend to invite authors from multiple disciplines, including data/information science, computer science, the social sciences, and psychology. The resulting special issue may also be of great interest to practitioners (e.g., in government, non-profit organisations, or companies) and educators (e.g., in digital literacy).

*Submission and Publication*

Authors are invited to submit original and unpublished papers. All submissions will be peer-reviewed and judged on correctness, originality, significance, quality of presentation, and relevance to the special issue topics of interest. Submitted papers should not have appeared in or be under consideration for another journal.
Instructions for authors:
Paper submission via
Please select the correct issue to submit to: “Social and Cultural Biases in Information, Algorithms, and Systems”.

*Important Dates*

- Submission Deadline: October 2018
- First Round Notification: December 2018
- Revision Due Date: February 2019
- Final Notification: April 2019
- Final Manuscript Due Date: June 2019
- Publication Date: July 2018

*Guest Editors*

Dr. Jo Bates, Information School, University of Sheffield, UK
Prof. Paul Clough, Information School, University of Sheffield, UK
Prof. Robert Jäschke, Humboldt-Universität zu Berlin, Germany
Prof. Jahna Otterbacher, Open University of Cyprus
Prof. Kristene Unsworth, Stockton University, New Jersey, USA

Related Resources

DSN 2019   Dependable Systems and Networks
IJIBM 2019   Call For Papers - International Journal of Information, Business and Management
SIPRO 2019   5th International Conference on Signal and Image Processing
ICDSST 2019   EWG-DSS 5th International Conference on Decision Support System Technology
IJGCA 2018   International Journal of Grid Computing & Applications
WLEC 2019   10th Annual Women's Leadership and Empowerment Conference
DIS 2019   Designing Interactive Systems Conference 2019
Luhmann@Dubrovnik 2019   CFP | Political communication: Observed with Niklas Luhmann's systems theory
Scopus-EI-AIACT 2019   2019 3rd International Conference on Artificial Intelligence, Automation and Control Technologies (AIACT 2019)
CCSIT 2018   8th International Conference on Computer Science and Information Technology