The Workshop on Online Abuse and Harms - WOAH 2020

Страна: Dominica

Город: Virtual

Тезисы до: 01.09.2020

Даты: 20.11.20 — 20.11.20

Е-мейл Оргкомитета: sharedexploration AT



Digital technologies have brought myriad benefits for society, transforming how people connect, communicate and interact with each other. However, they have also enabled harmful and abusive behaviours, from interpersonal aggression to bullying and hate speech, to reach large audiences and for their negative effects to be amplified. The negative effects are further compounded as marginalised and vulnerable communities are disproportionately at the risk of receiving abuse. As policymakers, civil society and tech companies devote more resources and time to tackle online abuse, there is a pressing need for scientific research that rigorously investigates how we define harms, how it is detected, moderated and countered.

Call for Papers:

For the fourth edition of the Workshop on Online Abuse and Harms (4th WOAH!) our theme is: Social Bias and Unfairness in Online Abuse Detection. This iteration of the workshop will bring together NLP researchers, members of groups that are likely to be discriminated against, ethics researchers, sociologists, technologists working on fairness and accountability, and legal experts to discuss the nuances of justice, fairness, accountability, and ethics of autonomous systems. The workshop will have four components:
1. Regular paper submissions, both short (4 pages) and long (8 pages).
2. Submissions from civil society (5 to 20 pages).
3. Shared Exploration; to get researchers to creatively engage in the problem area.
4. a multidisciplinary panel discussion.

We continue to emphasize the need for inter-, cross- and anti- disciplinary work on online abuse and harms, and invite paper submissions from a range of fields. These include but are not limited to: NLP, machine learning, computational social sciences, law, politics, psychology, network analysis, sociology and cultural studies. Additionally, in this iteration we invite civil society, in particular individuals and organisations working with women and marginalised communities who are often disproportionately affected by online abuse, to submit reports, case studies, findings, data, and to record their lived experiences. We hope that through these engagements we can develop computational tools which address the issues faced by those on the front-lines of tackling online abuse.

For research papers, we invite long (8 pages) and short (4 pages) academic/research papers on any of the following general topics.
Related to developing computational models and systems:
> NLP models and methods for detecting abusive language online, including, but not limited to hate speech, gender-based violence, cyberbullying etc.
> Application of NLP tools to analyze social media content and other large data sets
> NLP models for cross-lingual abusive language detection
> Computational models for multi-modal abuse detection
> Development of corpora and annotation guidelines
> Critical algorithm studies with a focus on content moderation technology
> Human-Computer Interaction for abusive language detection systems
> Best practices for using NLP techniques in watchdog settings
> Submissions addressing interpretability and social biases in content moderation technologies

Веб-сайт конференции: