Ritesh Kumar
2020
Proceedings of the Second Workshop on Trolling, Aggression and Cyberbullying
Ritesh Kumar
|
Atul Kr. Ojha
|
Bornini Lahiri
|
Marcos Zampieri
|
Shervin Malmasi
|
Vanessa Murdock
|
Daniel Kadar
Proceedings of the Second Workshop on Trolling, Aggression and Cyberbullying
Evaluating Aggression Identification in Social Media
Ritesh Kumar
|
Atul Kr. Ojha
|
Shervin Malmasi
|
Marcos Zampieri
Proceedings of the Second Workshop on Trolling, Aggression and Cyberbullying
In this paper, we present the report and findings of the Shared Task on Aggression and Gendered Aggression Identification organised as part of the Second Workshop on Trolling, Aggression and Cyberbullying (TRAC - 2) at LREC 2020. The task consisted of two sub-tasks - aggression identification (sub-task A) and gendered identification (sub-task B) - in three languages - Bangla, Hindi and English. For this task, the participants were provided with a dataset of approximately 5,000 instances from YouTube comments in each language. For testing, approximately 1,000 instances were provided in each language for each sub-task. A total of 70 teams registered to participate in the task and 19 teams submitted their test runs. The best system obtained a weighted F-score of approximately 0.80 in sub-task A for all the three languages. While approximately 0.87 in sub-task B for all the three languages.
Developing a Multilingual Annotated Corpus of Misogyny and Aggression
Shiladitya Bhattacharya
|
Siddharth Singh
|
Ritesh Kumar
|
Akanksha Bansal
|
Akash Bhagat
|
Yogesh Dawer
|
bornini lahiri
|
Atul Kr. Ojha
Proceedings of the Second Workshop on Trolling, Aggression and Cyberbullying
In this paper, we discuss the development of a multilingual annotated corpus of misogyny and aggression in Indian English, Hindi, and Indian Bangla as part of a project on studying and automatically identifying misogyny and communalism on social media (the ComMA Project). The dataset is collected from comments on YouTube videos and currently contains a total of over 20,000 comments. The comments are annotated at two levels - aggression (overtly aggressive, covertly aggressive, and non-aggressive) and misogyny (gendered and non-gendered). We describe the process of data collection, the tagset used for annotation, and issues and challenges faced during the process of annotation. Finally, we discuss the results of the baseline experiments conducted to develop a classifier for misogyny in the three languages.
Search
Co-authors
- Atul Kr. Ojha 3
- Marcos Zampieri 2
- Shervin Malmasi 2
- Bornini Lahiri 1
- Vanessa Murdock 1
- show all...
Venues
- TRAC3