Skip to content

Software and Resources for Mitigating Online Gender Based Violence in India

License

Notifications You must be signed in to change notification settings

hardik-pratap-singh/Uli

 
 

Repository files navigation

Uli : Reclaim your Online Space

Uli is a browser plugin that :

  • De-normalizes the everyday violence that people of marginalized genders experience online in India
  • Provides tools for relief and collective response.

It is an attempt to invert the top-down logics of platform moderation and center the experiences of those subject to online gender-based violence.

🎉 Contribution Pathways 🎉

We cherish diversity of experiences and perspectives. It adds value to our work. To this end, we strongly encourage candidates who find alignment with the project and are driven to learn, to contribute to Uli. There are both code and no-code issues that you can contribute to.

To contribute effectively, we recommend doing some of these:

  • Peruse our Wiki. It will help you navigate our repository, and adhere to our standards for contributions.
  • We've labeled beginner frienly issues with good first issue.
  • Read our Setup Guides on the Uli Wiki or watch a video tutorial
  • Join the our Slack to interact with the team and get any clarificatios. Introduce yourself in the #introductions channel and feel free to discuss any Hacktoberfest-related questions in the #issue_uli_hacktoberfest channel.

Technologies we use

If you are new to any of these, we've created a learning guide for you.

Repository Structure

Directory Description
browser extension Uli browser extension that helps moderate and mitigate online gender based violence on twitter. All hacktoberfest work is limited to this directory
annotators a web app to annotate tweets, unmaintained
slur-replacement Python notebook that documents our exact and approximate slur replacement techniques
ogbv-ml-rest Hosted access for our machine learning model

Quick Guide

  1. Motivation
  2. Approach
  3. Roadmap
  4. Contribute
  5. Contact

Motivation

The graphic narrative titled ‘Personal (Cyber) Space’ published in 2016 by Parthasarthy and Malhotra narrates an experience of a young internet user. The animated short comic hosted by Kadak, a South Asian women collective, asks: ‘If one says something, there’s the fear of hateful response. But if one doesn’t say something, isn’t that silence counterproductive?’ only to end with the question, ‘so what does one say?’

Violence, abuse, and hate speech on the web has become pervasive to one’s experience of social media and the existing scholarship suggests that it is those situated at the margins who are worst affected. People of marginalized genders in India are disproportionately affected. Simultaneously, the business models of social media platforms skew the incentives against protecting users who use social media in non-dominant languages. Uli is an attempt to build tools to protect and enable collective response for social media users of marginalizd genders.

Approach

The problem of online violence encompasses within itself legal, political, social, cultural and technological complexities that make any easy solution impossible. This overdetermined nature mandates that we seek solutions from multiple avenues. As with all Tattle projects, we don't expect technology to provide all the answers, but for it to be intertwined in human action.

Specifically around redaction, the project borrows from feminist approaches to Machine Learning technology and aims to intervene into the ongoing debate around content moderation. The existing algorithmic approaches to automated content moderation strategies are generally biased towards English-language content paying very limited attention to social, cultural and linguistic diversity elsewhere. Moreover, the existing approaches understand moderation through a binary logic of: leave content up or remove it. With multiple political and legal implications emerging from these biases, the existing approaches threaten to pose more problems rather than solving them. With this tool, the project aims to redress these problems and find creative ways in which moderation can empower multiple users, especially the ones that are most affected.

We started the projeect through a period of qualitative data collection methods and participatory analysis. Based on the needs articulated by gender rights activists and researchers we focused on building the following features:

  • Detection/ filtering of abuse
  • Tools for archiving locally and through email.
  • Localized resources for understanding the effects of online gender based violence.
  • Invoking networks for action on abusive content.

Based on feedback from the beta on potential misuse of the feature, it has been removed from the current version till more checks and balances can be built in.

The ultimate aim of the project is to envision creative and collective responses to the structural problem of violence experienced online and help build solidarity and shared understanding while empowering users to take back control of their digital experience.

Situating machine learning:

Machine learning based approaches are a commonly used technique to automate decision making when the volume of data is large. To put it briefly, machine learning works by finding patterns in existing data to ascribe a value to future queries. Instead of telling an algorithm what to do, in machine learning, the algorithm figures out what to do based on the data it is fed. The data used to train a machine learning system as well as the algorithm used to classify the data, can encode social beliefs and values. These are perpetuated in the performance of the machine learning systems.

The moderation decisions of social media platforms often make international news. Some decisions can be attributed to error. Machine learning system, like every prediction system, makes errors. But some decisions reflect the social values in the data and algorithms behind the model. So, what many communities find harmful may not be harmful as per the guidelines set by social media platforms. Machine learning tools can also be designed to reflect the values of those at the forefront of tackling violence, to protect those who will be at the receiving end of the violence. This is precisely the goal of our project.

The ML model is based on needs articulated by communities, rather than the priorities of powerful institutions. We are working to publish our methodology, annotation guidelines, datasets, and the limitations in the dataset. Our goal remains to make the models interpretable to the users of the plugi. We believe this will help to raise awareness about content moderation systems as well as gender-based violence online.

Contributing

We've made a list of good first issues to get started on. You can also track the project here Find an issue or domain that interests you and reach out to us. Learn More

Contact

For more details on this project please send an email to one of the following email IDs:

Funding:

The pilot of the project was managed by the Centre for Internet and Society and Tattle Civic Tech. It was funded by Omidyar Network India as part of their Digital Society Challenge grant. In addition to revenue from Tattle's other projects, Uli is supported by Mozilla's Digital Society Challenge.

Website

Visit the Uli website at https://uli.tattle.co.in/

About

Software and Resources for Mitigating Online Gender Based Violence in India

Resources

License

Code of conduct

Security policy

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 45.7%
  • JavaScript 28.3%
  • MDX 18.8%
  • Python 6.6%
  • CSS 0.2%
  • HTML 0.2%
  • Other 0.2%