- Open Access
- Authors : Shrii Sudhan , Narain C. A , Tirugnanam , S. Rajesh Kumar
- Paper ID : IJERTV11IS120103
- Volume & Issue : Volume 11, Issue 12 (December 2022)
- Published (First Online): 23-12-2022
- ISSN (Online) : 2278-0181
- Publisher Name : IJERT
- License: This work is licensed under a Creative Commons Attribution 4.0 International License
Social Media Toxicity Analyser
Narain C A , Tirugnanam ,Shrii Sudhan , Mr. S. Rajesh Kumar
1Department of Computer Science and Engineering, National Engineering College, Kovilpatti, Tamil nadu
2 Assistant Professor Department of Computer Science and Engineering, National Engineering College, Kovilpatti, Tamil Nadu
Abstract: Cyberbullying and online abuse have been continuously increasing at an alarming rate. This is detrimental to the mental state of youngsters and one of the biggest factors leading to mental depression. Manually determining the toxicity of the comments in the avalanche of data generated daily is an impossible task. Automating the detection and censorship of such toxic comments by social media platforms can go a long way in solving this issue. But detection of toxic comments is a very daunting task because various factors such as context, perception, vocabulary, etc matters here. The motto of our project is to find the toxicity of their media profiles by the activities of the respective person involved in it. The toxicity can be identified by the captions, bio, and other wordings a person uses on social media like Twitter. The whole profile of the person is analyzed and compared with the predefined data and a report is generated containing the toxicity of the social media profile. This report can be used by the cyber-security department to find whether that person is toxic or bullying anyone on the internet. Through this, we can decently manage social media content.
Keywords:- Cyber bullying, Online abusing, abusing. Bullying, cyber attack
Nowadays social media are becoming more toxic like fake accounts, usage of abusive words, and posting content that hurts other users. Therefore, to reduce toxicity, social media toxicity has been introduced in which the toxicity detector generates a report on the toxicity level of the given account or user id. The project's scope is to verify the social media account by scrapping all the posts, and comments posted by the user and detecting toxicity, and providing a report for the toxicity level in the account. It leads to harmless social media usage. The architecture was well-designed and organized. It makes social media more safe and more secure. Authentication works well.
The main objective of the project is to find the toxicity level of a particular given user's social media accounts, especially for a Twitter account. For doing this, we have used python language as a platform and imported the snscrape library module, JSON library, and flask library module. The project consists of an HTML page that gets the user's Twitter account name and the total number of last tweets to be retrieved to get the contents of posts and comments by the user. These details have been sent to the python flask login page function using an HTTP POST/GET request. The details are given to the snscrape's twitterScraper.TwitterUserScraper module and tweet contents up to the specified number will be fetched and append the data into a tweets array in python. The tweets will be saved into a newly created result.json file in the current location of the code. The data will be parsed through the for loops and it checks for the swear words from the dataset.json. If it matches the swear words, then the matched word will be added to the swear category strings and counts will be taken accordingly. The count of swear words based on category as specified in the dataset file and categorized swear words will be sent to the success function of the python module and this function consists of an embedded HTML file and connect the sentence with that embedded code. Finally, display the contents in the new HTML success page for the user.
Fig 3.1 Home page
Fig 3.2 Result Page
Fig 3.3 Saved Tweet contents in files for particular user
Fig 3.4 Dataset file for data checking
Fig 3.4 Invalid Username Check
Kazi Saeed Alam; Shovan Bhowmik; Priyo Ranjan Kundu Prosun et al  (2019) – Proposed This research is to automate the detection process of offensive language or cyberbullying. The main aim is to build a single and double ensemble-based voting model to classify the contents into two groups: `offensive' or `non-offensive'. For this purpose, four machine learning classifiers have been chosen and three ensemble models with two different feature extraction techniques combined with various n-gram analysis on a dataset extracted from Twitter.
Divya Iyyani; Keshin Jani; Swati Mali et al  (2018) – outlined troll has been an issue in social media for a long time. This explores the currently existing Anti-trolling systems, their methodologies and technological challenges. In particular, we consider machine learning and existing expert systems that detect and prevent trolls. The paper concludes with a discussion of the issues faced by this technology, the current functioning and a few suggestions along with a modified architecture and the vision of a trolling-free internet .
Cheniki Abderrouaf; Mourad Oussalah et al  (2019) – presented The main content in this topic is by leveraging novel natural language processing, machine learning, and feature engineering techniques. The proposed approach advocates a classification-like technique that makes use of a special data design procedure. The latter enforces a balanced training scheme by exploring the negativity of the original dataset.
Kanwal Yousaf; Tabassam Nawaz et al  (2022) – expressed the exponential growth of videos on YouTube has attracted billions of viewers among which the majority belongs to a young demographic. Malicious uploaders also find this platform as an opportunity to spread upsetting visual content, such as using animated cartoon videos to share inappropriate content with children. Therefore, an automatic real-time video content filtering mechanism is highly suggested to be integrated into social media platforms.
Nadira Boudjani, Yannis Haralambous, Inna Lyubareva et al  (2020)- proposed supervised approach has been proposed for toxic comment classification for the French language. We choose a set of features proposed for toxic comment detection for English and use it for French toxic comment detection. Our approach is based on N-gram features, linguistic features and a dictionary of insulting words and expressions.
Chetna Sharma ,Rahul Ramakrishnan,Ayusha Pendse,Priya Chimurkar Kiran T. Talele: et al  (2020) presented cyberbullying is a major issues in social media. The work here is to identifying cyberbullying at is origin, meaning when it is being drafted in real-time. Identifying traces of cyberbullying before the content is uploaded on the internet can help reduce circulation of hurtful messages.
Muhammad Amien Ibrahim; Noviyanti Tri Maretta Sagala; Samsul Arifin; Rinda Nariswari; Nerru Pranuta Murnaka et al  (2022) presented the main aim is to classify hate speech, abusive language, and normal messages on Indonesian Twitter. Several machine learning models, such as logistic regression and BERT models, are utilized to accomplish text classification tasks.
Mahamat Saleh Adoum Sanoussi; Chen Xiaohua; George K. Agordzo; Mahamed Lamine Guindo; Abdullah MMA Al Omari et al  (2022) Identifying hate speech on social media has become increasingly crucial for society. It has been shown that cyberbul-lying significantly affects the social tranquillity of the Chadian population, mainly in laces of conflict. This article aims to detect hate speech for texts written in lingua franca, a mix of the local Chadian and French languages.
Jayant Singh; Kishorjit Nongmeikapam et al  (2011) presented aggression and related activities such as trolling peoples, harassing online involves hate comments in various forms. There are numerous such cases coming in present time and sites respond by closing down their remark areas. After the introduction of Machine Learning and having data in massive amounts now its quite logical to build a tool which can tackle this situation.
Paraskevas Tsantarliotis; Evaggelia Pitoura; Panayiotis Tsaparas et al  (2014) took a novel approach to the trolling problem: our goal is to identify the targets of the trolls, so as to prevent trolling before it happens. We thus define the troll vulnerability prediction problem, where given a post we aim at predicting whether it is vulnerable to trolling. Towards this end, we define a novel troll vulnerability metric of how likely a post is to be attacked by trolls, and we construct models for predicting troll-vulnerable posts, using features from the content and the history of the post.
Lu Cheng , Ahmadreza Mosallanezhad , Yasin N. Silva , Deborah L. Hall ,Huan Liu et al  (2021) Increased social media use has contributed to the greater prevalence of abusive, rude, and offensive textual comments. Machine learning models have been developed to detect toxic comments online, yet these models tend to show biases against users with marginalized or minority identities.
Paula Reyero Lobo, Enrico Daga, Harith Alani et al  (2022) presented the use of a knowledge graph to help in better understanding such toxic speech annotation issues. Our empirical results show that 3% in a sample of 19k texts mention terms associated with frequently attacked gender and sexual orientation groups that were not correctly identified by the annotators.
B.S. Sahana; G. Sandhya; R.S. Tanuja; Sushma Ellur; A. Ajina et al  (2022) expressed unlike most of the work in toxic content detection where the nature of toxicity is determined, we treat the detection of toxic content as a binary classification task. Here, we have explored Support Vector Machine, Boosting and deep neural networks for classification. We have trained the model on twitter datasets. With a goal of better predictive performance, our approach uses a majority voting ensemble to aggregate the predictions of individual classifiers.
Ishan Sanjeev Upadhyay, KV Aditya Srivatsa, Radhika Mamidi et al  (2022) presented a dataset for toxic positivity classification from Twitter and an inspirational quote website. We then perform benchmarking experiments using various text classification models and show the suitability of these models for the task.
Yau-Shian Wang, Yingshan Chang et al  (2022) outlined the generative variant of zero-shot prompt-based toxicity detection with comprehensive trials on prompt engineering. We evaluate on three datasets with toxicity labels annotated on social media posts. Our analysis highlights the strengths of our generative classification approach both quantitatively and qualitatively. Interesting aspects of self-diagnosis and its ethical implications are discussed.
The act of bullying is a punishable offense, people on the internet are covering their faces and performing these kinds of actions. This should be condemned and thrown away from the internet and should be punished. Through this web application anyone can scrape the twitter data and find the toxicity of a particular user id and can avoid them. This helps in finding out peoples behavior on the internet. The project will be improved with more features like reporting and intimating the cyber security.
 Yau-Shian Wang, Yingshan Chang: (2022) – Toxicity Detection with Generative Prompt-based Inference