ترغب بنشر مسار تعليمي؟ اضغط هنا

Sieving Fake News From Genuine: A Synopsis

107   0   0.0 ( 0 )
 نشر من قبل Shahid Alam
 تاريخ النشر 2019
  مجال البحث الهندسة المعلوماتية
والبحث باللغة English




اسأل ChatGPT حول البحث

With the rise of social media, it has become easier to disseminate fake news faster and cheaper, compared to traditional news media, such as television and newspapers. Recently this phenomenon has attracted lot of public attention, because it is causing significant social and financial impacts on their lives and businesses. Fake news are responsible for creating false, deceptive, misleading, and suspicious information that can greatly effect the outcome of an event. This paper presents a synopsis that explains what are fake news with examples and also discusses some of the current machine learning techniques, specifically natural language processing (NLP) and deep learning, for automatically predicting and detecting fake news. Based on this synopsis, we recommend that there is a potential of using NLP and deep learning to improve automatic detection of fake news, but with the right set of data and features.



قيم البحث

اقرأ أيضاً

In dealing with altered visual multimedia content, also referred to as fake news, we present a ready-to-deploy extension of the current public key infrastructure (PKI), to provide an endorsement and integrity check platform for newsworthy visual mult imedia content. PKI, which is primarily used for Web domain authentication, can directly be utilized with any visual multimedia file. Unlike many other fake news researches that focus on technical multimedia data processing and verification, we enable various news organizations to use our developed program to certify/endorse a multimedia news content when they believe this news piece is truthiness and newsworthy. Our program digitally signs the multimedia news content with the news organizations private key, and the endorsed news content can be posted not only by the endorser, but also by any other websites. By installing a web browser extension developed by us, an end user can easily verify whether a multimedia news content has been endorsed and by which organization. During verification, our browser extension will present to the end user a floating logo next to the image or video. This logo, in the shape of a shield, will show whether the image has been endorsed, by which news organization, and a few more pieces of essential text information of the news multimedia content. The proposed system can be easily integrated to other closed-web system such as social media networks and easily applied to other non-visual multimedia files.
Disinformation through fake news is an ongoing problem in our society and has become easily spread through social media. The most cost and time effective way to filter these large amounts of data is to use a combination of human and technical interve ntions to identify it. From a technical perspective, Natural Language Processing (NLP) is widely used in detecting fake news. Social media companies use NLP techniques to identify the fake news and warn their users, but fake news may still slip through undetected. It is especially a problem in more localised contexts (outside the United States of America). How do we adjust fake news detection systems to work better for local contexts such as in South Africa. In this work we investigate fake news detection on South African websites. We curate a dataset of South African fake news and then train detection models. We contrast this with using widely available fake news datasets (from mostly USA website). We also explore making the datasets more diverse by combining them and observe the differences in behaviour in writing between nations fake news using interpretable machine learning.
The advent of social media changed the way we consume content favoring a disintermediated access and production. This scenario has been matter of critical discussion about its impact on society. Magnified in the case of Arab Spring or heavily critici zed in the Brexit and 2016 U.S. elections. In this work we explore information consumption on Twitter during the last European electoral campaign by analyzing the interaction patterns of official news sources, fake news sources, politicians, people from the showbiz and many others. We extensively explore interactions among different classes of accounts in the months preceding the last European elections, held between 23rd and 26th of May, 2019. We collected almost 400,000 tweets posted by 863 accounts having different roles in the public society. Through a thorough quantitative analysis we investigate the information flow among them, also exploiting geolocalized information. Accounts show the tendency to confine their interaction within the same class and the debate rarely crosses national borders. Moreover, we do not find any evidence of an organized network of accounts aimed at spreading disinformation. Instead, disinformation outlets are largely ignored by the other actors and hence play a peripheral role in online political discussions.
The dynamics and influence of fake news on Twitter during the 2016 US presidential election remains to be clarified. Here, we use a dataset of 171 million tweets in the five months preceding the election day to identify 30 million tweets, from 2.2 mi llion users, which contain a link to news outlets. Based on a classification of news outlets curated by www.opensources.co, we find that 25% of these tweets spread either fake or extremely biased news. We characterize the networks of information flow to find the most influential spreaders of fake and traditional news and use causal modeling to uncover how fake news influenced the presidential election. We find that, while top influencers spreading traditional center and left leaning news largely influence the activity of Clinton supporters, this causality is reversed for the fake news: the activity of Trump supporters influences the dynamics of the top fake news spreaders.
Recent years have seen a strong uptick in both the prevalence and real-world consequences of false information spread through online platforms. At the same time, encrypted messaging systems such as WhatsApp, Signal, and Telegram, are rapidly gaining popularity as users seek increased privacy in their digital lives. The challenge we address is how to combat the viral spread of misinformation without compromising privacy. Our FACTS system tracks user complaints on messages obliviously, only revealing the messages contents and originator once sufficiently many complaints have been lodged. Our system is private, meaning it does not reveal anything about the senders or contents of messages which have received few or no complaints; secure, meaning there is no way for a malicious user to evade the system or gain an outsized impact over the complaint system; and scalable, as we demonstrate excellent practical efficiency for up to millions of complaints per day. Our main technical contribution is a new collaborative counting Bloom filter, a simple construction with difficult probabilistic analysis, which may have independent interest as a privacy-preserving randomized count sketch data structure. Compared to prior work on message flagging and tracing in end-to-end encrypted messaging, our novel contribution is the addition of a high threshold of multiple complaints that are needed before a message is audited or flagged. We present and carefully analyze the probabilistic performance of our data structure, provide a precise security definition and proof, and then measure the accuracy and scalability of our scheme via experimentation.
التعليقات
جاري جلب التعليقات جاري جلب التعليقات
سجل دخول لتتمكن من متابعة معايير البحث التي قمت باختيارها
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا