Revealing The Secret Power: How Algorithms Can Influence Content Visibility on Twitter/X††thanks: Published in the Proceedings of the 33rd Network and Distributed System Security Symposium (NDSS 2026) – please cite the NDSS version.
Abstract
In recent years, the opaque design and the limited public understanding of social networks’ recommendation algorithms have raised concerns about potential manipulation of information exposure. Reducing content visibility, aka shadow banning, may help limit harmful content; however, it can also be used to suppress dissenting voices. This prompts the need for greater transparency and a better understanding of this practice.
In this paper, we investigate the presence of visibility alterations through a large-scale quantitative analysis of two Twitter/X datasets comprising over 40 million tweets from more than 9 million users, focused on discussions surrounding the Ukraine–Russia conflict and the 2024 US Presidential Elections. We use view counts to detect patterns of reduced or inflated visibility and examine how these correlate with user opinions, social roles, and narrative framings. Our analysis shows that the algorithm systematically penalizes tweets containing links to external resources, reducing their visibility by up to a factor of eight, regardless of the ideological stance or source reliability. Rather, content visibility may be penalized or favored depending on the specific accounts producing it, as observed when comparing tweets from the Kyiv Independent and RT.com or tweets by Donald Trump and Kamala Harris. Overall, our work highlights the importance of transparency in content moderation and recommendation systems to protect the integrity of public discourse and ensure equitable access to online platforms.
1 Introduction
Digital spaces play a critical role in shaping today’s information environment. Across the world, users rely on social platforms to access information on a wide range of topics, from personal interests to issues of societal importance. Within this ecosystem, recommendation algorithms are pivotal in governing the visibility and circulation of content. These algorithms suggest or outright select content for users based on interests inferred from, e.g., their interactions and connections, as well as the content they create or engage with.
Crucially, recommendation algorithms can also limit the visibility of some content, e.g., hashtags related to dangerous or inappropriate topics or information that may encourage harmful behaviors [36]. This is known as “shadow banning,” “visibility alteration,” or “reduction,” and typically aims to moderate online discussions, promote the dissemination of reliable content, and minimize the likelihood of toxic interactions to foster a healthier information ecosystem [25]. In practice, shadow banning and, in general, algorithmic alteration aimed at reducing the visibility of a user/their content might occur to control platform dynamics and for a number of reasons, ranging from mitigating spam/bot activity, algorithmic cleaning (e.g., throttling low-performing, repetitive, or self-promotional content), to prioritizing user retention (e.g., by penalizing posts linking to competing platforms).
Alas, the details of how shadow banning works and the rationale behind penalizing procedures remain mostly opaque. In some cases, altering visibility may be a deliberate action by a platform’s moderation team [44, 38]; in others, it could result from certain types of content being automatically flagged [16]. Whether humans or automated systems make these decisions, the lack of transparency surrounding shadow banning may introduce a certain degree of arbitrariness, real or perceived, as users often have little or no way of determining whether their content has been subject to visibility restrictions [24].
This opacity also encourages doubts as to whether or not shadow banning may be used to suppress specific viewpoints or communities [38]. The difficulty in detecting it, compared to more visible actions like account suspensions or bans, further exacerbates these concerns. Users may remain unaware that their content is being deliberately suppressed, which can lead to accusations of bias or censorship [29]. For example, journalists have reported internal mechanisms on platforms like Twitter (now X) and Facebook that reduce the visibility of certain types of content without notifying users [39].
While Twitter/X previously denied engaging in shadow banning [29], the platform’s updates to its terms of service in 2024 began to acknowledge the use of such practices [42]. This has fueled debates about the fairness and transparency of content moderation. The ability to shape or manipulate online debates to serve corporate or political interests arguably poses a significant threat to democratic discourse. Moreover, some have criticized the potential for platforms to silence political or ideological views under the guise of content moderation [11] – for instance, Elon Musk claimed that Twitter previously penalized conservative viewpoints, asserting that the so-called “Twitter Files” suggested algorithmic interventions biased against certain ideologies [37].
Beyond anecdotal evidence, researchers have begun looking into shadow banning through technical and statistical analyses (see Section 2). Overall, prior work suggests that shadow banning is relatively widespread across social media platforms [29, 7]. However, the lack of transparency in applying these techniques raises serious concerns. This prompts the need for systematic techniques to detect and measure visibility alterations and assess whether platforms employ shadow banning to favor specific narratives, limit dissenting voices, or prioritize corporate interests.
Technical Roadmap. Prior work has investigated shadow banning by analyzing search results or auditing recommendation algorithms, revealing potentially systematic penalization of specific users or content. However, these methods have not been able to determine whether content is penalized based on specific features (e.g., the presence of URLs), particularly at the debate level or across different topics.
In this paper, we aim to address this gap and evaluate the presence of visibility alterations in the wild using view counts, a largely underexplored metric in related work introduced on Twitter in 2022 and once available through the now mostly defunct API. As academic access to the API has been restricted since June 2023 [34, 22], we turn to two publicly available datasets that include view counts: 1) almost 35M tweets related to the 2024 US Presidential Election, released by Balasubramanian et al. [4], and 2) over 17M tweets related to military aid in the Ukraine-Russia war, released by Baqir et al. [5].111As Twitter rebranded to X in July 2023, tweets are now called posts. Since one of our datasets was collected before the rebranding and one after, to ease presentation, in the rest of the paper, we stick to “tweets.” Our methodology relies on data analysis and network science techniques to uncover traces of shadow banning at the content, user, and network levels, shedding light on visibility alterations. Although the inaccessibility of the platform’s recommendation algorithm prevents us from establishing an explicit causal link to algorithmic design, our findings provide reasonable evidence in support of this hypothesis. While alternative explanations are possible, we discuss why they are unlikely to hold in Section 6.
Research Questions. Overall, we identify and address the following research questions:
-
•
RQ1. Does the visibility of content depend on the characteristics of the information it contains? Our analysis examines how various content features, including the presence of external links, political bias, factuality of the sources, and ideological stance, may influence visibility.
-
•
RQ2. Do users experience different levels of visibility? We analyze the activities of prominent accounts to detect systematic variations in individual content visibility, i.e., whether some users’ content consistently receives less exposure.
-
•
RQ3. Does visibility vary across different communities? Social media users tend to form ideologically homogeneous communities through repeated interactions; thus, we investigate whether the content generated within these communities exhibits varying levels of visibility.
Main Findings. We introduce and use a new metric that accounts for author popularity when evaluating content visibility, called p-score, testing its dependence on content, users, and community characteristics. We find that visibility can vary substantially depending on content characteristics, both for content related to the Ukraine-Russia War and the 2024 US Elections. Even users with comparable profiles experience drastic differences in visibility—sometimes differing by orders of magnitude.
Among other things, we emphasize the following findings:
-
1.
At the content level (RQ1), tweets containing links to external websites have a significantly lower median p-score than those without links—respectively, approximately eight times for Ukraine-Russia and four times for the US Elections dataset.
-
2.
From the point of view of user-level visibility (RQ2), accounts that frequently post URLs and automated accounts are further penalized. Furthermore, in the Ukraine-Russia dataset, two prominent news outlets with comparable popularity (namely, RT.com and The Kyiv Independent) exhibit a median p-score differing by one order of magnitude (0.0069 vs. 0.084). A similar trend emerges in the US Elections dataset, where Donald Trump’s median p-score is approximately four times higher than that of Kamala Harris (0.132 vs. 0.030).
-
3.
Analyzing visibility variations at the community level (RQ3), using retweet and reply information, does not reveal systematic penalization of specific groups.
These discrepancies support the hypothesis that algorithmic interventions like shadow banning may be occurring, as Twitter/X’s recommendation algorithm systematically favors tweets without external URLs. Although we do not find evidence that specific topics, viewpoints, or communities are systematically penalized, interesting individual-level discrepancies do emerge, e.g., in the cases of The Kyiv Independent vs. RT.com and Donald Trump vs. Kamala Harris.
Overall, our work introduces a data-driven methodology using view counts and data science tools for detecting alterations in content exposure across different contexts. It also further highlights the importance of accessing information about content exposure to enhance transparency in social media algorithms and promote fairness and trustworthiness in online information ecosystems.
Disclaimer: Throughout the paper, we refer to systematic suppression, penalization, etc., in a broad sense, also to encompass selective content treatment. That is, we do not imply that suppression is necessarily ideologically motivated.
2 Related Work
In recent years, researchers have begun to work on identifying and categorizing various types of algorithmic moderation techniques, examining their consequences on the dynamics of information environments, and exploring their broader societal impact. Several studies highlight how algorithmic moderation shapes the visibility of content and influences public discourse. For instance, Zannettou et al. [45] present a large-scale analysis of Twitter, identifying content suppression instances, particularly in political debates, where specific hashtags were systematically downranked, limiting their reach. Jiang et al. [26] explore the impact of political bias in comment moderation on YouTube, finding that the likelihood of comment removal is independent of the political leanings of the video, even though other potential biases may persist in moderation practices.
Prior work has also examined the effects of algorithmic curation on users’ exposure to diverse content. Algorithms have been shown to influence what users see on Twitter [7], Facebook [3], TikTok [41], YouTube [21], and Google [35]. Bakshy et al. [3] analyze how Facebook’s algorithm reduces exposure to ideologically diverse news, contributing to filter bubbles and echo chambers [9]. On TikTok, Vombatkere et al. [41] find that the recommendation system primarily favors content based on likes and friendship relations, potentially contributing to users’ segregation.
In an effort to improve transparency, some platforms may offer explanations for why they remove content. However, these are often implausible or misleading, as observed by Mousavi et al. [31] for TikTok. Whereas, Jhaver et al. [23] demonstrate that providing users with meaningful explanations for content moderation decisions on Reddit reduces the likelihood of future post removals, indicating that transparency can positively influence user behavior.
Specific to shadow banning, Le Merrer et al. [29] investigate Twitter algorithmic interventions to reduce the visibility of certain content or accounts without outright removal. They highlight widespread traces of visibility interventions across various user groups and communities. Compared to our work, their study focuses on account-level interventions based on appearances in suggestions, search results, or discussion cascades, without considering alterations at the content level.
Beyond technical analysis, scholars have also examined the broader implications of algorithmic moderation, particularly its impact on free speech, public discourse, and user trust. Gillespie [16] argues that content recommendation itself functions as a form of moderation, emphasizing the lack of transparency in how platforms manage visibility. This opacity can obscure biases in content curation and contribute to inequities among different communities, eroding trust in these systems. Others have explored societal and political consequences. For example, DeVito [10] analyzes the limited information released about Facebook’s News Feed algorithm, uncovering the underlying values that shape its decisions and calling for greater regulatory oversight.
Overall, previous studies analyzing the presence of shadow banning and biases in algorithmic suggestion systems have mostly relied on bots [21], developed auditing models [35, 31], or examined algorithmic interventions by analyzing search query results [29]. However, algorithmic interventions that shape content visibility exist at different levels and in various forms, ranging from downranking in suggestion systems to complete hiding in search results. In this paper, we approach this issue from a broader perspective, leveraging the number of times a piece of content has been viewed to detect the presence of shadow banning. This allows us to capture the effects of multiple types of interventions without being limited to a specific form of algorithmic action.
Dataset | Transformation | #Posts | #Replies | #Retweets | #Quotes | #Accounts | #Links | #Domains |
---|---|---|---|---|---|---|---|---|
US Elections | Main Dataset – All Content | 34,762,696 | 23,906,163 | - | 3,883,501 | 3,844,610 | 2,320,840 | 75,876 |
Ukraine-Russia | with Valid View Count | 4,306,644 | 2,208,374 | - | 252,849 | 1,509,877 | 631,960 | 53,975 |
US Elections | Network & Latent | - | 4,457,343 | - | - | 493,149 | - | - |
Ukraine-Russia | Ideology Estimation | - | - | 9,939,787 | - | 67,926 | - | - |
US Elections | Claim & | 2,451,532 | - | - | - | 652,851 | - | - |
Ukraine-Russia | Topic Tracking | 1,845,285 | - | - | - | 583,803 | - | - |
3 Dataset
We now present the datasets analyzed in our study, specifically, two sets of Twitter/X posts that include view counts, i.e., the total number of times they have been viewed.
Ukraine-Russia War Dataset. In 2023, Baqir et al. [5] collected tweets related to military aid in the Ukraine-Russia war through the Academic API.222The dataset is available at https://osf.io/5m3vr/. More precisely, they used the following set of keywords: ‘military aid,’ ‘military support,’ ‘tanks,’ ‘Abrams,’ ‘Leopard,’ ‘Challenger,’ ‘jet,’ ‘aircraft,’ ‘munitions,’ ‘HIMARS,’ ‘rockets,’ and ‘missile.’ In addition to tweet metadata typically found in other datasets, this dataset also includes view counts for each piece of content.
Overall, the Ukraine-Russia war dataset includes over 17 million tweets from more than 5.2 million users, posted between November 22, 2022, and March 1, 2023, and collected in April 2023. The temporal gap between tweet creation and collection ensures that engagement and view metrics have stabilized, thus enabling robust analysis [33] and providing a meaningful snapshot of the discourse surrounding the Ukraine-Russia conflict on Twitter [5].
2024 US Elections. The second dataset was collected by Balasubramanian et al. [4] using a custom-built scraper.333The dataset is available at https://github.com/sinking8/usc-x-24-us-election. The authors curated a list of 44 keywords related to the 2024 US Elections, including ‘2024 Elections,’ ‘Donald Trump,’ ‘letsgobrandon,’ ‘ultramaga,’ etc. Following retrieval, we filter the data as discussed in Section 4.1 and obtain approximately 35 million tweets with valid view counts by 3.8 million users.
Arguably, these two datasets are well-suited to the purpose of our analysis, i.e., studying the occurrence and characteristics of shadow banning. More precisely, the topic of the Ukraine-Russia war has polarized public opinion into supporters and opponents with clearly distinct stances [28]. This allows us to identify the factions involved in the debate and evaluate whether Twitter/X has been altering the visibility of specific narratives. Similarly, discourse around US Elections is inherently polarized; thus, Balasubramanian’s dataset [4] lends itself to identifying opposing factions (i.e., supporters of different political candidates) and enables the analysis of visibility alterations at the community level.
4 Methods
For both datasets, we pre-process the data to filter out possible inconsistencies in the view count statistics. We also merge content domain information with third-party data to assign political leaning and factuality labels to tweets. Finally, we select a set of the most prominent accounts, referred to as “influencers,” representing the two opposing factions of each debate. The influencers are then used to infer users’ ideological stances and analyze visibility variations with respect to ideology. In this section, we discuss these actions in detail.
4.1 Pre-Processing
For a sample of tweets, we manually verify that the view counts in our datasets are broadly consistent with those displayed on the Twitter/X platform. While we observe minor discrepancies, these are likely due to the interval between data collection and manual verification, during which view counts may have slightly changed. Overall, our sanity check confirms the reasonable reliability of the datasets from [5, 4].
We also perform a filtering procedure to remove potentially inconsistent records, specifically, content for which the view count is unavailable and content for which the number of interactions (i.e., likes, comments, retweets, or quotes) exceeds the recorded number of views, as users must view the content before engaging with it.
4.2 Factuality and Political Bias Estimation
To test the dependence of visibility variations on content-level features like political bias and factuality (cf. RQ1 in Section 1), we rely on third-party classifications to infer the political bias and factuality of the news sources referenced in the debate. More precisely, we use a dataset obtained from Media Bias/Fact Check (MBFC), an independent fact-checking organization that categorizes news outlets based on their factuality and political bias.444https://mediabiasfactcheck.com/. This includes 2,190 different news outlets, along with their domain names, political leaning, and factuality, and classifies outlets across the political spectrum, from “extreme left” to “extreme right.” Additionally, some outlets are labeled as “questionable” or “conspiracy-pseudoscience” if they frequently publish misinformation, false content, or endorse conspiracy theories.
We use the MBFC dataset to assign factuality (from ‘Very High’ to ‘Very Low’) and political bias (‘Extreme Right’ to ‘Extreme Left’) to tweets based on the domain referenced. For example, a tweet linking to a CNN article is classified as ‘Mostly Factual’ with a ‘Left-Center’ bias, and one linking to RT News has a ‘Very Low’ factuality with a right-center bias.
4.3 Influencers Selection
Studying the relationship between variations in visibility and individual features like ideological stance (cf. RQ2) requires us to infer users’ opinions. To this end, we use Latent Ideology Estimation [15], a flexible and powerful method for inferring users’ ideological stances that relies on the identification of a set of influential accounts actively engaged in the debate. This group of users, referred to as influencers, must encompass several subcategories and represent a broad range of opinions across the ideological spectrum.
For the Ukraine-Russia dataset, we follow the procedure explained in [5] and build a network using retweet interactions, ranking accounts based on their in-degree, i.e., the number of unique users who retweeted them. We then manually select a set of prominent accounts representing both sides of the debate, prioritizing those with the highest in-degree. This initial set serves as a seed, and we expand it by utilizing Twitter’s “Who to follow” recommendations from the selected accounts’ profiles. We repeat this process until no new relevant accounts are suggested. The selection is then refined by excluding accounts with an in-degree below 100 or those whose content was unrelated to the Ukrainian conflict. Ultimately, this yields a final set of 204 influencers, representing both supporters and opponents of military support to Ukraine.
For the 2024 US Elections dataset, retweet interactions are not sufficiently rich to enable a reliable estimation of users’ opinions, as the dataset contains only a few hundred retweets. Therefore, we use reply interactions to reconstruct the interaction network, which, although noisier, still allows us to distinguish between the two groups. We identify the most replied-to users and manually extract the top 100 accounts exhibiting a clear stance, either supporting or opposing Trump. This allows us to select a balanced set of influencers representing both pro- and anti-Trump positions. Using Hartigan’s dip test of unimodality [19], we confirm the bimodal nature of the latent ideological estimation distribution for users and influencers in both datasets (Ukraine-Russia War: D = 0.027, p-value for users and D = 0.074, p-value for influencers; US Elections: D = 0.012, p-value for users and D = 0.106, p-value for influencers). This result highlights the polarized nature of the debate, as well as the reliability of our inference [15].
4.4 Users’ Ideological Stance Estimation
Latent Ideology Estimation is a proven method for reliably estimating user ideology in diverse contexts [15, 6, 13]. Using the set of accounts identified using the methodology described above, we apply this algorithm to the Ukraine-Russia war and the US 2024 Elections debates to infer users’ stances and contribute to answering RQ2. Crucially, this algorithm depends on identifying a subset of key influencers, as their selection significantly affects the accuracy of the ideology estimation.
Once the influencer set is identified, we apply the Correspondence Analysis algorithm [17], performing three main steps: 1) constructing an interaction matrix, 2) normalizing it, and 3) performing singular value decomposition (SVD). More precisely, in the first step, we build the matrix , where each element represents the number of interactions (retweets for Ukraine-Russia, replies for US 2024 elections) that user directs to influencer . We then normalize by dividing it by the total number of interactions, yielding . Using the following quantities: , , , and , we normalize further to obtain . In the third step, we perform singular value decomposition of , such that , where and are orthogonal matrices and is a diagonal matrix of singular values. We estimate user ideological leaning by taking the subspace associated with the first singular value of the decomposition. Specifically, the latent ideology of user is the -th entry in the first column of the matrix , while the ideology of each influencer is estimated as the median ideology score of their retweeters/repliers.
4.5 Claim Detection and Tracking
To get a more nuanced understanding of the different types of textual content and visibility patterns associated with them, we focus on claim-level analysis of tweets. This is an additional vantage point to help us answer RQ1, with an emphasis on the claims being spread around different topics. To this end, we use Natural Language Processing (NLP) and Learning to Rank (LTR) techniques to group tweets into a coherent structure of claims.
First, we use the claim extraction pipeline from Lambretta [32] to extract claims from tweets that make an objective claim and are candidates for being tracked through downstream keyword extraction or semantic search methodologies. These claims are structurally similar to those manually curated by fact-checking organizations like Snopes or PolitiFact.555See www.snopes.com, https://www.politifact.com/. We sample 1,000 unique tweets sorted by several different criteria of engagement: 1) Retweet count, 2) Reply count, 3) Like count, 4) Quote count, and 5) Views count. This sampling gives us a set of tweets that have a higher likelihood of containing a claim (and thus garnered the level of interaction on Twitter), rather than sampling tweets randomly. We then use ClaimSpotter [20] to identify tweets that contain a claim: ClaimSpotter returns a score between 0 and 1, representing how “check-worthy” an input text is. Consistent with [32], we select 0.5 as the threshold for classifying tweets containing a claim. For the filtered tweets containing a claim, we use Lambretta’s keyword identification component to extract the best set of keywords representing the claim. Finally, we retrieve similar tweets that make the same claim as the seed set using the extracted set of keywords and use these subsets of tweets to analyze different visibility patterns.



Since the tweets in the datasets span longitudinally and contain a multitude of events (both related to the Ukraine-Russia war and the 2024 US Elections), we further label the claims identified in the seed set of tweets to different themes. This makes the analysis tractable while allowing us to achieve the granular content level analysis. We use the unsupervised topic modeling library BertTopic [18] to identify coherent topics within the seed set of claims. We select a laxer threshold (number of documents in a topic, and similarity measure for two claims to be assigned the same topic), resulting in a larger number of initial topics. These topics are then manually cleaned for noise (artifacts of keywords used in the initial data sampling process) and iteratively assigned to themes. This way, we have a hierarchy of themes within the seed set of claims and a subset of tweets across themes whose visibility patterns can be compared.
4.6 P-score
As mentioned earlier, we use the “view count” metric to examine differences in visibility across the Ukraine-Russia War and the 2024 US Elections datasets. We do so as shadow banning essentially is the act of reducing the visibility of specific profiles or content through various techniques [29, 16]. Thus, view counts are an appropriate proxy for quantifying visibility. However, directly comparing the visibility of users may prompt some challenges – more precisely, content visibility on social networks heavily depends on the popularity of the author, as social media algorithms use follower-following relationships to suggest content [41].
As a result, we introduce and use a metric denoted as p-score, which accounts for the author’s popularity as follows:
(1) |
In other words, the p-score of a tweet quantifies the ratio between the number of views it receives and the number of followers of its author. Hence, a higher p-score indicates that content receives more views per follower, while lower values suggest more limited circulation.
5 Detecting Visibility Variations
In this section, we set out to answer our three research questions (see Section 1) related to studying the occurrence of shadow banning at content, user, and network levels.



5.1 RQ1: Content Level
To analyze the occurrence of shadow banning at the content level, we examine whether the p-scores of the tweets depend on the type of information they convey. To ensure a fair comparison and avoid possible confounding factors, we exclude replies, retweets, and quotes from the analysis, ending up with 1.8M tweets for the Ukraine-Russia dataset and 7.2M tweets for the 2024 US Elections dataset.
We classify tweets based on the domain they refer to using the MBFC dataset (see Section 4.2), dividing them into five categories: “News Outlets,” “No Domain,” “Other,” “Other Social,” and “Twitter.” These refer, respectively, to tweets that contain a link to a news outlet, no link, a link to an unclassified website, a link to another social platform, and a link to other Twitter content.
Content P-Score Distributions. We begin by examining whether tweets, categorized by the type or absence of external URLs, exhibit differences in p-score distributions. In Figure 1, we plot the distribution of the p-scores for all original content (i.e., without considering replies, retweets, and quotes) for each category of content. For both datasets, the distribution has a log-normal shape, even though the categories contribute differently to the distribution density. For instance, the “Other” and “News Outlets” categories contribute more to lower values, while “No Domain” appears to be skewed toward higher values. Also, the spectrum of this distribution can vary widely, spanning nine orders of magnitude.
Although the wide range of p-scores can be explained by the typical dynamics of social networks, whereby some content goes viral, some is ignored, and the majority receives a few interactions, the spectrum’s unbalanced and asymmetric constitution suggests the presence of different visibility levels for different categories. Tweets that do not link to other domains seem favored over others in both datasets. However, this difference seems to be considerably more pronounced in the Ukraine-Russia dataset than for the 2024 US Elections. This may suggest a varying degree of intervention over time, while also highlighting the algorithm’s systematic promotion of content without external URLs.
Artifacts in Content Visibility Distributions. We also observe the presence of peaks in certain intervals of the distributions in the Ukraine-Russia dataset. This may depend on the systematic penalization of specific users or the algorithm detecting inappropriate content and limiting its circulation. We analyze the distribution of user contributions within each interval to better understand the causes of these outliers and the role of users in shaping them. To do so, we use the Gini index, a common metric to measure the inequality of distributions, computing it on the distribution of the number of tweets per interval for each author [43]. In our case, if all users contribute to the population of an interval with approximately the same number of tweets, the Gini index is close to 0, whereas if a significant disparity exists, with some users generating most tweets in the interval, the index approaches 1.
In Figure 2, we present the distribution of p-scores across tweet categories for both datasets (also see Table 2). In both cases, the median p-score (marked by the red line) for the “No Domain” category is higher than for all others, confirming observations from Figure 1. Specifically, tweets from news outlets exhibit the lowest median p-scores (0.033 for Ukraine-Russia, 0.012 for US Elections), while “No Domain” content shows significantly higher medians (Ukraine-Russia: 0.246; US Elections: 0.056). This further supports the hypothesis that Twitter’s algorithm favors content that does not contain external URLs. Interestingly, tweets linking to unclassified websites have median p-scores close to those of news outlets (Ukraine-Russia: 0.036; US Elections: 0.014), whereas tweets quoting other tweets or linking to other social platforms are slightly less penalized (median p-scores of 0.097 and 0.150 for Ukraine-Russia and 0.029 and 0.024 for US Elections). To validate these observations, we perform a Mann-Whitney U test between all pairs of content categories to assess whether the values in one group tend to be greater than those in another [30]. The Mann-Whitney U test does not assume normality and instead evaluates whether there is a statistically significant shift between distributions. In our case, a low p-value indicates strong evidence that the values from one group are systematically higher than those from another. For both the Ukraine-Russia and US Elections datasets, the tests confirm that the “No Domain” category statistically significantly dominates all others (p-value for all comparisons).
Content | Ukraine-Russia | US Elections |
---|---|---|
News Outlet | 252,046 | 1,564,923 |
No Domain | 796,673 | 4,317,988 |
Other | 730,078 | 851,246 |
Other Social | 65,025 | 483,346 |
1,463 | 24,509 |





For the Ukraine-Russia dataset, the peaks in the “Other” distribution correspond to the intervals with the highest Gini scores, indicating the presence of users with a high volume of tweets who receive a similar level of visibility. Indeed, most of the content populating these intervals belongs to a few accounts explicitly labeled as automated by the platform, suggesting that the algorithm may be limiting the visibility of their content. A similar behavior, although less pronounced, emerges in the 2024 US Elections dataset. The highest level of imbalance appears in the left tail of the “Other” category distribution, where a substantial portion of tweets comes from a few highly active accounts that primarily discuss news by frequently linking to external websites. For instance, the most represented account in the bins with the highest disparity values has an average activity of 38 tweets per day, with 97.5% of its content linking to external websites, suggesting that the algorithm may further penalize very active users who frequently post URLs.
Overall, these artifacts in the distributions of content visibility reflect an unbalanced contribution by users for both datasets, pointing to interventions aimed at reducing the visibility of specific accounts. These patterns also confirm the consistent application of such practices over time.
Bias | Ukraine-Russia | US Elections |
---|---|---|
Extreme Right | 929 | 124,547 |
Right | 4,211 | 427,986 |
Right-Center | 9,586 | 165,889 |
Least Biased | 9,838 | 200,011 |
Left-Center | 11,935 | 484,472 |
Left | 3,204 | 148,036 |
Factuality | Ukraine-Russia | US Elections |
---|---|---|
Very Low | 845 | 46,747 |
Low | 940 | 101,638 |
Mixed | 10,485 | 668,796 |
Mostly Factual | 4,068 | 218,133 |
High | 19,158 | 513,152 |
Very High | 2,818 | 3,230 |


Political Bias and Factuality. Next, we analyze the dependence of visibility on the political bias and factuality of the content being shared. We apply the same methodology to assess whether visibility varies based on the political bias and factuality of the sources referenced in tweets, for both the Ukraine-Russia War (Figure 3(a)–3(b)) and the 2024 US Elections (Figure 3(c)–3(d)) debates (also see Table 3). For both datasets, the results show a relatively contained variation in visibility across both political bias (Ukraine-Russia: min 0.019, max 0.05; US Elections: min 0.0081, max 0.015) and factuality (Ukraine-Russia: min 0.018, max 0.04; US Elections: min 0.0086, max 0.024), suggesting that the algorithm does not systematically penalize content based on these attributes. This is also confirmed via the Mann-Whitney U test, which did not find a dominant category for political leaning in the Ukraine-Russia dataset (all categories exhibit p-values 0.05 in at least one comparison). As for factuality, the tests highlight that the “Mixed” and “Low” categories tend to be penalized (maximum p-values and 0.0019, respectively), but reject the hypothesis that “Very Low” content is consistently outperformed by “High” or “Very High” categories. Similarly, in the US Elections dataset, there is no evidence of clear dominance among political leanings (all categories have p-values 0.05 in at least one comparison), although the tests detect the lower performance of the “Left” category (p-value ). Interestingly, for factuality, “Very High” content outperforms all other categories (p-values in all comparisons), while no clear dominance emerges among the remaining classes. However, the number of tweets in the “Very High” category is considerably smaller than in others (see Table 3), suggesting an occasional rather than systematic consumption of highly factual content in the US Elections debate.
Notably, the “Right-Center” and “Mixed” categories in the Ukraine-Russia dataset exhibit a higher Gini index in some lower p-score bins, suggesting the possible presence of penalized domains belonging to those categories. For the US Elections dataset, this effect is less pronounced, with only the “Right-Center” category showing minor anomalies with slightly higher Gini indices. Additionally, the latter displays greater variation in the Gini index across the distribution, with lower p-score bins having higher Gini values. However, this behavior appears to affect all categories of political bias and factuality – except “Very High,” which has considerably fewer observations – and thus does not suggest targeted interventions to limit the visibility of specific categories. Overall, the analysis of content features and their influence on visibility highlights that content circulation on Twitter is influenced primarily by the presence of external URLs, rather than by the political stance or factuality of the information conveyed, with tweets containing links to external sources systematically penalized by the recommendation algorithm.



Topics. Next, we use the textual content of tweets to investigate whether the different topics being discussed experienced varying levels of visibility. Following the claim detection and tracking methodology discussed in Section 4.5, for the Ukraine-Russia dataset, we first filter out 2,735 seed tweets making a claim. These seed tweets are further categorized into five different topics: 1) Europe/NATO Military Support to Ukraine, 2) Criticism of US Military Support to Ukraine, 3) US Military Support to Ukraine, 4) War Related Reporting and Events, and 5) Pro-Russian Support. Keyword extraction and the subsequent matching procedure return 151,697 tweets discussing claims related to the seed topics. We applied the same procedure for US elections dataset, categorizing the seeds into the following topics: 1) Anti-Trump Campaign Messaging, 2) International Conflicts & Electoral Impact, 3) Policy Debates & Issues, 4) Pro-Trump Campaign Messaging, 5) Public Reaction & Media Coverage, 6) Scandals, Conspiracies & Legal Controversies
Figure 4 shows the distribution of the p-score values for the claim-matched tweets, divided by each theme of claims. Similar to Figure 1, the distribution follows a log-normal shape. However, in both cases the p-scores from the different themes have a largely balanced and symmetric distribution, contributing almost uniformly to the distribution density.




This is indicative of similar visibility levels across different thematic content and suggests the lack of interventions at the topic level. We also investigate the distribution of users within each interval by plotting the Gini index on the distribution of the number of tweets per interval for each author. Figure 5 reports the distribution of p-scores across tweet categories, with red lines indicating the median value for each category, alongside the Gini index for each interval for the Ukraine-Russia and the 2024 US Elections datasets. For the Ukraine-Russia dataset, the median p-score for the different topical themes is very close to each other (within a Standard Deviation of 0.0296), with the topic “War Related Reporting and Events” having the highest median (0.143), and “Criticism of US Military Support to Ukraine” the lowest median p-score (0.071828). A similar pattern is observed in the US Elections dataset, where the highest median p-score corresponds to the “International Conflicts & Electoral Impact” theme (0.0561), while the lowest to the “Policy Debates & Issues” theme (0.0349). Taken together, these results suggest the lack of any topic-oriented intervention on the algorithm’s part.
5.2 RQ2: User Level
Besides content-level visibility, reductions can also affect individual accounts, regardless of what they post. Hence, we analyze whether users experience different levels of visibility based on their role in the debate and their ideological stance, specifically, their support for or opposition to military aid to Ukraine as well as support for or opposition to Donald Trump in the lead-up to the 2024 US Elections.
Inferring Users’ Ideology. To perform account-level analysis, we first need to infer users’ attributes, such as their stance on the debated issue. To this end, we use latent ideology estimation (see Section 4.4) to classify users’ ideological stances. Given a set of prominent accounts known as influencers (see Section 4.3), this technique allows us to infer users’ stances based on interaction information. Matrix decomposition maps users who interacted with a similar set of influencers close to each other in a real space, while users with fewer common influencers are positioned farther apart.
Due to the absence of retweet statistics for the 2024 US Elections dataset, we estimate latent ideology using retweet interactions for the Ukraine-Russia and reply interactions for the US Elections datasets. As mentioned, we do so as the latter only includes a few hundred retweets that did not have a valid view count. Although users’ estimaton results a bit noisier, the use of replies does not affect the overall ability to infer accounts’ stances. Indeed, the bar plots in Figure 6 show the distribution of ideology scores for users and influencers in the Ukraine–Russia War (panel a) and the US Elections (panel b), together with the distribution of ideology scores for repliers/retweeters of a subset of ten influencers in each debate. Both datasets exhibit two communities positioned at opposite ends of the space, revealing polarized debates with bimodal opinion distributions, as also highlighted by Baqir et al. [5] in the context of the Ukraine-Russia war and by the Dip test as reported in Section 4.3.
In each debate, we categorize accounts into two stances: supporters (ideology score ) or opponents of military aid (ideology score ), and supporters (ideology score ) or opponents of Donald Trump (ideology score ). We then calculate the average p-score for users and influencers in our dataset. To ensure the robustness of our results, we only consider accounts with more than five original tweets.
Visibility vs Ideology. After inferring user opinions, we analyze whether the visibility of the account is influenced by the ideology they support. Figure 7 illustrates the distributions of the average p-score for users and influencers representing opposing ideological stances in the two debates. Influencers tend to have higher average p-scores than users in both debates, likely due to their prominent role and greater popularity. However, there is little difference between the p-score distributions of users and influencers when comparing ideological stances in both the Ukraine-Russia and US Elections datasets. For the Ukraine-Russia dataset, the median p-scores are 0.118 for supporters vs. 0.136 for opponents among users, and 0.479 for supporters vs. 0.499 for opponents among influencers. For the US Elections dataset, the median p-scores are 0.049 for Trump supporters vs. 0.048 for opponents among users, and 0.280 for Trump supporters vs. 0.358 for opponents among influencers. These results are confirmed by the Mann-Whitney U test, which shows significantly higher performance for influencers compared to users (p-value 0.0005 for both supporter and opponent influencers versus users; p-value for pro- and anti-Trump influencers versus users). The test also reveals comparable visibility between users and influencers of different stances (p-value 0.07 for the Ukraine-Russia case and p-value 0.29 for the US elections).
Influencers p-score distributions. Although we have ruled out the penalization of specific ideologies, visibility alterations may still occur for specific accounts. This is suggested by the presence of outliers among influencers in Figure 7, which may indicate accounts experiencing either reduced or increased content circulation due to algorithmic visibility adjustments. To investigate the presence of algorithmic interventions and to mitigate the influence of potential confounding factors, we exclusively focus on influencers and divide them based on their popularity in different tiers, as done in recent work [40]. Figure 8 displays the p-score distribution for the accounts used as influencers in the analysis of the two debates that posted more than five original pieces of content, with accounts grouped according to the popularity tier they belong to.
For both datasets, within the same group, influencers tend to exhibit distributions with similar shapes but different medians, reinforcing the possibility of algorithmic intervention at the individual level. We also observe that some influencers exhibit multiple peaks in their p-score distributions. This pattern could be attributed to the use of different types of content, each achieving varying levels of visibility, as illustrated in Figure 1. Alternatively, it may reflect temporal fluctuations in visibility due to adjustments in the suggestion algorithm. Notably, in the Ukraine-Russia war debate, mfa_russia and RT_com (accounts associated with the Ministry of Foreign Affairs and a Russian state-sponsored news outlet, respectively) significantly underperform relative to other accounts within the same tier. In the US Elections debate, the accounts linked to Joe Biden and Kamala Harris exhibit p-score distributions skewed lower than that of Donald Trump.


Case Studies. To shed light on the individual differences observed earlier, we select pairs of users with comparable roles in the two debates and analyze the distribution of their p-scores. In the case of the Ukraine-Russia war debate, we analyze RT_com and compare it to KyivIndependent, a news outlet supporting Ukraine. We select these accounts as they have comparable levels of popularity and belong to news agencies representing opposing factions, thereby minimizing potential confounding factors that could influence visibility. For the 2024 US Elections, we select the final candidates from the two major parties, i.e., Donald Trump and Kamala Harris.
One might argue that the circulation of content also depends on how much other users engage with it; thus, a more active user base could enhance the diffusion of the author’s content. To assess this, Figure 9 presents a 2-D density plot of the p-score versus the ratio of retweets to views. This ratio serves as a proxy for quantifying how actively consumers re-share content after being exposed to it. In particular, Figure 9(a) reveals that RT_com and KyivIndependent have rather similar user bases in terms of active sharing, as highlighted by the top marginal distribution (Mann-Whitney U test p-value: 0.093). However, they display significant differences in their p-score distributions (Mann-Whitney U test p-value: , median p-score: KyivIndependent=0.084, RT_com=0.007). Given the similarities across various dimensions, these findings strongly support the hypothesis that content from RT_com has been subject to visibility reduction.




In the case of the US Elections, the separation between Trump-Harris distributions is less clear (Figure 9(b)), as the two 2D densities overlap in some areas. Nevertheless, their marginal distributions differ significantly along both dimensions. Specifically, the distribution of retweet count over view count reveals that Donald Trump’s audience is less prone to explicitly endorse his content through active engagement compared to Kamala Harris’s audience (median retweets/views: 0.0025 for Trump, 0.0043 for Harris, Trump vs Harris Mann-Whitney U test p-value=1, Harris vs Trump Mann-Whitney U test p-value ). However, the p-score marginal distributions show that most of the content produced by Donald Trump achieves greater visibility than that produced by Kamala Harris (p-score median: 0.13 Trump, 0.03 Harris, Trump vs Harris Mann-Whitney U test p-value ). Hence, despite having a less active user base, Donald Trump’s content experienced increased visibility, suggesting the possible influence of algorithmic intervention.
5.3 RQ3: Network Level
Last but not least, to answer RQ3, we build retweet and reply interaction networks and study whether users interacting with each other experience similar visibility. We compute the joint density of the p-score for each account with respect to the average p-score of the accounts that interacted with it, as illustrated in Figure 10. To avoid potential conflating factors related to different roles, we compute separate density distributions for influencers and users based on their ideological stance. For the Ukraine-Russia war (Figure 10(a)), we use retweet information to capture interaction, while, for US Elections (Figure 10(b)), we employ replies to build the interaction network due to the unavailability of retweets.
The analysis indicates an absence of correlation between individual p-scores and the average p-scores of retweeters/repliers across both account classes and ideological stances for both datasets. Thus, users who interacted with accounts with reduced visibility do not necessarily experience decreased visibility themselves, regardless of account type, stance, or debate. This is further supported by Pearson’s correlation test, which yielded a maximum correlation value of 0.13 for the Influencer-Supporter case in the Ukraine-Russia dataset and a value of 0.094 in the US elections debate.
6 Discussion
The visibility of content online is inherently shaped by recommendation algorithms, as these determine what users see based on various engagement metrics and platform-specific goals [26]. Many platforms report reducing visibility as part of their moderation strategies to limit the spread of harmful or borderline content, purportedly to safeguard public discourse [16]. However, algorithmic visibility reduction may occur for several reasons, for instance, to penalize low-performing content or to prioritize user retention.
Regardless, the lack of transparency around these interventions contributes to a growing distrust towards platforms [36]. Moreover, while interventions can reduce the spread of harmful content or limit the influence of problematic users (including bots or coordinated disinformation campaigns), there is no guarantee that the same mechanisms are not used to silence minority voices or suppress narratives contrary to corporate interests. These interventions are subtle, far less noticeable than bans or content removal, and users may remain unaware that the reach of their posts is being reduced [16].
Our work investigated visibility reduction in the context of two Twitter datasets related to two polarized topics (the Ukraine-Russia war and the 2024 US Elections) using the content views metric. This allowed us to directly observe the effects of algorithmic interventions and assess how visibility reductions vary based on several factors. In doing so, we analyze both the behavior of recommendation algorithms at a granular level and how views can help identify otherwise unnoticed visibility suppression.
Data availability. Practices like shadow banning can significantly influence online debates and shape public opinion. As a result, the ability to study these interventions is crucial. However, while data availability is essential in doing so, in recent years, access to social media data for research purposes has become increasingly restricted [27], exacerbating the opacity of platforms and further eroding public trust. For instance, there is growing concern that social media companies could reduce content visibility to reduce the spread of misinformation, but may be choosing not to do so [1].
Indeed, our study reveals that Twitter did not systematically apply visibility reduction techniques to limit the spread of unreliable sources both in the Ukraine-Russia war and the 2024 US Presidential Elections debates. Instead, interventions appear to target specific accounts flagged as problematic, while all content containing external links has been penalized, regardless of the credibility of the information shared.
Longitudinal analysis. The comparison between multiple datasets collected during different periods and on different topics also helped us highlight the evolution of visibility depending on the context considered. For example, although the algorithm tends to reduce the visibility of tweets containing URLs, the strength of such interventions appears to differ between the two cases. Similarly, the differences in the visibility analysis at the individual level may result from variations in the behavior of the recommendation algorithm.
Additional factors. Finally, other factors influencing content visibility, including coordinated behaviors, user engagement strategies, and the topic under discussion, should also be taken into consideration. While these factors can exploit algorithmic design to boost visibility, our methodology proves effective in detecting visibility disparities between types of content and user behavior independently from the origin of these differences. In other words, alternative explanations for our findings are, arguably, far less likely.
For instance, one possible reason for visibility alteration could be related to bots, however, our visibility metric adjusts for follower count to account for potential bot-driven inflation. Furthermore, our network analysis is designed to detect clusters of users with altered visibility, including those that may be the result of coordinated behavior. Since no penalized communities were identified, such groups, if they exist, either evade detection or do not benefit from increased visibility, making coordinated manipulation an unlikely explanation. At the user level, we include retweets to capture engagement-driven visibility; nevertheless, users with similar roles and activity levels still exhibit significant differences in visibility, which suggests that user-level factors alone cannot explain the observed penalization. Lastly, the systematic reduction in visibility of content containing URLs in both debates, regardless of the type of resource referenced, strongly suggests algorithmic intervention rather than varying user interest.
Ethics Considerations. Our work relies on publicly available data obtained from other studies [5, 4], retrieved using the Twitter public API or by scraping public data, and augmented with publicly available tools. As such, it is not considered human subject research by our institutions. We acknowledge that one of our datasets was collected by [4] using a custom scraper, which may potentially contravene Twitter’s terms of service (ToS). However, we believe our research to be in the public interest, as studying platform manipulation arguably outweighs ToS concerns, in a manner not dissimilar to, e.g., violating robots.txt when crawling cybercrime forums [8]. In general, ethics researchers have made the case that ethics might not depend on breaking the ToS of the platforms [14].
Furthermore, to preserve anonymity, we do not report any information about the users in the dataset other than those of public entities like news organizations. For accounts that play an important role in the debate (i.e., influencers), we anonymize those with fewer than 50,000 followers, aligning with previous research [13]. Overall, we believe that our work aligns with the Menlo Report [2]’s principles of Beneficence, Respect for Persons, and Public Interest.
7 Conclusion
This paper examined differences in content visibility on Twitter/X – at content, user, and network levels – in the context of the Ukraine-Russia war and the 2024 US Presidential Elections. We relied on view counts, a largely unexplored feature of Twitter/X posts, to estimate content circulation and detect forms of visibility penalization that are not captured by techniques previously used in the literature. We found evidence that Twitter/X consistently favors tweets that do not contain links to external websites. While the platform does not systematically alter content visibility based on political views, the factuality of the referenced sources, or ideological stances, we do find evidence of visibility alterations at the individual account level, as illustrated by the cases of RT.com versus The Kyiv Independent and Donald Trump versus Kamala Harris. Our work demonstrates how visibility alterations can be detected using engagement metrics like view counts, an underexplored and often unavailable metric.
As mentioned, our work mainly focuses on algorithmic penalization in a broad sense, while we do not aim to determine whether/how this may be ideologically motivated.
Recommendations. We believe that the methods proposed in this paper can help researchers and civil society alike improve transparency by uncovering penalization practices adopted by social media platforms, especially if they might end up infringing on free speech. Concretely, non-profits, research outfits, and fact-checkers can leverage our methods to raise awareness and increase transparency by publishing dashboards of content, keywords, and domains that are anomalously penalized, and record historical information about potential changes in visibility dynamics in public reports.
Limitations. Our study revolves around two datasets covering two different topics and collected using different methods, i.e., the Academic API and Web scraping. This may result in varying levels of representativeness of the debate. To mitigate this potential issue, we pre-processed the data to ensure strong coherence and minimize the likelihood of any artifacts influencing our results. Also, although our methodology proved effective in detecting visibility alterations, future and/or more sophisticated strategies may still go undetected.
Finally, going forward, concerns about the limited availability of social network data for researchers cast doubts on the feasibility of conducting at-scale audits and studies of algorithmic behavior and practices adopted by platforms. Without data, ensuring accountability and maintaining trust in these platforms will remain an uphill battle. Initiatives like the Digital Services Act (DSA) [12] aim to enforce greater transparency, however, the real-world implementation of these laws may prove challenging.
Future Work. Our work could be enhanced in several ways. Integrating other factors – e.g., the influence of bots and coordinated accounts – could provide a more thorough understanding of how these elements affect visibility and how the algorithm responds to their presence. Also, as part of our data was retrieved using official APIs, we acknowledge that platforms may influence the data they provide to mask intervention patterns. While we took several integrity-enhancing actions to ensure the reliability of the data, one could combine API-based and scraping techniques to collect the same dataset to offer even more robust and nuanced insights.
Moreover, a comparison across additional datasets covering additional topics over extended periods could help identify changes in algorithmic behavior. Finally, comparing visibility patterns across other platforms would tease out unique traits of Twitter’s recommendation algorithm while also studying practices shared with other systems.
Acknowledgments. A.G. gratefully acknowledges the financial support provided by CY4GATE, by the National Recovery and Resilience Plan (NRRP) project “Securing Software Platforms – SOP” (CUP H73C22000890001), and by the European Union under the NRRP, Mission 4 Component 2 Investment 1.3 – Call for Proposals No. 341 of March 15, 2022, Italian Ministry of University and Research – NextGenerationEU, Project Code PE00000014, Concession Decree No. 1556 of October 11, 2022 (CUP D43C22003050001), “SEcurity and RIghts in the CyberSpace (SERICS) – Spoke 2 Misinformation and Fakes: DEcision supporT systEm foR cybeR intelligENCE (Deterrence). G.S.’s work was partially supported by NSF Grant CNS-2419829.
References
- [1] C. Bagchi, F. Menczer, J. Lundquist, M. Tarafdar, A. Paik, and P. A. Grabowicz. Social media algorithms can curb misinformation, but do they? arXiv:2409.18393, 2024.
- [2] M. Bailey, D. Dittrich, E. Kenneally, and D. Maughan. The menlo report. IEEE Security & Privacy, 10(2), 2012.
- [3] E. Bakshy, S. Messing, and L. A. Adamic. Exposure to ideologically diverse news and opinion on Facebook. Science, 348(6239), 2015.
- [4] A. Balasubramanian, V. Zou, H. Narayana, C. You, L. Luceri, and E. Ferrara. A public dataset tracking social media discourse about the 2024 US Presidential Election on Twitter/X. arXiv:2411.00376, 2024.
- [5] A. Baqir, Y. Chen, F. Diaz-Diaz, S. Kiyak, T. Louf, V. Morini, V. Pansanella, M. Torricelli, and A. Galeazzi. Unveiling the drivers of active participation in social media discourse. Scientific Reports, 15(1), 2025.
- [6] P. Barberá, J. T. Jost, J. Nagler, J. A. Tucker, and R. Bonneau. Tweeting from left to right: Is online political communication more than an echo chamber? Psychological science, 26(10), 2015.
- [7] N. Bartley, A. Abeliuk, E. Ferrara, and K. Lerman. Auditing algorithmic bias on twitter. In ACM WebSci, 2021.
- [8] R. Brewer, B. Westlake, T. Hart, and O. Arauza. The ethics of web crawling and web scraping in cybercrime research: Navigating issues of consent, privacy, and other potential harms associated with automated data collection. In Researching cybercrimes: methodologies, ethics, and critical approaches. 2021.
- [9] M. Cinelli, G. De Francisci Morales, A. Galeazzi, W. Quattrociocchi, and M. Starnini. The echo chamber effect on social media. Proceedings of the National Academy of Sciences, 118(9), 2021.
- [10] M. A. DeVito. From editors to algorithms: A values-based approach to understanding story selection in the Facebook news feed. Digital Journalism, 5(6), 2017.
- [11] P. Dunn. Online hate speech and intermediary liability in the age of algorithmic moderation. PhD thesis, University of Luxembourg, 2024.
- [12] European Commission. The Digital Service Act. https://www.europarl.europa.eu/RegData/etudes/STUD/2020/654180/EPRS˙STU(2020)654180˙EN.pdf, 2022.
- [13] M. Falkenberg, A. Galeazzi, M. Torricelli, N. Di Marco, F. Larosa, M. Sas, A. Mekacher, W. Pearce, F. Zollo, W. Quattrociocchi, et al. Growing polarization around climate change on social media. Nature Climate Change, 12(12), 2022.
- [14] C. Fiesler, N. Beard, and B. C. Keegan. No robots, spiders, or scrapers: Legal and ethical regulation of data collection methods in social media terms of service. In AAAI ICWSM, 2020.
- [15] J. Flamino, A. Galeazzi, S. Feldman, M. Macy, B. Cross, Z. Zhou, M. Serafino, A. Bovet, H. Makse, and B. Szymanski. Political polarization of news media and influencers on Twitter in the 2016 and 2020 US presidential elections. Nature Human Behaviour, 7, 2023.
- [16] T. Gillespie. Do not recommend? Reduction as a form of content moderation. Social Media + Society, 8(3), 2022.
- [17] M. J. Greenacre. Correspondence Analysis. Wiley Interdisciplinary Reviews: Computational Statistics, 2(5), 2010.
- [18] M. Grootendorst. BERTopic: Neural topic modeling with a class-based TF-IDF procedure. arXiv:2203.05794, 2022.
- [19] J. A. Hartigan and P. M. Hartigan. The dip test of unimodality. The Annals of Statistics, 1985.
- [20] N. Hassan, F. Arslan, C. Li, and M. Tremayne. Toward automated fact-checking: Detecting check-worthy factual claims by claimbuster. In ACM KDD, 2017.
- [21] H. Hosseinmardi, A. Ghasemian, M. Rivera-Lanas, M. Horta Ribeiro, R. West, and D. J. Watts. Causally estimating the effect of YouTube’s recommender system using counterfactual bots. Proceedings of the National Academy of Sciences, 121(8), 2024.
- [22] J. Jaursch, J. Ohme, and U. Klinger. Enabling Research with Publicly Accessible Platform Data: Early DSA Compliance Issues and Suggestions for Improvement. Weizenbaum Institut, 2024.
- [23] S. Jhaver, A. Bruckman, and E. Gilbert. Does transparency in moderation really matter? User behavior after content removal explanations on reddit. Proceedings of the ACM on Human-Computer Interaction, 3(CSCW), 2019.
- [24] S. Jhaver, A. Q. Zhang, Q. Z. Chen, N. Natarajan, R. Wang, and A. X. Zhang. Personalizing content moderation on social media: User perspectives on moderation choices, interface design, and labor. Proceedings of the ACM on Human-Computer Interaction, 7(CSCW2), 2023.
- [25] J. A. Jiang, P. Nie, J. R. Brubaker, and C. Fiesler. A trade-off-centered framework of content moderation. ACM Transactions on Computer-Human Interaction, 30(1), 2023.
- [26] S. Jiang, R. E. Robertson, and C. Wilson. Reasoning about political bias in content moderation. In AAAI, 2020.
- [27] K. Kupferschmidt. Twitter’s threat to curtail free data access angers scientists. Science, 379(6633), 2023.
- [28] V. La Gatta, C. Wei, L. Luceri, F. Pierri, and E. Ferrara. Retrieving false claims on Twitter during the Russia-Ukraine conflict. In The WebConf Companion, 2023.
- [29] E. Le Merrer, B. Morgan, and G. Trédan. Setting the record straighter on shadow banning. In IEEE INFOCOM, 2021.
- [30] H. B. Mann and D. R. Whitney. On a test of whether one of two random variables is stochastically larger than the other. The Annals of Mathematical Statistics, 1947.
- [31] S. Mousavi, K. P. Gummadi, and S. Zannettou. Auditing Algorithmic Explanations of Social Media Feeds: A Case Study of TikTok Video Explanations. In ICWSM, 2024.
- [32] P. Paudel, J. Blackburn, E. De Cristofaro, S. Zannettou, and G. Stringhini. Lambretta: Learning to Rank for Twitter Soft Moderation. In IEEE Symposium on Security and Privacy, 2023.
- [33] J. Pfeffer, D. Matter, and A. Sargsyan. The half-life of a tweet. In ICWSM, 2023.
- [34] S. Rathje. To tackle social-media harms, mandate data access for researchers. Nature, 633(8028), 2024.
- [35] R. E. Robertson, S. Jiang, K. Joseph, L. Friedland, D. Lazer, and C. Wilson. Auditing partisan audience bias within google search. Proceedings of the ACM on Human-Computer Interaction, 2(CSCW), 2018.
- [36] L. Savolainen. The shadow banning controversy: perceived governance and algorithmic folklore. Media, Culture & Society, 44(6), 2022.
- [37] The New York Times. Elon Musk, Matt Taibbi, and a Very Modern Media Maelstrom. https://www.nytimes.com/2022/12/04/business/media/elon-musk-twitter-matt-taibbi.html, 2022.
- [38] The New York Times. Elon Musk’s Free Speech Problem. https://www.nytimes.com/interactive/2025/04/23/business/elon-musk-x-suppression-laura-loomer.html, 2025.
- [39] The Washington Post. Shadowbanning is real: Here’s how you end up muted by social media. https://www.washingtonpost.com/technology/2022/12/27/shadowban/, 2022.
- [40] P. P. Tricomi, S. Kumar, M. Conti, and V. Subrahmanian. Climbing the Influence Tiers on TikTok: A Multimodal Study. In ICWSM, 2024.
- [41] K. Vombatkere, S. Mousavi, S. Zannettou, F. Roesner, and K. P. Gummadi. TikTok and the Art of Personalization: Investigating Exploration and Exploitation on Social Media Feeds. In The WebConf, 2024.
- [42] X. Terms of Service. https://archive.is/J6FsZ, 2024.
- [43] S. Yitzhaki and E. Schechtman. The Gini methodology: a primer on a statistical methodology. Springer, 2013.
- [44] YouTube. Continuing our work to improve recommendations on YouTube. https://youtube.googleblog.com/2019/01/continuing-our-work-to-improve.html, 2019.
- [45] S. Zannettou, T. Caulfield, W. Setzer, M. Sirivianos, G. Stringhini, and J. Blackburn. Who let the trolls out? towards understanding state-sponsored trolls. In ACM WebSci, 2019.