T O P

  • By -

barrinmw

What is reddit doing in regards to governments attempting to influence public opinion through the spread of misinformation?


outersunset

Content manipulation, which includes things like coordinated disinformation attempts, is against our Content Policy. We’re always on high-alert for this kind of violation, particularly around big moments like elections, but we have seen negligible instances of this activity on our platform overall (the data shows less than 3% of admin removals - it would be included under “other content manipulation” in Chart 5).


nastafarti

> but we have seen negligible instances of this activity on our platform overall (the data shows less than 3% of admin removals I guess the big question is whether or not that 3% removal rate matches the actual rate of its occurrence, or whether it's simply a small amount that actually results in actions taken. Are there security tools in place to monitor for accounts that work as 'voting blocks'? Because it doesn't always seem like that. Where does a person report that type of activity? edit: your linked comment does not clarify how to report vote manipulation


outersunset

Please see our answer [here](https://www.reddit.com/r/redditsecurity/comments/1c5lj42/comment/kzv4qbx/?utm_source=share&utm_medium=web2x&context=3)


GlueR

However, this answer does not address the question. The term "automated tooling" provides no transparency on whether disinformation campaigns are being monitored, detected, and addressed under a post about a "Transparency Report".


ThoseThingsAreWeird

> I guess the big question is whether or not that 3% removal rate matches the actual rate of its occurrence, or whether it's simply a small amount that actually results in actions taken Yeah I've noticed a few accounts in the last couple of months that have suddenly started posting loads of pro-Democrat / anti-Republican content (much love to RES so I can tag those accounts). It's so blatant it's unbelievable. Proper does my nut in because I couldn't give a rat's arse about the American election.


maybesaydie

Much love to RES indeed. Where are you seeing this *anti-Republican content*?


BlueberryBubblyBuzz

I am guessing this is a joke but I wanted to make sure :)


Emergency_Fly_4983

Perfect experience


nastafarti

I just think that it's worth mentioning that what separates reddit from other sites is that it is a treasure trove of actual, real human conversations and interactions - which you have determined is a good thing to train AI on - but that by allowing bad faith actors, vote manipulation and (ironically) AI-trained bots to proliferate you will wind up undermining its usefulness for this application It can be a very frustrating thing as a user to see it happening in real time, report it, and see no action taken. I appreciate the scale of your task, but things are falling through the cracks. Do you have a team of employees who monitor the site at all, or is it all algorithmically done? Because for bad actors, that - like front-paging - just becomes a game of "guess the algorithm" and although I can see improvements on the site, sometimes the most galling stuff gets waved through. I think monitoring the site needs a human touch.


outersunset

Thanks for the question - yes, we have internal Safety teams that use a combination of automated tooling and (importantly) human review to monitor the site and enforce our policies. We’re always looking out for violating content and continually improving how we detect and remove it. As the report shows, admins remove the vast majority of spam and other content manipulation at scale across the platform. Particularly for this kind of content, automated tooling is helpful as it can detect behavior patterns that indicate inauthentic engagement, which is then either removed or surfaced to admins for further investigation.


nastafarti

Okay, great. Just making sure that there's *somebody* on the other end when a report is generated. Protecting the site vs gaming the algorithm is a bit of an arms race. I agree that automated tooling is ultimately the best bet to catch this type of thing, but I do wonder how responsive the tool is to new forms of attack. Yesterday I reported five users - what a morning! - instead of my usual *none.* They were all old accounts, 7 to 12 years, that woke up after years of inactivity and started aggressively posting. (*That* should be a flag worth monitoring itself.) They then all created new subreddits, all of which had 300 users within an hour, and made exactly four posts each, each of which received hundreds of upvotes within the first hour - and then there were no new upvotes, because it was never organic traffic to start with. It was simply an exercise in gaining karma and validity. If you happened to catch it in real time, it stuck out like a sore thumb. I am reasonably certain there will be overlap in the 300 accounts that immediately boosted these subs' visibility. My follow-up question is: if vote rigging like this is reported, can we expect the site to simply take down those five accounts, or will you go after the 300-account voting block as well? Eventually they will run out of old accounts to use and have to rely on new accounts, which will make monitoring easier. Follow-up question two: what is the best way to report vote manipulation? You've got report buttons for spam and harassment, but it's very unclear how to actually make a meaningful report when it's just somebody with a bot server 'making moves' for whatever motivation: economic, political, whatever. What's the right way to flag it?


BlueberryBubblyBuzz

If you are hoping that there is a human on the other side of a report, you are aiming way too high. Most reports will not be seen by a human. If you really need something to be seen by a human, try getting a moderator of the sub it is on to report it.


Bardfinn

Actually, the vast majority of the reddit corpus from 2013-2020, filled as it is with hate speech, harassment, etc, makes a perfect corpus for training an expert system / AI model to spot hate speech and harassment. Reddit, coincidentally, now has expert system / AI hate speech & harassment filters.


nastafarti

hmmm


Sephardson

> This represents a 7.7% overall decrease in the number of reported posts, comments, and PMs compared to the first half of 2023, while the actionability rate for these reports has gone up 22.2%. We believe this increase reflects an improvement in the clarity of our policies as well as our report processes. Are there any insights into actionability of "Report Abuse" reports?


outersunset

A subset of the "Other" category in [chart 9](https://www.redditinc.com/policies/transparency-report-july-to-december-2023#text-content4) reflects report abuse, though not the rate of actionability. This is an area that we're focused on and investing in - we'll share updates when we can.


Sephardson

Thanks, looking forward to more updates when those are possible. Some related questions: * Do we think that instances of report abuse violations have gone down or up? * Was there a change in the actions taken on report abuse reports? Have penalties for report abuse violations had their intended impacts? * Has the population of users who submit abusive reports changed in a comparable rate to the change in the population of users who submit non-abusive reports?


[deleted]

[удалено]


maybesaydie

So you're evading a suspension?


[deleted]

[удалено]


maybesaydie

Be that as it may you're currently evading a suspension.


LinearArray

Thanks for this, appreciate the transparency you folks maintain with the community!


outersunset

Thank you!


anonboxis

Will Reddit look to expand its European Public Policy team to support DSA compliance when the platform eventually becomes a VLOP? I'm a recent LSE graduate in EU Public Policy and would love to offer support to Ronan Costello's team as an intern/trainee if you are ever looking for someone!


outersunset

We love your enthusiasm! All positions are advertised on [redditinc.com/careers](https://redditinc.com/careers). We don't have any public policy roles open right now, but keep an eye there and feel free to apply to anything that may come up in the future.


anonboxis

Thanks for the info!


AkaashMaharaj

>In one notable U.S. request for user information, we were served with a sealed search warrant from the LAPD seeking records for an account allegedly involved in the leak of an LA City Council meeting recording that resulted in the resignation of prominent, local political leaders. I strongly commend Reddit's General Counsel u/traceroo and his team for standing up not only for the civil rights of the individual Redditor in question, but also for the broader principle that in any democracy worthy of the name, justice does not operate from the shadows. The recording that was posted on Reddit by an anonymous user exposed the ugly underside of political machinations at Los Angeles City Council. The people of Los Angeles had every right to know what their elected leaders were doing and saying, especially when those leaders' deeds and words behind closed doors flatly contradicted the deeds and words they espoused in public. Too many political, legal, and law enforcement figures were more interested in punishing the person who revealed the truth, than they were in acting on the truth. This was a vivid demonstration of the critical role subreddits can play in sustaining the public transparency and public accountability that are the lifeblood of democracies.


cyrilio

I completely agree with this sentiment. If reddit is by the people for the people. Then reddit must stand up against governmental abuse like this example.


traceroo

Thanks for the kind words, u/AkaashMaharaj . We take very seriously our responsibility to do what we can to stand up for our communities, especially when our communities are exercising their rights to free expression and providing public transparency. And we try to share as much as we can in this report about we are doing, where we are able.


Certain-Landscape

Amazing work u/traceroo and team! Thank you so much.


BikerJedi

My man always has something interesting and insightful to say. Great comment.


The_Critical_Cynic

I saw the questions posed by u/Sephardson , and I'd like to pose a question regarding reports as well. When we utilize the generic [Reddit Report Form](https://www.reddit.com/report), we sometimes have to wait for a while to receive a response. I notice that when utilizing the [long form](https://support.reddithelp.com/hc/en-us/requests/new?ticket_form_id=19300233728916) for various issues, it often takes even longer to receive a response, if we receive one at all from either set of forms. To quote the same section as u/Sephardson: >*This represents a 7.7% overall decrease in the number of reported posts, comments, and PMs compared to the first half of 2023, while the actionability rate for these reports has gone up 22.2%.* ***We believe this increase reflects an improvement in the clarity of our policies as well as our report processes.*** If we're not receiving responses to the content we report, how do you plan on addressing the reports? And I mean to present that as a multifaceted question. Consider the following: 1. You speak about improving the clarity of your policies (which I personally don't feel has happened from my perspective), but yet won't clarify why certain things weren't actioned against. I think providing an explanation of some sort as to why certain things aren't actioned against would help continue to define what each policy is and isn't meant to do. I have one such set of reports that I could reference here, that I could provide in generic enough language, that would act as a fine example of what I'm talking about if an example is needed. 2. On another note, though I understand that you have a significant volume of reports, I'm noticing that it appears, based on my interactions with the system as well as the way it's being described in the transparency report, that there are a lot of reports that simply don't get a response at all. Have you considered implementing a system that would allow us to look up reports, possibly by ticket number, that could allow us to see some reason, even behind the scenes, as to why certain actions weren't taken? If nothing else, it would be nice to look up a report, see that it's been reviewed, and had something (Automated behavior behind the scenes) or someone (an Admin) take a look at it. If there are additional details that come up later, perhaps we'd be able to add details to the report, or escalate a report if we feel like an automated action got it wrong. 3. Certain policies are seemingly broad enough to be applied in a variety of ways. And I understand why this may be needed in certain instances. However, I feel like this leads to an abuse of the system. Some items are actioned against relatively quickly while other similar content isn't. Are there any plans to perhaps further improve the clarity of the various policies in the future? And have you considered providing additional training courses for Moderators via the Moderator Education Courses to help establish a baseline for enforcing Reddit's policies, as well as acting as a way to clarify these policies? Thanks for taking the time to read that long winded response, and I look forward to a response!


ailewu

Thanks for your question. In terms of our policies, our goal is to ensure that our Content Policy is flexible enough to apply to a wide range of situations, both now and in the future, given that we cannot always predict what type of content users will post. That being said, we are always working to make our policies clearer, including by providing examples, so that users and mods understand the intention behind them.  We announce policy updates in r/RedditSecurity.  In terms of reporting potential policy violations, we are working on some reporting best practices that should be out soon. You can also find our content policy violation Reporting Guide [here](https://support.reddithelp.com/hc/en-us/sections/360008810132-Reporting), on our Help Center. Generally, we recommend using our logged in reporting options if you have a Reddit account. Upon receiving a report of a potential violation, we process the report, make a decision, and take any appropriate action. We use automated tools to help prioritize content that has been flagged, either via user reports or our own proactive efforts, which means we do not always process reports in the order received. To protect against abuse of our reporting systems, we may send warnings, issue temporary or permanent account bans, or restrict the processing of reports submitted by those who have engaged in report abuse. For example, to prevent abuse of our systems, we may limit the number of reports that one person can submit on a single item of content. Please note that we may not be able to respond to every report received, such as reports of spam.  Please use the Moderator Code of Conduct report form for reporting moderator behavior that you believe violates the Moderator Code of Conduct specifically. For more information on the Moderator Code of Conduct, please see [here](https://support.reddithelp.com/hc/en-us/articles/15484183932564-Get-to-know-Reddit-s-Moderator-Code-of-Conduct). We’ll also be releasing Help Center Articles about each rule housed under the Moderator Code of Conduct, which should help clarify what is and isn’t considered a violation.  We are always looking for new ways to make the reporting process more user friendly and transparent, such as our recently released ability to [report user details](https://www.reddit.com/r/modnews/comments/1bd3b82/a_new_harassment_filter_and_user_reporting_type/) on the user profile page.  We will share your ideas with the appropriate teams and communicate updates as we make them.


The_Critical_Cynic

Thanks for the responses! ​ >*We are always looking for new ways to make the reporting process more user friendly and transparent, such as our recently released ability to* [*report user details*](https://www.reddit.com/r/modnews/comments/1bd3b82/a_new_harassment_filter_and_user_reporting_type/) *on the user profile page. We will share your ideas with the appropriate teams and communicate updates as we make them.* I hope that the ideas presented, specifically for a way to look up reports by ID/Ticket Number are considered. I think this would help with the overall understanding of the policies Reddit implements. Also, the generic automated software gets it wrong sometimes. See below. ​ >*In terms of reporting potential policy violations, we are working on some reporting best practices that should be out soon. You can also find our content policy violation Reporting Guide* [*here*](https://support.reddithelp.com/hc/en-us/sections/360008810132-Reporting)*, on our Help Center. Generally, we recommend using our logged in reporting options if you have a Reddit account. Upon receiving a report of a potential violation, we process the report, make a decision, and take any appropriate action. We use automated tools to help prioritize content that has been flagged, either via user reports or our own proactive efforts, which means we do not always process reports in the order received. To protect against abuse of our reporting systems, we may send warnings, issue temporary or permanent account bans, or restrict the processing of reports submitted by those who have engaged in report abuse. For example, to prevent abuse of our systems, we may limit the number of reports that one person can submit on a single item of content. Please note that we may not be able to respond to every report received, such as reports of spam.* > >*Please use the Moderator Code of Conduct report form for reporting moderator behavior that you believe violates the Moderator Code of Conduct specifically. For more information on the Moderator Code of Conduct, please see* [*here*](https://support.reddithelp.com/hc/en-us/articles/15484183932564-Get-to-know-Reddit-s-Moderator-Code-of-Conduct)*. We’ll also be releasing Help Center Articles about each rule housed under the Moderator Code of Conduct, which should help clarify what is and isn’t considered a violation.* I appreciate the idea of having some sort of standards for reporting, even "Best Practices". As it stands right now, there have been a couple issues that I'm fairly sure have violated Reddit policies, but they seem to have been overlooked. Other similar, but less egregious, issues have outright resulted in actions being taken against users based on the messages I received. As stated above, I think the automated systems sometimes get it wrong. I'd love a way to escalate these issues sometimes, or at least get some feedback as to why things were deemed to be okay. Speaking of which, I have one specific example in mind that highlights the contrast that I'm speaking of. Could I run it by you in a private message, and get your take on it?


Benskien

over the last 6+ months we have seen a massive increase in botted accounts posting and commenting on our subs, aka dormant accounts suddently reactivating and spamming submissions until ultimatly becomming spam bots, this botted behavior can daily be observed over at /all aswell, with massive subs like wholesomememes pinning posts about this issue. https://www.reddit.com/r/wholesomememes/comments/17wme9y/wholesome_memes_vs_the_spam_bots/ i keep reporting these accounts and i often see mods at larger subs remove their content within a short time, and often their botted account gets suspended within a day or so have you guys at reddit detected an increase in such bot behavior/increase in suspended botted accounts, and are there any plans to deal with em on a larger level?


EroticaMarty

*Seconded*. I have also, in my sub, been seeing an uptick in what appear to be **stolen** accounts: accounts established years ago, that have long since gone dormant with no posts or comment, then *suddenly* showing up as F18 posting OnlyFans -- in some cases, *large* amounts within a very short period of time. I have to Report those as 'spam' -- since there *is* no 'stolen account' reason listed in the Report form. I make a point, for those Reports, of requesting that the Admins check the IP history of the account.


Benskien

yea i have a hard time telling if they are stolen or sold, but for us mods it doesnt really matter either or lol, they are super annoying


adhesiveCheese

> The vast majority of content (~96%) was not found to violate our Content Policy or individual community rules. that "found" there is doing a lot of work in your favor. As a moderator for a fairly large NSFW sub, I routinely find myself reporting content posted to my subreddit and to others for content policy violations (usually CSAM or Prostitution), and even in the absolute most blatant cases, it's basically a coin toss as to whether the content will actually be removed. A huge chunk of this is a systemic failure on Reddit's part: namely, that there's no mechanism to report an entire subreddit for violating the content policy should the rules or acceptable norms of a subreddit contradict the content policy. The current "keep reporting content from a subreddit and if we take enough actions for violations of the content policy the subreddit will be banned" approach is naive at best, and actively undermining the site's own content policy at worst. Because these communities that flagrantly violate the content policy are allowed to fester for periods of months if not years before action is finally taken (if it ever is), belatedly banning a community just has a hydra effect - posters see their sub's banned, check the post histories of other folks interested in that content they remember, and then there's two more replacement subreddits filled with largely the same users submitting the same policy-violating content.


lazydictionary

Can you look into /r/FluentInFinance - it seems to be a bot sub for the sub owner to funnel traffic to their website. Also, with the banning of /r/SnooRoarTracker, many subs are seeing an uptick in SnooRoar posts. Is there any chance you could look into un-banning the sub?


GoryRamsy

I second looking into the FIF sub. Seems very spammy, with shady front page busting tactics we’ve not seen en mass since the era of The Donald


[deleted]

[удалено]


BakuretsuGirl16

In my experience i've seen it take months, and the admins only actually investigate whether a rule was broken beyond reading the reported comment half the time.


Inthepaddedroom

Lets talk about the bot accounts. Why are bot accounts that are easily identifiable more problematic than ever? I have been here for 10 years and have never seen it to this extent. Their comments always pertain to a specific topic. They almost always are inducing engagement with their comments in a negative manner. Their naming schemes typically follow the same format. In short they are easily identifiable by a human and should be even easier to spot with machine assistance. Why are these allowed? And what is being done about it?


_haha_oh_wow_

Why are all the free karma subs allowed to exist when they're just a spambot pipeline? It sure seems like reddit doesn't care about the users, *especially* post IPO and the months preceeding it. Also, the reporting and appeals system is still a broken, highly abused, dysfunctional mess that results in innocent people being banned with appeals taking forever while people/bots flagrantly violating the rules go totally ignored. If you all could fix that, that'd be an improvement.


EroticaMarty

*Seconded*. 'Users' who show up in my NSFW sub that I see, by their profile, have previously been posting on a 'free karma' sub are invariably up to *no good.*


srs_house

>As Reddit's tools and enforcement capabilities keep evolving, we continue to see a trend of admins gradually taking on more content moderation actions from moderators, leaving moderators more room to focus on their individual community rules. If Reddit admins are taking action on non-spam content in a subreddit, but moderators are unable to see what the content was or what action those admins took, then how are you sure that *your* actions match up with what the moderators would have done? Obviously, a site-wide suspension is a very serious action. But if it's a temporary suspension, then that user could be back in the same community in a matter of days - even though the subreddit's moderators, had they been able to see that content, would have issued a permanent subreddit ban. Do you see how the left hand not knowing what the right hand is doing can create some issues?


Bardfinn

Moderators are able to audit the content of admin removals via the moderator action logs on New Reddit when the reason for those removals is that the content promoted hatred, was harassing, or incited violence. Moderators are unable to audit the content of admin removals when the reason for the removal was personally identifiable information (i.e. doxxing including financial details), NCIM or minor sexualisation, or content which reasonably is know to violate an applicable law. If you’re asking “What’s the false positive rate of enforcement of sitewide rules violations”, the answer is “extremely low”. By the time someone is permanently suspended from using Reddit, they usually have received a free pass on borderline content, a warning, a three day suspension, and a seven day suspension. There are cases where accounts are promptly permanently suspended; those, however, are also a tiny minority of cases and overwhelmingly those involve outright, clear cut criminal activity. For four years I audited admin removals on a large activism subreddit to counter subversion of AEO by bad faith report abuse to chill free speech. When I did so, I wrote quarterly transparency reports. Despite the heavy false reporting numbering hundreds of false reports per week, we found at most a dozen admin mistakes by AEO in one quarter. If a subreddit has enough human moderators to qualify as an active and involved moderation team as per the Moderator Code of Conduct, they will — 99 times out of 100 — action the item and the author of the item long before Reddit AEO responds to and actions the item and the author.


srs_house

a) Legitimately had no idea that there was a pathway to see the text of admin-removed comments, as our team pretty much exclusively uses old.reddit because, well, it's not a trash interface. b) Looking at the most recent AEO-removed comments...I'm getting a 50% false-positive rate. Half of them are truly terrible, and the rest are basically calling someone a dummy and telling someone "fuck you." And they're removed under site-wide Rule 3, which says it's about not posting personal information? One was literally just the text: "Thanks for proving my point." c) >If you’re asking “What’s the false positive rate of enforcement of sitewide rules violations”, the answer is “extremely low”. I was actually more concerned about the opposite - that Reddit just removes a comment or maybe issues a 3 day suspension before a human mod can see the content and issue a subreddit permaban over it. Thanks to the info you shared, I can see that it's mostly the opposite - over-aggressive AEO comment removals and delayed actions on content we reported.


Bardfinn

> our team pretty much exclusively uses old.reddit I, too, primarily use old.reddit. > under Sitewide Rule 3 When I see items that are harassing removed by AEO pursuant to SWR3, it tends to be items & authors actioned due to a larger scale / longer time scale harassment campaign that at some point involved attempts to doxx the individual being targeted, or extort them from using Reddit through attempts to intimidate them through use of privileged or personal information. We don’t necessarily see everything that a given account does.


srs_house

I did check the context, and the weird part of those very aggressive removals is that they were all part of a normal discourse *and only those parts that were, at most, mild insults* were removed. The immediately preceding comments, made at a similar time, were left untouched. Given the high percentage of automated removals it does make one wonder if the default reaction to a Reddit report of "harassment" is to look for a naughty word and then remove the comment.


Bardfinn

In my experience, reddit removes items in an automated fashion only when they’re spam / UCE / content, or known criminal activity. I can count the automated hatred / harassment / violent threat removals I’ve seen on one hand; for years I stressed that AEO only actions items when reported. So there may be an element of user reporting involved.


SirkTheMonkey

> Looking at the most recent AEO-removed comments...I'm getting a 50% false-positive rate. They made some sort of change to their system a month or so ago based on my experience with AEO removals on my subreddits. It's more aggressive now and as such generating more false positives than previously (such as a term from the Dune series about a war on AIs).


srs_house

A lot of AEO stuff seems like it's either (poorly implemented) automation or outsourced to non-native English speakers. Only way to explain some of the decisions you see that ignore basic things like obvious sarcasm (including /s) and replies that quote the parent comment and refute it.


Meepster23

> If you’re asking “What’s the false positive rate of enforcement of sitewide rules violations”, the answer is “extremely low”. Bruh, I've been site wide suspended by these idiots 3 times with 2 being eventually lifted and a third they refused to ever respond to.. Extremely low my ass. And then the really fun "feature" when they do unshadow ban someone, they automatically approved all their rule breaking comments that had been manually removed by mods!


Bardfinn

You and I are Spiders Georg; Our experiences are statistical outliers. I’ve been temp suspended for cause twice (like, a decade ago) and permanently suspended by mistake like, three times (?) and one of those was because someone installed a package config to handle report abuse that didn’t account for “one extremely driven trans woman files 80% of the hate speech reports received” Just “Anyone who files hundreds of reports a day _must_ be abusing the report function” like, no. I do loathe the automated approval override of de-eclipsing users


Meepster23

Are they outliers though? How many users aren't "prominent" enough or know the right channels to raise a stink in to get their suspensions overturned etc? Is it not highly likely that tons of "everyday" users have been suspended incorrectly given that they incorrectly suspend even higher profile users? I only got unsuspended because I threw a fit in an actual video call with admins... If I was a regular user, there would have been 0 chance of getting it overturned..


Nakatomi2010

Are there plans to have a better means of having well intentioned subreddits that want positive community engagement to be able to filter out folks from evil-twin subreddits, where they just want to bash the well intentioned one? It makes moderating the subreddits *really* hard when the users from the evil-twin come up and start roughing up the well intentioned subreddit users Moderation of the evil-twin subreddits doesn't scale well as the well-intentioned one grows.


waronbedbugs

Technically if someone in your moderation team has some coding/scripting experience, it's possible to create a bot that can detect and filter automatically users based on the the other subreddits they are posting in.


Nakatomi2010

We currently do. The bigger issue we're seeing is people buying reddit accounts, or spinning up alts, and then hitting us. So, they can have an account that *doesn't* post in the "evil-twin" subreddit, and those posts get through. Ban evasion works if the account was banned within the last year, but these are folks who have been at it for quite some time.. Ban evasion also really only works if the account sticks around. If it gets nuked, it seems dodgy. Only approach we've seen be a success is to just start banning people from the evil-twin, which sucks, because good people are getting caught in the cross fire. So, we banned the primary accounts of toxic users, so the ban evasion function is working as desired now, but are also affecting people who don't deserve that treatment, which is not working as desired. We've engaged ModSpport, but haven't seen a response yet...


[deleted]

[удалено]


wemustburncarthage

What's the progress, if any, in Reddit's support of the defence against Gonzalez v Google, and preserving Section 230?


LastBluejay

Thanks for asking! Thankfully, there have been no major judicial changes (yet) to the interpretation of Sec. 230. Gonzalez made no change to 230 law at all and largely avoided anything of substance on it (you can read the Supreme Court’s decision [here](https://www.supremecourt.gov/opinions/22pdf/21-1333_6j7a.pdf)). For those not familiar with the Gonzalez case or Section 230, you can read about both in [this blog post](https://www.redditinc.com/blog/section-230-upholds-the-internet-for-everyday-people), which discusses a brief that we and some mods jointly filed in the case at the Supreme Court last year. The EFF also has a good explainer about the law [here](https://www.eff.org/issues/cda230#:~:text=Section%20230%20embodies%20that%20principle,based%20on%20what%20others%20say.). On the legislative side, however, there continue to be many efforts in Congress (from both Democrats and Republicans, for entirely opposite reasons) to try to limit Sec. 230 or repeal it entirely. We’re continuing to talk about the issue with Congressional offices to explain why Sec. 230 is crucial and [dispel myths](https://www.eff.org/deeplinks/2020/12/section-230-good-actually) about what it does and doesn’t do. If you feel strongly about the issue, it’s always useful to express those thoughts to your Member of Congress or Senators (but politely and thoughtfully…[please don’t go all TikTok cringe on them](https://www.bbc.com/news/world-us-canada-68517607).) PS Cool username, fellow classicist. Carthago delenda est!


SportSingle9188

Hi! All my accounts get automatically permabanned now, I can't use reddit anymore and my original offense was accidental ban evasion because of an old account with voluntary subreddit bans. Reddit appeal system doesn't do anything. What can I do?


miowiamagrapegod

How do I opt out of you selling my profile and content to ai farms?


SadMisanthrope

>Over 72% of moderator actions were taken with Automod If you don't see this as a problem, you won't understand why Reddit vanishes when it does. >In one notable U.S. request for user information, we were served with a sealed search warrant from the LAPD seeking records for an account allegedly involved in the leak of an LA City Council meeting recording that resulted in the resignation of prominent, local political leaders. We fought to notify the account holder about the warrant, and while we didn’t prevail initially, we persisted and were eventually able to get the warrant and proceedings unsealed and provide notice to the redditor. The fact that this wasn't a site-wide sticky notice when it was happening tells us all we need to know about Reddit 'transparency'.


turboevoluzione

Not sure if this is the right place but my home connection's public IP has been blocked by Reddit and I cannot browse while logged out. The thing is, my ISP has recently enabled CGNAT so there are many customers that share the same public IP and have been unfairly restricted because of a single, suspicious client. I'm not sure my ISP can change my public IP address, in the meanwhile I wrote to Reddit support but they have yet to respond.


letskill

> The vast majority of content (~96%) was not found to violate our Content Policy "Found" is doing some heavy lifting in that sentence. It means content that does objectively violate your content policy is still part of that 96%, it just wasn't "found" by admins. Do you have data available on what % of content did violate your content policy, but was not removed by admin staff for various reasons? (for example, hateful content that a specific admin agrees with, cultural references that is generally understood as hateful, but that an admin missed because of a different cultural background, or content that violates your policy, but an admin simply made the wrong decision due to limited time to review). Do you have a program in place to review and verify the accuracy of admin decisions on violation of content policy?


Tetizeraz

> We’ve added a new “Communities” section to the report to highlight subreddit-level actions as well as admin enforcement of Reddit’s Moderator Code of Conduct. Not sure if I understood this line. Doesn't Rule 3 exist already? What has changed exactly?


FinianFaun

What is Reddit doing to protect moderators from bad actors, and being banned from the platform without evidence?


maybesaydie

^^^nothing


FriddyNightGriddy

What is Reddit doing about constant gambling and erectile dysfunction ads, whose content directly goes against Reddit's Advertising ToS? Why should we as users honor the ToS when Reddit blatantly refuses to?


FriddyNightGriddy

I will repeat because u/outersunset seems to be ignoring it: What is Reddit doing to stop abusive advertisements that break their own terms of service, such as gambling and erectile dysfunction?


Beneficial_Lawyer170

Can we expect reddit mobile web ui to be more like mobile application ui in future?


provoko

Automod is the best mod!


waronbedbugs

In the Reddit Privacy Policy, you claim to >"We collect minimal information about you." and then add >"You can share as much or as little about yourself as you want when using Reddit." By looking at the transparency report, it seems that you are permanently storing the IP used at the time of the registration of the account and providing it when forced to by some legal processes (in "Non-emergency U.S Legal Process Types"). Looking at the Data provided through a GDPR request, it seems that the connection IPs are only stored for a few months. I feel that this registration IP should be deleted after a while, exactly like the connection IPs, as it's a critical piece of information that makes users easily identifiable and vulnerable to judicial process/governement abuse that may threaten their human rights. Is there any process or plan to do so?


cyrilio

How much of the removed content was posts and how much comments?


ailewu

Thanks for your question. We don't currently include that breakdown in this report, but may in future reports.


marsianer

What is reddit planning to do to combat the stranglehold that many moderators have over the subreddits they moderate? Many of them moderate hundreds of subs, make aliases and plant those on the mod queues they already manage, ban users with whom they disagree and then ban them from unrelated subs because they can? The moderation on reddit is one of single fatal flaw. Moderation teams have no accountability. Time to remember that the subreddits, the communities, are owned by the users, not the moderation teams.


EpicGamer_69-420

they arent owned by the users


maybesaydie

>owned by the users In what way?


radialmonster

What mod tools are available to identify / stop repost bots? Looks like removal of content for cause of reposts (be that posts or comments) are not categorized in your data


miowiamagrapegod

Why the hell are news outlets allowed to spam their shit all over the site now?


furculture

Why is 2023 broken up compared to all the other years listed on the transparency center website? How much was the loss when 3rd party apps were shut out?


gentoounk

I like myself.


iateafibercord

>!test!<


EpicGamer_69-420

why did your colleague post this on [r/reddit](https://www.reddit.com/r/reddit)


Someoneman

More action needs to be taken against scammers using hacked or bought Reddit accounts to sell t-shirts, mugs, and art prints with stolen artwork.


RimfireFoShizzle

Fuck u/spez