[HN Gopher] Ex-Reddit CEO on Twitter moderation
___________________________________________________________________
 
Ex-Reddit CEO on Twitter moderation
 
Author : kenferry
Score  : 580 points
Date   : 2022-11-03 12:07 UTC (10 hours ago)
 
web link (twitter.com)
w3m dump (twitter.com)
 
| pphysch wrote:
| Musk is betting on the $8 membership being a big hit, which
| immediately addresses a lot of the moderation issues.
| 
| It's gonna be a completely different paradigm than reddit.
| Herding cats into a box painted onto the ground vs. herding cats
| into a 8' high cage.
 
  | eachro wrote:
  | Reddit had the benefit of subreddit moderators policing their
  | own. Twitter has no such thing. Maybe if you squint enough, big
  | accounts block/muting bad actors in their replies can sort of
  | count as self-policing but that does not prevent the bad actor
  | from being a troll in someone else's replies.
 
| danuker wrote:
| > Spam is typically easily identified due to the repetitious
| nature of the posting frequency, and simplistic nature of the
| content (low symbol pattern complexity).
| 
| Now that we have cheap language models, you could create endless
| variations of the same idea. It's an arms race.
 
| dbrueck wrote:
| At least one missing element is that of _reputation_. I don 't
| think it should work exactly like it does in the real world, but
| the absence of it seems to always lead to major problems.
| 
| The cost of being a jerk online is too low - it's almost entirely
| free of any consequences.
| 
| Put another way, not everyone deserves a megaphone. Not everyone
| deserves to chime in on any conversation they want. The promise
| of online discussion is that everyone should have the _potential_
| to rise to that, but just granting them that privilege from the
| outset and hardly ever revoking it doesn 't work.
| 
| Rather than having an overt moderation system, I'd much rather
| see where the reach/visibility/weight of your messages is driven
| by things like your time in the given community, your track
| record of insightful, levelheaded conversation, etc.
 
  | 22c wrote:
  | > The cost of being a jerk online is too low - it's almost
  | entirely free of any consequences.
  | 
  | Couldn't agree more here.
  | 
  | Going back to the "US Postal service allows spam" comment made
  | by Yishan, well yes, the US postal service will deliver mail
  | that someone has PAID to have delivered, they've also paid to
  | have it printed. There's not a zero cost here and most
  | businesses will not send physical spam if there weren't at
  | least some return on investment.
  | 
  | One big problem not even touched by Yishan is vote
  | manipulation, or to put it in your terms, artificially boosted
  | reputation. I consider those to be problems with the platform.
  | Unfortunately, I haven't yet seen a platform that can solve the
  | problem of "you, as an individual, have ONE voice". It's too
  | easy for users to make multiple accounts, get banned, create a
  | new one, etc.
  | 
  | At the same time, nobody who's creating a platform for users
  | will want to make it HARDER for users to sign up. Recently
  | Blizzard tried to address this (in spirit) by forcing users to
  | use a phone number and not allowing "burner numbers" (foolishly
  | determined by "if your phone number is pre-paid"). It
  | completely backfired for being too exclusionary. I personally
  | hate the idea of Blizzard knowing and storing my phone number.
  | However, the idea that it should be more and more difficult or
  | costly for toxic users to participate in the platform after
  | they've been banned is not, on its own, a silly idea.
 
  | ledauphin wrote:
  | I agree with the basic idea that we want reputation, but the
  | classic concept of reputation as a single number in the range
  | (-inf, inf) is useless for solving real-world problems the way
  | we solve them in the real world.
  | 
  | Why? Because my reputation in meatspace is precisely 0 with
  | 99.9% of the world's population. They haven't heard of me, and
  | they haven't heard of anyone who has heard of me. Meanwhile, my
  | reputation with my selected set of friends and relatives is
  | fairly high, and undoubtedly my reputation with some small set
  | of people who are my enemies is fairly low. And this is all
  | good, because no human being can operate in a world where
  | everyone has an opinion about them all the time.
  | 
  | Global reputation is bad, and giving anyone a megaphone so they
  | can chime into any conversation they want is bad, full stop.
  | Megaphone-usage should not be a democratic thing where a simple
  | majority either affirms or denies your ability to suddenly make
  | everyone else listen to you. People have always been able to
  | speak to their tribes/affinity groups/whatever you want to call
  | them without speaking to the entire state/country/world, and if
  | we want to build systems that will be resilient then we need to
  | mimic that instead of pretending that reputation is a zero-sum
  | global game.
 
    | jacobr1 wrote:
    | Social reputation IRL also has transitive properties -
    | vouching from other high-rep people, or group affiliations.
    | Primitive forms of social-graph connectedness have been
    | exposed in social networks but it doesn't seem like they've
    | seen much investment in the past decade.
 
  | paul7986 wrote:
  | The Internet needs to have a verified public identity /
  | reputation system especially with deep fakes becoming more and
  | more pervasive/easier to create. Trolls can troll all they want
  | but if they want to be serious with their words then back it up
  | with his or her online verified public Internet /reputation ID.
  | 
  | If this is one of Musk's goals with Twitter he didn't overpay.
  | The Internet definitely needs such a system..has for awhile
  | now!
  | 
  | He might connect Twitter into the crypto ecosystem and that
  | along with a verified public Internet / Reputation ID system i
  | think could be powerful.
 
    | pixl97 wrote:
    | How does this system work worldwide across multiple
    | governments, is resistant to identity theft, and prevents
    | things like dictatorships from knowing exactly who you are?
 
    | jonny_eh wrote:
    | Remember keybase.io? They still exist, but not as a cross-
    | platform identity system anymore.
 
    | runako wrote:
    | It's worth noting that Twitter gets a lot of flak for
    | permanently banning people, but that those people were all
    | there under their real names. Regardless of your opinion on
    | the bans, verifying that they were indeed banning e.g. Steve
    | Bannon would not have helped the decision making process
    | around his ban any easier.
 
  | runako wrote:
  | This is a good idea, except that it assumes _reputation_ has
  | some directional value upon which everyone agrees.
  | 
  | For example, suppose a very famous TV star joins Twitter and
  | amasses a huge following due to his real-world popularity
  | independent of Twitter. (Whoever you have in mind at this
  | point, you are likely wrong.) His differentiator is he's a
  | total jerk all the time, in person, on TV, etc. He is popular
  | because he treats everyone around him like garbage. People love
  | to watch him do it, love the thrill of watching accomplished
  | people debase themselves in attempts to stay in his good
  | graces. He has a reputation for being a popular jerk, but
  | people obviously like to hear what he has to say.
  | 
  | Everyone would expect his followers to see his posts, and in
  | fact it is reasonable to expect those posts to be more
  | prominent than those of lesser-famous people. Now imagine that
  | famous TV star stays in character on the platform and so is
  | also total jerk there: spewing hate, abuse, etc.
  | 
  | Do you censor this person or not? Remember that you make more
  | money when you can keep famous people on the site creating more
  | engagement.
  | 
  | The things that make for a good online community are not
  | necessarily congruent with those that drive reputation in real
  | life. Twitter is in the unfortunate position of bridging the
  | two.
 
    | dbrueck wrote:
    | I posted some additional ideas in a reply to another comment
    | that I think addresses some of your points, but actually I
    | think you bring up a good point of another thing that is
    | broken with both offline and online communities: reputation
    | is transferrable across communities far more than it should
    | be.
    | 
    | You see this anytime e.g. a high profile athlete "weighs in"
    | on complicated geopolitical matters, when in reality their
    | opinion on that matter should count next to nothing in most
    | cases, unless in addition to being a great athlete they have
    | also established a track record (reputation) of being expert
    | or insightful in international affairs.
    | 
    | A free-for-all community like Twitter could continue to
    | exist, where there are basically no waiting periods before
    | posting and your reputation from other areas counts a lot.
    | But then other communities could set their own standards that
    | say you can't post for N days and that your incoming
    | reputation factor is 0.001 or something like that.
    | 
    | So the person could stay in character but they couldn't post
    | for awhile, and even when they did, their posts would
    | initially have very low visibility because their reputation
    | in this new community would be abysmally low. Only by really
    | engaging in the community over time would their reputation
    | rise to the point of their posts having much visibility, and
    | even if they were playing the long game and faking being good
    | for a long time and then decided to go rogue, their
    | reputation would drop quickly so that the damage they could
    | do would be pretty limited in that one community, while also
    | potentially harming their overall reputation in other
    | communities too.
    | 
    | As noted in the other post, there is lots of vagueness here
    | because it's just thinking out loud, but I believe the
    | concepts are worth exploring.
 
      | runako wrote:
      | These are good ideas that might help manage an online
      | community! On the other hand, they would be bad for
      | business! When a high-profile athlete weighs in on a
      | complicated geopolitical matter and then (say) gets the
      | continent wrong, that will generate tons of engagement
      | (money) for the platform. Plus there's no harm done. A
      | platform probably wants that kind of content.
      | 
      | And the whole reason the platform wants the athlete to post
      | in the first place is because the platform wants that
      | person's real-world reputation to transfer over. I believe
      | it is a property of people that they are prone to more
      | heavily weigh an opinion from a well-known/well-liked/rich
      | person, even if there is no real reason for that person to
      | have a smart opinion on a given topic. This likely is not
      | something that can be "fixed" by online community
      | governance.
 
  | sydd wrote:
  | But the how would you address "reputable" people spreading
  | idiotic things or fake news? How would you prevent Joe Rogan
  | spreading COVID conspiracy theories? Or Kanye's antisemitic
  | comments? Or a celebrity hyping up some NFT for a quick cash
  | grab? Or Elon Musk falling for some fake news and spreading it?
 
  | brookst wrote:
  | Maybe? Reputation systems can devolve into rewarding
  | groupthink. It's a classic "you get what you measure"
  | conundrum, where once it becomes clear that an opinion / phrase
  | / meme is popular, it's easy to farm reputation by repeating
  | it.
  | 
  | I like your comment about "track record of insightful,
  | levelheaded conversation", but that introduces another
  | abstraction. Who measures insight or levelheadedness, and how
  | to avoid that being gamed?
  | 
  | I general I agree that reputation is an interesting and
  | potentially important signal, I'm just not sure I've ever seen
  | an implementation that doesn't cause a lot of the problems it's
  | trying to solve. Any good examples?
 
    | dbrueck wrote:
    | Yeah, definitely potential for problems and downsides. And I
    | don't know of any implementations that have gotten it right.
    | And to some degree, I imagine all such systems (online or
    | not) can be gamed, so it's also important for the designers
    | of such a system to not try to solve every problem either.
    | 
    | And maybe you do have some form of moderation, but not in the
    | sense of moderation of your agreement/disagreement with ideas
    | but moderation of behavior - like a debate moderator - based
    | on the rules of the community. Your participation in a
    | community would involve reading, posting as desired once
    | you've been in a community for a certain amount of time,
    | taking a turn at evaluating N comments that have been
    | flagged, and taking a turn at evaluating disputes about
    | evaluations, with the latter 2 being spread around so as to
    | not take up a lot of time (though, having those duties could
    | also reinforce your investment in a community). The
    | reach/visibility of your posts would be driven off your
    | reputation in that community, though people reading could
    | also control how much they see too - maybe I only care about
    | hearing from more established leaders while you are more open
    | to hearing from newer / lower reputation voices too. An
    | endorsement from someone with a higher reputation counts more
    | than an endorsement from someone who just recently joined,
    | though not so huge of a difference that it's impossible for
    | new ideas to break through.
    | 
    | As far as who measures, it's your peers - the other members
    | of the community, although there needs to be a ripple effect
    | of some sort - if you endorse bad behavior, then that
    | negatively effects your reputation. If someone does a good
    | job of articulating a point, but you ding them simply because
    | you disagree with that point, then someone else can ding you.
    | If you consistently participate in the community duties well,
    | it helps your reputation.
    | 
    | The above is of course super hand-wavy and incomplete, but
    | something along those lines has IMO a good shot of at least
    | being a better alternative to some of what we have today and,
    | who knows, could be quite good.
 
      | brookst wrote:
      | > Your participation in a community would involve reading,
      | posting as desired once you've been in a community for a
      | certain amount of time, taking a turn at evaluating N
      | comments that have been flagged, and taking a turn at
      | evaluating disputes about evaluations, with the latter 2
      | being spread around so as to not take up a lot of time
      | (though, having those duties could also reinforce your
      | investment in a community).
      | 
      | This is an interesting idea, and I'm not sure it even needs
      | to be that rigorous. Active evaluations are almost a chore
      | that will invite self-selection bias. Maybe we use
      | sentiment analysis/etc to passively evaluate how people
      | present and react to posts?
      | 
      | It'll be imperfect in any small sample, but across a larger
      | body of content, it should be possible to derive metrics
      | like "how often does this person compliment a comment that
      | they also disagree with" or "relative to other people, how
      | often do this person's posts generate angry replies", or
      | even "how often does this person end up going back and
      | forth with one other person in an increasingly
      | angry/insulting style"?
      | 
      | It still feels game-able, but maybe that's not bad? Like, I
      | am going to get such a great bogus reputation by writing
      | respectful, substantive replies and disregarding bait like
      | ad hominems! That kind of gaming is maybe a good thing.
      | 
      | One fun thing is this could be implemented over the top of
      | existing communities like Reddit. Train the models,
      | maintain a reputation score externally, offer an API to
      | retrieve, let clients/extensions decide if/how to re-order
      | or filter content.
 
    | mjjjjjjj wrote:
    | This is pure hypothetical, but I bet Reddit could derive an
    | internal reputational number that is a combination of both
    | karma (free and potentially farmable) and awards (that people
    | actually pay for or that are scarce and shows what they
    | value) that would be a better signal to noise ratio than just
    | karma alone.
 
      | pixl97 wrote:
      | So a wealthy bot farmer rules this system?
 
    | fblp wrote:
    | Google search is an example of ude of site reputation (search
    | ranking) driven by where backlinks and various other site
    | quality metrics.
    | 
    | I would also say the Facebook feed also ranks based on the
    | reputation and relevance of the poster of the content.
 
      | pixl97 wrote:
      | Is google supposed to be a positive or negative example
      | here, that is with all the recent complaints about SEO spam
      | and search quality dropping?
 
  | jonny_eh wrote:
  | Soon reputation will cost only $8 a month.
 
    | VonGallifrey wrote:
    | I don't know why this meme is being repeated so much. I see
    | it everywhere.
    | 
    | Twitter Verification is only verifying that the account is
    | "authentic, notable, and active".
    | 
    | Nothing about the Verification Process changed. At least I
    | have not heard about any changes other then the price change
    | from free to $8.
 
      | dragonwriter wrote:
      | > Twitter Verification is only verifying that the account
      | is "authentic, notable, and active".
      | 
      | Musk has been very clear that it will be open to anyone who
      | pays the (increased) cost for Twitter's Blue (which also
      | will get other new features), and thus no longer be tier to
      | "notable" or "active".
      | 
      | > At least I have not heard about any changes other then
      | the price change from free to $8.
      | 
      | Its not a price change from free to $8 for Twitter
      | Verification. It is a _discontinuation_ of Twitter
      | Verification as a separate thing, but moving the (revised)
      | process and resulting checkmark to be an open-to-anyone-
      | who-pays component of Blue, which increases in cost to $8
      | /mo (currently $4.99/mo).
 
| billiam wrote:
| The best part of his engrossing Twitter thread is that he inserts
| a multitweet interstitial "ad" for his passion project promoting
| reforestation right in the middle of his spiel.
 
  | baby wrote:
  | that's the best approach to growth:
  | 
  | 1. find what's trendy
  | 
  | 2. talk about what's trendy
  | 
  | 3. in the middle or at the end of that, talk about how that
  | relates to you and your work
 
    | anigbrowl wrote:
    | I'm sure it works across the population at large but I avoid
    | doing business with people who engage in that kind of
    | manipulation. They're fundamentally untrustworthy in my
    | experience.
 
    | CamperBob2 wrote:
    | It's the best approach for flipping your bozo bit in the
    | minds of most of your readers, but I don't see how that leads
    | to "growth."
 
  | EarlKing wrote:
  | Yes... the irony of discussing signal-to-noise ratio issues
  | (i.e. spam) and then spamming up your own thread. This post
  | sponsored by Irony.
 
  | SilasX wrote:
  | Maybe it's as some kind of reinforcement of his point about
  | policing the lines between spam and non spam?
 
  | dang wrote:
  | I know, but:
  | 
  | " _Please don 't pick the most provocative thing in an article
  | or post to complain about in the thread. Find something
  | interesting to respond to instead._"
  | 
  | https://news.ycombinator.com/newsguidelines.html
 
    | choppaface wrote:
    | I find it interesting that dang's post about not just HN
    | rules but also his personal feelings about yishan's thread: *
    | appear in the same post--- there clearly is no personal
    | boundary with dang * direct replies to dang's post, now the
    | top of the comments section, are disabled
    | 
    | whenever dang tries to "correct the record" or otherwise
    | engage in frivolous squabbles with other HN commenters, I
    | really hope this article pops up in the linked material. some
    | may argue that yishan here is doing inapropriate self-
    | promotion and that might undermine trust in his message. i
    | hope HN readers notice how partial dang is, how he's used HN
    | technical features ti give his own personal feelings
    | privilege, and the financial conflicts of interest here.
 
| onetimeusename wrote:
| free speech might be self regulating. A place that gets excessive
| spam attracts no one and then there wouldn't be much motivation
| to spam it anymore.
| 
| I don't recall spam restrictions on old IRC. A moderator could
| boot you off. My own theory is having an exponential cool off
| timer on posts could be the only thing needed that still is
| technically 100% free speech.
 
  | bombcar wrote:
  | IRC had tons of independent little servers doing their own
  | thing.
  | 
  | We have huge companies spanning the entire globe; if you get
  | banned from one you're out world-wide.
  | 
  | This is where federation can help, IF it truly is a bunch of
  | smaller servers rather than ending up one large one.
 
| aksjdhmkjasdof wrote:
| I have actually worked in this area. I like a lot of Yishan's
| other writing but I find this thread mostly a jumbled mess
| without much insight. Here are a couple assorted points:
| 
| >In fact, once again, I challenge you to think about it this way:
| could you make your content moderation decisions even if you
| didn`t understand the language they were being spoken in?
| 
| I'm not sure what the big point is here but there are a couple
| parts to how this works in the real world:
| 
| 1) Some types of content removal do not need you to understand
| the language: visual content (images/videos), legal takedowns
| (DMCA).
| 
| 2) Big social platforms contract with people around the world in
| order to get coverage of various popular languages.
| 
| 3) You can use Google Translate (or other machine translation) to
| review content in some languages that nobody working in content
| moderation understands.
| 
| But some content that violates the site's policies can easily
| slip through the cracks if it's in the right less-spoken
| language. That's just a cost of doing business. The fact that the
| language is less popular will limit the potential harm but it's
| certainly not perfect.
| 
| >Here`s the answer everyone knows: there IS no principled reason
| for banning spam. We ban spam for purely outcome-based reasons: >
| >It affects the quality of experience for users we care about,
| and users having a good time on the platform makes it successful.
| 
| Well, that's the same principle that underlies all content
| moderation: "allowing this content is more harmful to the
| platform than banning it". You can go into all the different
| reasons why it might be harmful but that's the basic idea and
| it's not unprincipled at all. And not all spam is banned from all
| platforms--it could just have its distribution killed or even be
| left totally alone, depending on the specific cost/benefit
| analysis at play.
| 
| You can apply the same reasoning to every other moderation
| decision or policy.
| 
| The main thrust of the thread seems to be that content moderation
| is broadly intended to ban negative behavior (abusive language
| and so on) rather than to censor particular political topics. To
| that I say, yeah, of course.
| 
| FWIW I do think that the big platforms have taken a totally wrong
| turn in the last few years by expanding into trying to fight
| "disinformation" and that's led to some specific policies that
| are easily seen as political (eg policies about election fraud
| claims or covid denialism). If we're just talking about staying
| out of this business then sure, give it a go. High-level
| blabbering about "muh censorship!!!" without discussion of
| specific policies, is what you get from people like Musk or
| Sacks, though, and that's best met with an eye roll.
 
| Waterluvian wrote:
| If I wanted quality content, I would just do the Something Awful
| approach and charge $x per account.
| 
| If I wanted lots of eyeballs (whether real or fake) to sell ads,
| I would just pay lip service to moderation issues, while focusing
| on only moderating anything that affects my ability to attract
| advertisers.
| 
| But what I want, above all, because I think it would be hilarious
| to watch, is for Elon to activate Robot9000 on all of Twitter.
 
  | onetimeusename wrote:
  | Robot9000 really didn't improve quality in the places it was
  | deployed though and people just game it.
  | 
  | edit: that said I think Something Awful arguably has the best
  | approach to this does it not? The site is over 20 years old at
  | this point. That is absolutely ancient compared to all the
  | public message forums that exist.
 
    | Waterluvian wrote:
    | I agree. I think the SA approach is the best I've ever seen.
    | But as I'm flippantly pointing out: it only works if you
    | really only care about fostering quality social interaction.
    | 
    | The mistake SA is making is not fixating on revenue as the
    | chief KPI. ;)
 
    | idiotsecant wrote:
    | SA is also not the font of internet culture that it once was,
    | either, so clearly the price of admission is not sufficient
    | to make it successful. It seems to me it was, at most, a
    | partial contributor.
 
      | onetimeusename wrote:
      | I think it's an interesting argument about what SA is now.
      | I hear membership is growing again. It has a dedicated
      | group there. I think that's what's most interesting. That's
      | really not much different than a Reddit board in theory.
      | But Reddit boards seem to come and go constantly and suffer
      | from all sorts of problems. I am not a redditor but SA
      | seems like a better board than a specific sub reddit.
      | 
      | My point is that maybe what SA is now is the best you can
      | hope for on the internet, and it's going strong(?).
 
      | Apocryphon wrote:
      | Also, SA has "underperformed" as an internet platform-
      | Lowtax notoriously failed to capitalize on the community
      | and grow it into something bigger (and more lucrative). So
      | it remains a large old-school vBulletin-style internet
      | forum instead of a potential Reddit or even greater, albeit
      | with its culture and soul intact.
 
        | Waterluvian wrote:
        | Not suggesting you meant it this way, but there's an
        | amusing "money person with blinders on" angle to the
        | statement. It's the "what's the point of anything if
        | you're not making money?!"
 
| anonymid wrote:
| Isn't it inconsistent to both say "moderation decisions are about
| behavior, not content", and "platforms can't justify moderation
| decisions because of privacy reasons".
| 
| It seems like you wouldn't need to reveal any details about the
| content of the behavior, but just say "look, this person posted X
| times, or was reported Y times", etc... I find the author to be
| really hand-wavy around why this part is difficult.
| 
| I work with confidential data, and we track personal information
| through our system and scrub it at the boundaries (say, when
| porting it from our primary DB to our systems for monitoring or
| analysis). I know many other industries (healthcare, education,
| government, payments) face very similar issues...
| 
| So why don't any social network companies already do this?
 
  | ranger207 wrote:
  | For one, giving specific examples gives censured users an
  | excuse to do point-by-point rebuttals. In my experience, point-
  | by-point rebuttals are one of the behaviors that should be
  | considered bad behavior and moderated against because they
  | encourage the participant to think only of each point
  | individually and ignore the superlinear effect of every point
  | taken together. For another, the user can latch on specific
  | examples that seem innocuous out of context and allow them to
  | complain that their censorship was obviously heavy handed, and
  | if the user is remotely well known then it's famous person's
  | word versus random other commenters trying to add context. The
  | ultimate result is that people see supposed problems with
  | moderation far more than anyone ever says "man I sure am glad
  | that user's gone" so there's a general air of resentment
  | against the moderation and belief in its ineffectiveness
 
  | mikkergp wrote:
  | I would guess part of the problem is the the more specific the
  | social media company gets, the more nit picky the users get.
 
| ptero wrote:
| This topic was adjacent to the sugar and L-isomer comments. Which
| probably influenced my viewpoint:
| 
| Yishan is saying that Twitter (and other social networks)
| moderate bad behavior, not bad content. They just strive for
| higher SNR. It is just that specific types of content seems to be
| disproportionately responsible for starting bad behavior in
| discussions; and thus get banned. Sounds rational and while
| potentially slightly unfair looks totally reasonable for a
| private company.
| 
| But what I think is happening is that this specific moderation on
| social networks in general and Twitter in particular has pushed
| them along the R- (or L-) isomer path to an extent that a lot of
| content, however well presented and rationally argued, just
| cannot be digested. Not because it is objectively worse or leads
| into a nastier state, but simply because deep inside some
| structure is pointing in the wrong direction.
| 
| Which, to me, is very bad. Once you reach this state of mental R-
| and L- incompatibility, no middle ground is possible and the
| outcome is decided by an outright war. Which is not fun and
| brings a lot of causalties. My 2c.
 
  | wwweston wrote:
  | > a lot of content, however well presented and rationally
  | argued, just cannot be digested.
  | 
  | Can you name the topic areas where even cautious presentation
  | will not be sustained on twitter?
 
  | vanjajaja1 wrote:
  | This was my takeaway as well. Yishan is arguing that social
  | media companies aren't picking sides they're just aiming for a
  | happy community, but the end result of that is that the loudest
  | and angriest group(s) end up emotionally bullying the
  | moderation algorithm into conforming. This is precisely the
  | problem that Elon seems to be tackling.
 
| RickJWagner wrote:
| Reddit is a sewer. I don't think the Ex-CEO has demonstrated any
| moderation skills.
 
| LegitShady wrote:
| He was CEO of a company that has volunteer moderators, what he
| knows about handling moderation is tainted by the way reddit is
| structured. Also, reddit's moderation is either heavy handed or
| totally ineffective depending on the case so not sure he's the
| right person to talk to.
| 
| Also, I stopped reading when he did an ad break on a twitter
| thread. Who needs ads in twitter threads? It makes him seem
| desperate and out of touch. Nobody needs his opinion, and they
| need his opinion with ad breaks even less.
 
| protoman3000 wrote:
| I like the idea that you don't want to moderate content, but
| behavior. And it let me to these thoughts. I'm curious about your
| additions to these thoughts.
| 
| Supply moderation of psychoactive agents never worked. People
| have a demand to alter the state of their consciousness, and we
| should try to moderate demand in effective ways. The problem is
| not the use of psychoactive agents, it is the abuse. And the same
| applies to social media interaction which is a very strong
| psychoactive agent [1]. Nevertheless it can be useful. Therefore
| we want to fight abuse, not use.
| 
| I would like to put up to discussion the usage and extension of
| techniques for demand moderation in the context of social media
| interactions which we know to somewhat work already in other
| psychoactive agents. Think something like drugs education in
| schools, fasting rituals, warning labels on cigarettes, limited
| selling hours for alcohol, trading food stamps for drug addicts
| etc.
| 
| For example, assuming the platform could somehow identify abusive
| patterns in the user, it could
| 
| - show up warning labels that their behavior might be abusive in
| the sense of social media interaction abuse
| 
| - give them mandatory cool-down periods
| 
| - trick the allostasis principle of their dopamine reward system
| by doing things intermittently, e.g. by only randomly letting
| their posts to go through to other users, or only randomly allow
| them to continue reading the conversation (maybe only for some
| time), or only randomly shadow ban some posts
| 
| - make them read documents about harmful social media interaction
| abuse
| 
| - hint to them how abusive patterns in other people look like
| 
| - give limited reading or posting credits (e.g. "Should I
| continue posting in this flamewar thread and then not post
| somewhere else where I find it more meaningful at another time?")
| 
| - etc.
| 
| I would like to hear your opinions about this in a sensible
| discussion.
| 
| _________
| 
| [1] Yes, social media interaction is a psychoactive and addictive
| agent, just like any other drug or your common addiction like
| overworking yourself, but I digress. People use social media
| interactions to among others raise their anger, to feed their
| addiction to complaining, to feel a high of "being right"/owning
| it up to the libs/nazis/bigots/idiots etc., to feel like they
| learned something useful, to entertain themselves, to escape from
| reality etc. Many people suffer from compulsively or at least
| habitual abuse of social media interactions, which has been shown
| by numerous studies (Sorry, to lazy to find a paper now to cite).
| Moreover the societal effects of abuse of social media
| interactions and their dynamics and influence on democratic
| politics are obviously detrimental.
 
  | idiotsecant wrote:
  | Maybe this works on a long enough timeline, but by your analogy
  | entire generations of our population are now hopelessly
  | addicted to this particular psychoactive agent. We might be
  | able to make a new generation that is immune to it, but in the
  | mean time these people are strongly influencing policy,
  | culture, and society in ways that are directly based on that
  | addiction. This is a 'planting trees I know I will never feel
  | the shade of' situation.
 
| carapace wrote:
| > working on climate: removing CO2 from the atmosphere is
| critical to overcoming the climate crisis, and the restoration of
| forests is one of the BEST ways to do that.
| 
| As a tangent, Akira Miyawaki has developed a method for
| 'reconstitution of "indigenous forests by indigenous trees"'
| which "produces rich, dense and efficient protective pioneer
| forests in 20 to 30 years"
| 
| https://en.wikipedia.org/wiki/Akira_Miyawaki#Method_and_cond...
| 
| It's worth quoting in full:
| 
| > Rigorous initial site survey and research of potential natural
| vegetation
| 
| > Identification and collection of a large number of various
| native seeds, locally or nearby and in a comparable geo-climatic
| context
| 
| > Germination in a nursery (which requires additional maintenance
| for some species; for example, those that germinate only after
| passing through the digestive tract of a certain animal, need a
| particular symbiotic fungus, or a cold induced dorming phase)
| 
| > Preparation of the substrate if it is very degraded, such as
| the addition of organic matter or mulch, and, in areas with heavy
| or torrential rainfall, planting mounds for taproot species that
| require a well-drained soil surface. Hill slopes can be planted
| with more ubiquitous surface roots species, such as cedar,
| Japanese cypress, and pine.
| 
| > Plantations respecting biodiversity inspired by the model of
| the natural forest. A dense plantation of very young seedlings
| (but with an already mature root system: with symbiotic bacteria
| and fungi present) is recommended. Density aims at stirring
| competition between species and the onset of phytosociological
| relations close to what would happen in nature (three to five
| plants per square metre in the temperate zone, up to five or ten
| seedlings per square metre in Borneo).
| 
| > Plantations randomly distributed in space in the way plants are
| distributed in a clearing or at the edge of the natural forest,
| not in rows or staggered.
 
| atchoo wrote:
| I think you have to be quite credulous to engage in this topic of
| "twitter moderation" as if it's in good faith. It's not about
| about creating a good experience for users, constructive debate
| or even money. It's ALL about political influence.
| 
| > I`m heartened to know that @DavidSacks is involved.
| 
| I'm not. I doubt he is there because Twitter is like Zenefits,
| it's because his preoccupation over the last few years has been
| politics as part of the "New Right" Thiel, Master, Vance etc.
| running fund raisers for DeSantis and endorsing Musk's pro-
| Russian nonsense on Ukraine.
| 
| https://newrepublic.com/article/168125/david-sacks-elon-musk...
 
  | drewbeck wrote:
  | Very helpful and unfortunate context, thanks
 
  | mikkergp wrote:
  | Yeah, I saw David Sacks tweet:
  | 
  | " The entitled elite is not mad that they have to pay $8/month.
  | They're mad that anyone can pay $8/month."
  | 
  | There must be quite a few people in here who are well versed in
  | customer relations, at least in the context of a startup, can
  | anyone explain to me why Musk and Sacks seem to have developed
  | the strategy of insulting their customers and potential
  | customers?
  | 
  | I can think of two reasons
  | 
  | 1. They think twitter has a big enough most of obsessed people
  | that they can het away with whatever they want.
  | 
  | 2. They think that there really is a massive group of angry
  | "normies" they can rile up to pay $8 a month for twitter blue,
  | but isn't ironically the goal of twitter blue to get priority
  | access to the "anointed elite"? For sure I'm not paying $8 a
  | month to get access to the feeds of my friends and business
  | associates.
  | 
  | David Sacks' tweet does feel very Trumpian in a way though,
  | which supports the notion of bringing trump back and starting
  | the free speech social network.
 
    | atchoo wrote:
    | I think their general plan will be to discourage/silence
    | influential left-wing voices with enough cover to keep the
    | majority of the audience for an emboldened right-wing.
    | 
    | If thinking imaginatively, then the proposal framed as "$8/mo
    | or have your tweets deranked" is a deal they actively don't
    | want left-wingers to take. They want to be able to derank
    | their tweets with a cover of legitimacy.
    | 
    | The more they can turn this fee into a controversial "I
    | support Musk" loyalty test, the more they can discourage
    | left-wing / anti-Musk subscribers while encouraging right-
    | wing / pro-Musk subscribers who will all have their tweets
    | boosted.
    | 
    | Feels conspiratorial but it's a fee that mostly upsets
    | existing blue tick celebrities which _should_ be the last
    | group Twitter The Business would want to annoy but they are
    | the influential left-wingers. If you look at who Musk picked
    | fights with about it e.g. AOC and Stephen King, then that is
    | even more suggestive of deliberate provocation.
    | 
    | Whether planned or not, I suspect that this is how it play
    | out.
 
    | MetaWhirledPeas wrote:
    | > _can anyone explain to me why Musk and Sacks seem to have
    | developed the strategy of insulting their customers and
    | potential customers?_
    | 
    | I find it fascinating that so many people are whip-crack
    | quick to loudly criticize the $8 checkmark move.
    | 
    | How many of these critics even use Twitter?
    | 
    | And of those who do use Twitter, how can any of them know the
    | outcome of such a move? Why not just wait and observe?
 
      | mikkergp wrote:
      | I think because Elon and Co. are acting so dismissive and
      | entitled. They're acting frickin weird. Admittedly I think
      | who you think sounds more entitled depends on your
      | worldview. I do think the journalist reactions are strange,
      | but probably just because they're acting to something so
      | strange.
      | 
      | Elon is hardly describing a vision for this new version of
      | twitter that people might be inspired to spend $8 for, yes
      | something vague about plebs vs nobility, and half has many
      | ads, but his biggest call to action has been "Hey we need
      | the money". They're acting so shitty to everyone it's
      | hardly a surprise people aren't fawning in confidence back.
      | Plus I can't help but feel that these people are really
      | just echoing what everyone else is thinking. Why am I
      | paying $8 a month for Twitter?
 
        | dragonwriter wrote:
        | > Elon is hardly describing a vision for this new version
        | of twitter that people might be inspired to spend $8 for,
        | yes something vague about plebs vs nobility,
        | 
        | Yeah, Elon calls the status quo a "lords & peasants
        | system" and says that to get out of that model Twitter
        | should have a two-tier model where the paid users get
        | special visual flair, algorithmic boosts in their tweets
        | prominence and reach, and a reduced-ads feed experience
        | compared to free users.
        | 
        | And somehow doesn't see the irony.
 
    | ZeroGravitas wrote:
    | Aside from the weird elite baiting rhetoric, does this mean
    | that blue checkmark no longer means "yes, this is that famous
    | person/thing you've heard of, not an impersonator" but now
    | just means "this person gave us 8 dollars?"
 
      | mikkergp wrote:
      | I don't know if this has been clear yet. Or less than
      | famous person/thing, do you have to submit ID / use real
      | name to get verified.
 
  | lazyeye wrote:
  | I think they are arguing its about removing political influence
  | and restoring balance.
  | 
  | This will be seen as an "attack on the left" because Twitter
  | has been controlled by the left till now.
 
  | boredhedgehog wrote:
  | You are showcasing exactly the behavior he blames for the
  | degeneration of dialogue: take any kind of question and turn it
  | into a political us vs them.
 
    | atchoo wrote:
    | When the context of the discussion is twitter moderation in
    | the wake of Musk's takeover and who his team is, it's already
    | political. For Yishan to pump up Sacks and his confidence in
    | him to fix moderation, without acknowledging that today he is
    | a political operator, is close to dishonest. Contributing
    | this information is hopefully helpful.
 
    | gort19 wrote:
    | Isn't that what Sacks is doing when he talks about the
    | 'entitled elite' being 'mad'?
 
| deckard1 wrote:
| I did not see any mention of structure.
| 
| Reddit has a different structure than Twitter. In fact, go back
| to before Slashdot and Digg and the common (HN, Reddit) format of
| drive-by commenting was simply not a thing. Usenet conversations
| would take place over the course of days, weeks, or even months.
| 
| Business rules. Twitter is driven by engagement. Twitter is
| practically the birthplace of the "hot take". It's what drives a
| lot of users to the site and keeps them there. How do you control
| the temper of a site when your _goal_ is inflammatory to begin
| with?
| 
| Trust and Good Faith. When you enter into a legal contract, both
| you and the party you are forming a contract with are expected to
| operate in _good faith_. You are signaling your intent is to be
| fair and honest and to uphold the terms of the contract. Now, the
| elephant in the room here is what happens when the CEO, Elon
| Musk, could arguably (Matt Levine has done so, wonderfully) not
| even demonstrate good faith during the purchase of Twitter,
| itself. Or has been a known bully to Bill Gates regarding his
| weight or sex appeal, or simply enjoys trolling with conspiracy
| theories. What does a moderation system even mean in the context
| of a private corporation owned by such a person? Will moderation
| apply to Elon? If not, then how is trust established?
| 
| There is a lot to talk about on that last point. In the late '90s
| a site called Advogato[1] was created to explore trust metrics.
| It was not terribly successful, but it was an interesting time in
| moderation. Slashdot was also doing what they could. But then it
| all stopped with the rise of corporate forums. Corporate forums,
| such as Reddit, Twitter, or Facebook, seem to have no interest in
| these sorts of things. Their interest is in conflict: they need
| to onboard as many eyeballs as possible, as quickly as possible,
| and with as little user friction as possible. They also serve
| advertisers, who, you could argue, are the _real_ arbiters of
| what can be said on a site.
| 
| [1] https://en.wikipedia.org/wiki/Advogato
 
| jmyeet wrote:
| This is a good post.
| 
| I'm one of those who likes to bring out the "fire in a theater"
| or doxxing as the counterexample to disprove literally nobody is
| a free speech absolutist. This on top of it not being a 1A issue
| anyway because the first five words are "Congress shall pass no
| law".
| 
| But spam is a better way to approach this and show it really
| isn't a content problem but a user behaviour problem. Because
| that's really it.
| 
| Another way to put this is that the _total experience_ matters,
| meaning the experience of all users: content creators, lurkers
| _and advertisers_. Someone could go into an AA meeting and not
| shut up about scientology or coal power and you 'll get kicked
| out. Not because your free speech is being violated but because
| you're annoying and you're worsening the experience of everyone
| else you come in contact with.
| 
| Let me put it another way: just because you have a "right" to say
| something doesn't mean other people should be forced to hear it.
| That platform has a greater responsibility that your personal
| interests and that's about behaviour (as the article notes), not
| content.
| 
| As this thread notes, this is results-oriented.
 
| spaceman_2020 wrote:
| At how many tweets in the thread do you just go "maybe I should
| just write this as a blog post?"
 
| [deleted]
 
| whatshisface wrote:
| He says there is no principled reason to ban spam, but there's an
| obvious one, it isn't really speech. The same goes for someone
| who posts the same opinion everywhere with no sense of contextual
| relevance. That's not real speech, it's just posting.
 
  | ch4s3 wrote:
  | It's basically just public nuisance, like driving up and down a
  | street blaring your favorite club banger at 3AM. More
  | uncharitably its a lot like littering, public urination, or
  | graffiti.
 
    | mr_toad wrote:
    | Spammers are also using a public space for their own selfish
    | gain, which makes them freeloaders.
 
      | ch4s3 wrote:
      | That's sort of what I mean. It's like putting up a
      | billboard on someone else's property. Taking down the
      | billboard isn't about the content of the billboard but
      | rather the non-permitted use of the space.
 
| puffoflogic wrote:
| TL;DR: Run your platform to confirm to the desires of the loudest
| users. Declare anything your loudest users don't want to see to
| be "flamewar" content and remove it.
| 
| My take: "Flamebait" _is_ a completely accurate label for the
| content your loudest users don 't want to see, but it's by
| definition your loudest users who are actually doing the flaming,
| and by definition they disagree with the things they're flaming.
| So all this does is reward people for flamewars, while the
| moderators effectively crusade on behalf of the flamers. But it's
| "okay" because, by definition, the moderators are going to be
| people who agree with the political views of the loudest viewers
| (if they weren't they'd get heckled off), so the mods you
| actually get will be perfectly happy with this situation. Neither
| the mods nor the loudest users have any reason to dislike or see
| any problem with this arrangement. So why is it a problem?
| Because it leads to what I'll call a flameocracy: whoever flames
| loudest gets their way as the platform will align with their
| desires (in order to reduce how often they flame). The mods and
| the platform are held hostage by these users but are suffering
| literal Stockholm Syndrome as they fear setting off their abusers
| (the flamers).
 
| kalekold wrote:
| I wish we could all go back to phpBB forums. Small, dedicated,
| online communities were great. I can't remember massive problems
| like this back then.
 
  | sciencemadness wrote:
  | The bad actors were much less prevalent back in the heyday of
  | small phpBB style forums. I have run a forum of this type for
  | 20 years now, since 2002. Around 2011 was when link spam got
  | bad enough that I had to start writing my own bolt-on spam
  | classifier and moderation tools instead of manually deleting
  | spammer accounts. Captchas didn't help because most of the spam
  | was posted by actual humans, not autonomous bots.
  | 
  | In the past 2 years fighting spam became too exhausting and I
  | gave up on allowing new signups through software entirely. Now
  | you have to email me explaining why you want an account and
  | I'll manually create one for the approved requests. The world's
  | internet users are now more numerous and less homogeneous than
  | they were back when small forums dominated, and the worst 0.01%
  | will ruin your site for the other 99.99% unless you invest a
  | lot of effort into prevention.
 
    | pixl97 wrote:
    | Yep, if you're on the internet long enough you'll remember
    | the days before you were portscanned constantly. You'll
    | remember the days before legions of bots hammered at your
    | HTTP server. You'd remember it was rare to have some kiddie
    | DDOS your server off the internet and you had to hide behind
    | a 3rd party provider like cloudflare.
    | 
    | That internet is long dead, hence discussions like Dead
    | Internet Theory.
 
      | carapace wrote:
      | My mom still has a land-line. She gets multiple calls a
      | day, robots trying to steal an old lady's money. For this
      | we invented the telephone? the transistor?
 
        | pixl97 wrote:
        | Honestly the internet has a lot to do with this problem
        | too, stealing VOIP accounts and spoofing caller ID has
        | enabled a lot of this.
 
  | joshstrange wrote:
  | That's because they were small and often has strict rules
  | (written or not), aka moderation, about how to behave. You
  | don't remember massive problems because the bad actors were
  | kicked off. It falls apart at scale and when everyone
  | can't/won't agree on "good behavior" or "the rules" is/are.
 
  | root_axis wrote:
  | phpBB forums have always been notorious for capricious bans
  | based on the whims of mods and admins, it's just that getting
  | banned from a website wasn't newsworthy 10 years ago.
 
| PathOfEclipse wrote:
| Re: "Here`s the answer everyone knows: there IS no principled
| reason for banning spam."
| 
| The author is making the mistake that "free speech" has been
| about saying whatever you want and whenever you want. This was
| never the case, including at the time of the _founding_ of the
| U.S. constitution. There has always been a tolerance window which
| defines what you can say and what you can 't say without
| repercussions, often and usually enforced by society and societal
| norms.
| 
| The 1st amendment was always about limiting what the government
| can do to curtail speech, but, as we know, there are plenty of
| other actors in the system that have and continue to moderate
| communications. The problem with society today is that those in
| power have gotten really bad at defining a reasonable tolerance
| window, and in fact, political actors have worked hard to _shift_
| the tolerance window to benefit them and harm their opponents.
| 
| So, he makes this mistake and then builds on it by claiming that
| censoring spam violates free speech principles, but that's not
| really true. And then he tries to equate controversy with spam,
| saying it's not so much about the content itself but how it
| affects users. And that, I think leads into another major problem
| in society.
| 
| There has always been a tension between someone getting
| reasonably versus unreasonably offended by something. However, in
| today's society, thanks in part to certain identitarian
| ideologies, along with a culture shift towards the worship or
| idolization of victimhood, we've given _tremendous_ power to a
| few people to shut down speech by being offended, and vastly
| broadened what we consider reasonable offense versus unreasonable
| offense.
| 
| Both of these issues are ultimately cultural, but, at the same
| time, social media platforms have enough power to influence
| culture. If the new Twitter can define a less insane tolerance
| window and give more leeway for people to speak even if a small
| but loud or politically motivated minority of people get
| offended, then they will have succeeded in improving the culture
| and in improving content moderation.
| 
| And, of course, there is a third, and major elephant in the room.
| The government has been caught collaborating with tech companies
| to censor speech indirectly. This is a concrete violation of the
| first amendment, and, assuming Republicans gain power this
| election cycle, I hope we see government officials prosecuted in
| court over it.
 
  | slowmovintarget wrote:
  | I think that's a mischaracterization of what was written about
  | spam.
  | 
  | The author wrote that most people don't consider banning spam
  | to be free speech infringement because the act of moderating
  | spam has nothing to do with the content and everything to do
  | with the posting behavior in the communication medium.
  | 
  | The author then uses that point to draw logical conclusions
  | about other moderation activity.
  | 
  | Leading with a strawman weakens your argument, I think.
 
    | PathOfEclipse wrote:
    | Fortunately it's not a strawman. From the article:
    | 
    | =====
    | 
    | Moderating spam is very interesting: it is almost universally
    | regarded as okay to ban (i.e. CENSORSHIP) but spam is in no
    | way illegal.
    | 
    | Spam actually passes the test of "allow any legal speech"
    | with flying colors. Hell, the US Postal Service delivers spam
    | to your mailbox. When 1A discussions talk about free speech
    | on private platforms mirroring free speech laws, the
    | exceptions cited are typically "fire in a crowded theater" or
    | maybe "threatening imminent bodily harm." Spam is nothing
    | close to either of those, yet everyone agrees: yes, it`s okay
    | to moderate (censor) spam.
    | 
    | =====
    | 
    | He's saying directly that censoring spam is not supported by
    | any free speech principle, at least as he sees it, and in
    | fact our free speech laws allow spam. He also refers to the
    | idea of "allow any legal speech" as the "free-speech"-based
    | litmus test for content moderation, and chooses spam to show
    | how this litmus test is insufficient.
    | 
    | What about my framing of his argument is a strawman? it looks
    | like a flesh-and-blood man! I am saying that his litmus test
    | is an invalid or inaccurate framing, of what a platform that
    | supports free speech should be about. Even if the government
    | is supposed to allow you to say pretty close to whatever you
    | want whenever you want, it's never been an expectation that
    | private citizens have to provide the same support.
    | Individuals, institutions, and organizations have always
    | limited speech beyond what the government could enforce.
    | Therefore, "free speech" has never meant that you could say
    | whatever is legal and everyone else will just go along with
    | it.
    | 
    | On the other hand, Elon Musk's simple remark of saying that
    | he knows he's doing a good job if both political extremes are
    | equally offended shows to me that he seems to understand free
    | speech in practice better than this ex-Reddit CEO does!
    | (https://www.quora.com/Elon-Musk-A-social-media-platform-s-
    | po...)
 
      | slowmovintarget wrote:
      | For the record, I agree with your points in your original
      | post regarding the nature of free speech and with regard to
      | the Overton window for tolerable speech (if there is such a
      | thing).
      | 
      | I disagree with the notion that Yishan made a mistake in
      | how he wrote about spam. You used that as a basis for
      | disclaiming his conclusions.
      | 
      | Yishan was not making a point about free speech, he was
      | making the point that effective moderation is not about
      | free speech at all.
 
        | PathOfEclipse wrote:
        | That's a fair point. At the same time:
        | 
        | A) saying moderation is not about free speech is, I
        | think, making a point about free speech. Saying one thing
        | is unrelated to another is making a point about both
        | things.
        | 
        | B) Even framed this way, I think Yishan is either wrong
        | or is missing the point. If you want to do content
        | moderation that better supports free speech, what does
        | that look like? I think Yishan either doesn't answer that
        | question at all, or else implies that it's not solvable
        | by saying the two are unrelated. I don't think that's the
        | case, and I also think his approach of focusing less on
        | the content and more on the supposed user impact just
        | gives more power to activists who know how to use outrage
        | as a weapon. If you want your platform to better support
        | free speech, then I think the content itself should
        | matter as much or more than peoples' reaction to it, even
        | if moderating by content is more difficult. Otherwise,
        | content moderation can just be gamed by generating the
        | appropriate reaction to content you want censored.
 
| digitalsushi wrote:
| I can speak only at a Star Trek technobabble level on this, but
| I'd like it if I could mark other random accounts as "friends" or
| "trusted". Anything they upvote or downvote becomes a factor in
| whether I see a post or not. I'd also be upvoting/downvoting
| things, and being a possible friend/trusted.
| 
| I'd like a little metadata with my posts, such as how
| controversial my network voted it. The ones that are out of
| calibration, I can view, see their responses, and then I could
| see if my network has changed. It would be nice to click on a
| friend and get a report across months of how similar we vote. If
| we started drift, I can easily cull them and get my feed cleaned
| up.
 
  | cauthon wrote:
  | Unfortunately this is antithetical to the advertising-based
  | revenue model these sites operate on. There's no incentive for
  | the site to relinquish their control over what you see and
  | return it to the user.
  | 
  | On an anecdotal level, the fraction of tweets in my feed that
  | are "recommended topics" or "liked by people I follow" or
  | simply "promoted" has risen astronomically over the past few
  | months. I have a pretty curated list of follows (~100), I had
  | the "show me tweets in chronological order" box checked back
  | when that was an option, and the signal to noise ratio has
  | still become overwhelmingly unusable.
 
  | billyjobob wrote:
  | Slashdot started introducing features like this 20 years ago.
  | We thought "web of trust" would be the future of content
  | moderation, but subsequent forums have moved further and
  | further away from empowering users to choose what they want to
  | see and towards simple top down censorship.
 
    | guelo wrote:
    | It's not censorship, it's about optimizing for advertisers
    | instead of users. Which means users can't have too much
    | control. But since users won't pay advertising is the only
    | business model that works.
 
      | rdtwo wrote:
      | True you don't want to advertise on anti China stuff (even
      | if true)
 
      | rewgs wrote:
      | I wish we'd just stopped at "users won't pay" and realized
      | that, if people aren't willing to pay for it, maybe it's
      | not a product worth building.
 
    | bombcar wrote:
    | Web of trust works _too well_ and unfortunate ideas can be
    | entertained by someone you trust, which is a no-no once you
    | have top-down ideas.
    | 
    | Almost everyone has someone they trust so much that even the
    | most insane conspiracy theory would be entertained at least
    | for awhile if that trusted person brought it up.
 
      | rdtwo wrote:
      | I mean remember when the wuhan lab thing was a total
      | conspiracy theory. Or when masks were supposedly (not
      | airborne) not needed and you just had to wash your hands
      | more. All sorts of fringe stuff sometimes turns out to be
      | true. But you know sometimes you get pizzagate but it's the
      | price we pay.
 
      | TMWNN wrote:
      | >Almost everyone has someone they trust so much that even
      | the most insane conspiracy theory would be entertained at
      | least for awhile if that trusted person brought it up.
      | 
      | What's wrong with that?
      | 
      | If someone you trust brings up a crazy idea, it _should_ be
      | considered. Maybe for a day, maybe for an hour, maybe for a
      | half second, but it shouldn 't be dismissed immediately no
      | matter what.
 
      | SV_BubbleTime wrote:
      | Can't have people being exposed to diversity of thought!
      | Not by people they like, what if someone is wrong!
      | 
      | ... I'm all about low hanging fruit in moderation, and not
      | trying to fix human behavior. I'll keep waiting to see when
      | that is back in vogue.
 
  | Melatonic wrote:
  | Very interesting idea and I think this could definitely be
  | useful. Then again users could just create new accounts.
 
  | robocat wrote:
  | I would frame it more the opposite: I can easily see comments
  | that are worthless to me, and I want to automatically downvote
  | any comments by the same user (perhaps needs tags: there are
  | users whose technical opinions are great, but I'm not
  | interested at all in their jokes or politics).
  | 
  | One problem with your idea is that moderation votes are
  | anonymous, but ranking up positively or negatively depending on
  | another users votes would allow their votes to be deanonymised.
  | 
  | Perhaps adding in deterministic noise would be enough to offset
  | that? Or to prevent deanonimisation you need a minimum number
  | of posting friends?
  | 
  | In fact I would love to see more noise added to the HN comments
  | and a factor to offset early voting. Late comments get few
  | votes because they have low visibility, which means many great
  | comments don't bubble to the top.
 
    | mjamesaustin wrote:
    | In this thread, Yishan promotes an app called Block Party
    | that lets you block anyone, or even all accounts that liked
    | or retweeted something.
    | 
    | https://twitter.com/blockpartyapp_
 
      | bombcar wrote:
      | BlockParty is almost the definition of echo chamber
      | construction.
 
| wormslayer666 wrote:
| I got my first experience in running a small-medium sized (~1000
| user) game community over the past couple years. This is mostly
| commentary on running such a community in general.
| 
| Top-level moderation of any sufficiently cliquey group (i.e. all
| large groups) devolves into something resembling feudalism. As
| the king of the land, you're in charge of being just and meting
| out appropriate punishment/censorship/other enforcement of rules,
| as well as updating those rules themselves. Your goal at the end
| of the day is continuing to provide support for your product,
| administration/upkeep for your gaming community, or whatever else
| it was that you wanted to do when you created the platform in
| question. However, the cliques (whether they be friend groups,
| opinionated but honest users, actual political camps, or any
| other tribal construct) will always view your actions through a
| cliquey lens. This will happen no matter how clear or consistent
| your reasoning is, unless you fully automate moderation (which
| never works and would probably be accused of bias by design
| anyways).
| 
| The reason why this looks feudal is because you still must curry
| favor with those cliques, lest the greater userbase eventually
| buys into their reasoning about favoritism, ideological bias, or
| whatever else we choose to call it. At the end of the day, the
| dedicated users have _much_ more time and energy to argue, or
| propagandize, or skirt rules than any moderation team has to
| counteract it. If you 're moderating users of a commercial
| product, it hurts your public image (with some nebulous impact on
| sales/marketing). If you're moderating a community for a game or
| software project, it hurts the reputation of the community and
| makes your moderators/developers/donators uneasy.
| 
| The only approach I've decided unambiguously works is one that
| doesn't scale well at all, and that's the veil of secrecy or
| "council of elders" approach which Yishan discusses. The king
| stays behind the veil, and makes as few public statements as
| possible. Reasoning is only given insofar as is needed to explain
| decisions, only responding directly to criticism as needed to
| justify actions taken anyways. Trusted elites from the userbase
| are taken into confidence, and the assumption is that they give a
| marginally more transparent look into how decisions are made, and
| that they pacify their cliques.
| 
| Above all, the most important fact I've had to keep in mind is
| that the outspoken users, both those legitimately passionate as
| well as those simply trying to start shit, are a tiny minority of
| users. Most people are rational and recognize that
| platforms/communities exist for a reason, and they're fine with
| respecting that since it's what they're there for. When
| moderating, the outspoken group is nearly all you'll ever see.
| Catering to passionate, involved users is justifiable, but must
| still be balanced with what the majority wants, or is at least
| able to tolerate (the "silent majority" which every demagogue
| claims to represent). That catering must also be done carefully,
| because "bad actors" who seek action/change/debate for the sake
| of stoking conflict or their own benefit will do their best to
| appear legitimate.
| 
| For some of this (e.g. spam), you can filter it comfortably as
| Yishan discusses without interacting with the content. However,
| more developed bad actor behavior is really quite good at
| blending in with legitimate discussion. If you as king recognize
| that there's an inorganic flamewar, or abuse directed at a user,
| or spam, or complaint about a previous decision, you have no
| choice but to choose a cudgel (bans, filters, changes to rules,
| etc) and use it decisively. It is only when the king appears weak
| or indecisive (or worse, absent) that a platform goes off the
| rails, and at that point it takes immense effort to recover it
| (e.g. your C-level being cleared as part of a takeover, or a
| seemingly universally unpopular crackdown by moderation). As a
| lazy comparison, Hacker News is about as old as Twitter, and any
| daily user can see the intensive moderation which keeps it going
| despite the obvious interest groups at play. This is in spite of
| the fact that HN has _less_ overhead to make an account and begin
| posting, and seemingly _more_ ROI on influencing discussion (lots
| of rich /smart/fancy people _post_ here regularly, let alone
| read).
| 
| Due to the need for privacy, moderation fundamentally cannot be
| democratic or open. Pretty much anyone contending otherwise is
| just upset at a recent decision or is trying to cause trouble for
| administration. Aspirationally, we would like the general
| _direction_ of the platform to be determined democratically, but
| the line between these two is frequently blurry at best. To avoid
| extra drama, I usually aim to do as much discussion with users as
| possible, but ultimately perform all decisionmaking behind closed
| doors -- this is more or less the  "giant faceless corporation"
| approach. Nobody knows how much I (or Elon, or Zuck, or the guys
| running the infinitely many medium-large discord servers)
| actually take into account user feedback.
| 
| I started writing this as a reply to paradite, but decided
| against that after going far out of scope.
 
| blfr wrote:
| > Because it is not TOPICS that are censored. It is BEHAVIOR.
| 
| > (This is why people on the left and people on the right both
| think they are being targeted)
| 
| An enticing idea but simply not the case for any popular existing
| social network. And it's triply not true on yishan's reddit which
| both through administrative measures and moderation culture
| targets any and all communities that do not share the favoured
| new-left politics.
 
  | gambler wrote:
  | He is half-correct, but not in a good way. When people on the
  | left say something that goes against new-left agenda, they get
  | suppressed too. That is not a redeeming quality of the system
  | or an indicator of fairness. It simply shows that the ideology
  | driving moderation is even more narrow-minded and intolerant of
  | dissent than most observers assume at first sight.
  | 
  | At the same time, it's trivial to demonstrate that YouTube and
  | Twitter (easy examples) primarily target conservatives with
  | their "moderation". Just look at who primarily uses major
  | alternative platforms.
 
    | AhmedF wrote:
    | > it's trivial to demonstrate that YouTube and Twitter (easy
    | examples) primarily target conservatives with their
    | "moderation".
    | 
    | Terrible take.
    | 
    | Actual data analysis shows that at _worst_ conservatives are
    | moderated equally, and at best, less than non-conservatives.
    | 
    | Here's something to chew on: https://forward.com/fast-
    | forward/423238/twitter-white-nation...
 
    | archagon wrote:
    | Or consider that perhaps the right in particular tends to
    | harbor and support people who lean more towards
    | disinformation, hate speech, and incitement to violence.
 
      | [deleted]
 
      | [deleted]
 
  | pj_mukh wrote:
  | I mean that's the claim, the counter-claim would require a
  | social network banning _topics_ and not behavior. Note: As a
  | user you can see topics, you can 't see behavior. The fact that
  | some users flood other users' DM's is not visible to all users.
  | So how do you know?
  | 
  | "I don't trust left-y CEO's", is a fair enough answer, but
  | really that's where the counter-claim seems to end. Now that we
  | have a right-wing CEO, looks like the shoe is on the other
  | foot[1]
  | 
  | [1]
  | https://twitter.com/AOC/status/1588174959817658368?s=20&t=F9...
 
    | luckylion wrote:
    | > As a user you can see topics, you can't see behavior.
    | 
    | True, but not really a good argument for the "trust us, this
    | person needed banning, no we will not give any details"-style
    | of moderation that most companies have applied so far. And
    | you can see topics, so you'll generally notice when topic are
    | being banned, not behavior, because they usually don't align
    | perfectly.
 
  | jtolmar wrote:
  | All the communist subreddits are in constant hot water with the
  | admins. They banned ChapoTrapHouse when it was one of the
  | biggest subreddits. When a bunch of moderators tried to protest
  | against reddit hosting antivax content, the admins took control
  | over those subreddits.
  | 
  | So no, you're just factually wrong here.
 
    | throw_m239339 wrote:
    | > All the communist subreddits are in constant hot water with
    | the admins.
    | 
    | Yet, the biggest communist sub of all, r/politics is doing
    | perfectly fine.
    | 
    | "moderating behavior"? Bullshit, when unhinged redditors are
    | constantly accusing conservatives of being "traitors",
    | "nazis", there is so much unhinged comments on these subs
    | that it clearly demonstrate a general left wing bias when it
    | comes to moderation, in favor of the most extreme left.
    | 
    | Chapotraphouse was only banned because they harassed the
    | wrong sub, but when it was about harassing people on subs
    | deemed not progressives, reddit admins didn't care a bit.
 
      | archagon wrote:
      | r/politics is "communist"? That's... just a really dumb
      | take. If there is a far-left presence on Reddit, it is not
      | prominent. r/politics is left-leaning and angry, but,
      | objectively speaking, not really all that extreme.
      | 
      | And, for what it's worth, it seems perfectly reasonable to
      | label those who tried to overthrow our democratic
      | government "traitors".
 
        | throw_m239339 wrote:
        | > r/politics is "communist"? That's... just a really dumb
        | take. If there is a far-left presence on Reddit, it is
        | not prominent. r/politics is left-leaning and angry, but,
        | objectively speaking, not really all that extreme.
        | 
        | Obviously, none of the affluent idiots from chapo or at
        | hexbear controlling r/politics r/news or r/worldnews are
        | really communists, they are just rich asshats that just
        | pretend to be, my point still stands. They are still
        | spouting marxist non sense, violent speech, and their
        | behavior isn't moderated, as long as they don't target
        | "the wrong people".
 
        | ineptech wrote:
        | This kind of sentiment always shows up in this kind of
        | thread; I think a lot of people don't really grok that
        | being far enough to one side causes an unbiased forum to
        | appear biased against you. If you hate X enough, Reddit
        | and Twitter are going to seem pro-X, regardless of what X
        | is.
        | 
        | (And, separately, almost no one who argues about anything
        | being "communist" is using the same definition of that
        | word as the person they're arguing with, but that's a
        | different problem entirely)
 
      | jonwithoutanh wrote:
      | ... you ever try and post anything in /r/conservative? you
      | get banned. Doesn't matter if you are quoting something
      | president Trump had said 1000 times before. You get banned.
      | You can't even quote them back to themselves, or ask them a
      | question. You get banned.
      | 
      | Persecution fetish much buddy?
 
        | throw_m239339 wrote:
        | > Persecution fetish much buddy?
        | 
        | It seems that some people here can't help making
        | everything about their sick sexual thrills. And I'm not
        | your buddy.
 
      | AhmedF wrote:
      | Well at least you're mask-off in regards to being bad
      | faith.
 
        | throw_m239339 wrote:
        | > Well at least you're mask-off in regards to being bad
        | faith.
        | 
        | And you have no arguments, so you resort to personal
        | attacks. That makes you a troll.
 
  | Natsu wrote:
  | Yeah, the "there's no principled reason to ban spam" is just
  | silly. The recipients don't want to see it whereas people cry
  | censorship when messages they want to see are blocked.
  | 
  | It's literally the difference between your feed being filtered
  | by your choices & preferences and someone else imposing theirs
  | upon you.
 
    | archagon wrote:
    | I see many more people crying censorship when messages that
    | they want _others_ to see are blocked.
 
      | Natsu wrote:
      | You must hang out in a very different place, then. I see
      | much more outcry when 3rd parties come between willing
      | speakers and recipients, with most of the rest being people
      | misrepresenting censorship as moderation because it allows
      | them to justify it.
      | 
      | In that vein, this went up on HN the other day:
      | https://astralcodexten.substack.com/p/moderation-is-
      | differen...
 
  | kodt wrote:
  | There are plenty of right wing / politically conservative subs.
 
    | Fervicus wrote:
    | Funny that you never see them on the front page.
 
  | bnralt wrote:
  | Indeed, I've seen several subs put in new rules where certain
  | topics aren't allowed to be discussed at all, because the
  | administrators told them that the sub would get banned if the
  | users went against the beliefs held by the admins (even if the
  | admins had a minority opinion when it came to the country as a
  | whole).
  | 
  | Then there is just arbitrary or malicious enforcement of the
  | rules. /r/Star_Trek was told by admins they would be banned if
  | they talked about /r/StarTrek at all, so now that's a topic
  | that's no longer allowed in that sub. But there are tons of
  | subs set up specifically to talk about other subs, where just
  | about all posts are about other subs (such as
  | /r/subredditdrama), and the admins never bother them.
  | 
  | I don't think we can have a conversation about moderation when
  | people are pretending that the current situation doesn't exist,
  | and that moderation is only ever done for altruistic reasons.
  | It's like talking about police reform but pretending that no
  | police officer has ever done anything wrong and not one of them
  | could ever be part of a problem.
 
    | thepasswordis wrote:
    | /r/srd and /r/againsthatesubreddits exist with the _explicit_
    | purpose of brigading other subs, and yet they are not banned.
 
      | UncleMeat wrote:
      | SRD has an explicit "no brigading other subs" rule. How is
      | their explicit purpose brigading other subs?
 
        | Manuel_D wrote:
        | Kiwifarms has an explicit "don't interact with the
        | subjects" rule. Does that mean it never happens?
 
        | UncleMeat wrote:
        | Brigading absolutely happens in SRD. We can talk about
        | whether this style of content should exist, but it does
        | not "exist with the explicit purpose of brigading other
        | subs."
 
        | Manuel_D wrote:
        | Right, it exists with the tacit purpose of brigading
        | other subs. But like Kiwifarms, blurbs in the site rules
        | mean nothing given the context of the community.
 
        | thepasswordis wrote:
        | "Hey guys no brigading okay? ;-)" followed by a page
        | which directly links to threads for people to brigade.
        | 
        | They don't even bother to use the np.reddit "no
        | participation" domain. Most other subs don't even allow
        | you to _link_ outside the sub, because they 've been
        | warned by admins about brigading.
        | 
        | Their rules barely even mention brigading:
        | https://www.reddit.com/r/subredditdrama/wiki/rules, and
        | you have to go to the expanded version of the rules to
        | find even _this_ , which just says not to _vote_ in
        | linked threads.
        | 
        | Literally the entire purpose of this sub is to brigade
        | and harass other subs. Their politics align with those of
        | the admins, though, so it's allowed. It is _blatant_
        | bullying at the tacit encouragement of the people running
        | the site.
 
        | UncleMeat wrote:
        | IIRC, np was the norm for many years and it just didn't
        | actually change anything. Oodles of people do get banned
        | from SRD for commenting in linked threads. The easiest
        | way to see this is when month+ old threads get linked.
        | Only the admins can see downvoting patterns.
        | 
        | Is simply linking to other threads on reddit sufficient
        | for you to consider something promoting brigading?
 
        | bnralt wrote:
        | > Is simply linking to other threads on reddit sufficient
        | for you to consider something promoting brigading?
        | 
        | As I mentioned previously, linking to other subs, or even
        | simply _talking_ about /r/StarTrek, was enough for admins
        | to accuse /r/Star_Trek of brigading. They threatened to
        | shut them down unless they stopped members from doing
        | that, and so you're not allowed to do it in the sub
        | anymore.
        | 
        | Whether you think that linking to other subs is brigading
        | or not, it's clear that admins call it brigading when
        | they want to shut down subs, yet then let continue on
        | much larger subs dedicated to the act as long as the
        | admins like the sub.
        | 
        | Edit: For example, here's a highly upvoted SRD post
        | talking about the admins threatening /r/Star_Trek if they
        | mention /r/StarTrek[1]. They call /r/Star_Trek linking to
        | /r/StarTrek posts to complain about them "brigading," in
        | the same post that they themselves are linking to a
        | /r/Star_Trek post in order to complain about it.
        | 
        | [1] https://www.reddit.com/r/SubredditDrama/comments/tuem
        | 1m/the_...
 
        | bombcar wrote:
        | The big difference with Reddit is if _posters_ get
        | banned, or if the subreddit gets banned.
        | 
        | The "good" subreddits will get posters banned left and
        | right, but accounts are cheap and they return.
        | 
        | The "bad" subreddits get banned.
 
        | KarlKode wrote:
        | What I got from similar subreddits (e.g.
        | /r/bestoflegaladvice) is that you'll get (shaddow)banned
        | really fast if you click a link in the subreddit and
        | coment on the linked post.
        | 
        | Just mentioning this because I agree with the point you
        | make (in general).
 
        | Zak wrote:
        | > _Most other subs don 't even allow you to link outside
        | the sub, because they've been warned by admins about
        | brigading._
        | 
        | I joined reddit in 2005 and have moderated several
        | subreddits. The admins have never imposed anything
        | resembling that on any subreddit I have moderated. I have
        | a suspicion they impose it when they see a large amount
        | of brigading behavior.
        | 
        | Perhaps it's not applied in an entirely fair or even
        | manner, but I suspect it's only applied when there's an
        | actual problem.
 
  | shafoshaf wrote:
  | I'm not sure we could tell the difference. As Yishan states,
  | the proof of the behavior isn't being made public because of
  | the exposure to creating new issues. Without that, you would
  | never know.
  | 
  | As for specific platforms, aka Reddit, how can one be sure that
  | right wingers on that platform aren't in fact more likely to
  | engage in bad behavior that left wingers? It might be because
  | they are being targeted, but it could also be that that group
  | of people on that platform tend to act more aggressively.
  | 
  | I am NOT saying that I know if Reddit is fair in its
  | moderation, I just don't know.
 
  | realgeniushere wrote:
 
  | oneneptune wrote:
  | Can you elaborate with an example? I'm unfamiliar with reddit
  | and it's content management. I'm unsure if the premise of "AI"
  | moderation is true, how it could moderate beyond a pattern or
  | behavior since it can't reasonably be scanning every post and
  | comment for political affiliation?
 
| urbandw311er wrote:
| https://threadreaderapp.com/thread/1586955288061452289.html
 
| wasmitnetzen wrote:
| As once famously said by Mark Twain: "I didn't have time to write
| a short Twitter thread, so I wrote a long one instead."
 
  | marban wrote:
  | Make that "I didn't have time to write a tweet, so I wrote a
  | Twitter thread instead."
 
    | wasmitnetzen wrote:
    | I thought about that phrasing, but it was too far from the
    | original quote for me. But yes, it also works and is the
    | better joke as well.
 
| gryBrd1987 wrote:
| Twitter is text based. Video games have had text based profanity
| filters for online games for years.
| 
| Make it easy for users to define a regex list saved locally. On
| the backend train a model that filters images of gore and
| genitals. Aim users who opt in to that experience at that
| filtered stream.
| 
| This problem does not require a long winded thesis.
 
  | crummy wrote:
  | why do you think nobody else has tried this / had any success
  | with this approach?
 
    | gryBrd1987 wrote:
    | Because we focus on abstract problem statements, coded
    | appeals to authority (as if ex-Reddit CEO is that special;
    | there are a few), rather than concrete engineering?
    | 
    | User demand to control what they see is there. It's why TV
    | was successful; don't like what's on History? Check out
    | Animal Planet.
    | 
    | Tech CEOs need their genius validated and refuse to concede
    | anyone else knows what's best for themselves. What everyone
    | else sees is a problem for a smurt CEO to micromanage to
    | death, of course.
 
| thrwaway349213 wrote:
| What yishan is missing is that the point of a council of experts
| isn't to effectively moderate a product. The purpose is to
| deflect blame from the company.
| 
| It's also hilarious that he says "you can`t solve it by making
| them anonymous" because a horde of anonymous mods is precisely
| how subreddits are moderated.
 
| [deleted]
 
| StanislavPetrov wrote:
| >Why is this? Because it has no value? Because it`s sometimes
| false? Certainly it`s not causing offline harm.
| 
| >No, no, and no.
| 
| Fundamentally disagree with his take on spam. Not only does spam
| have no value, it has negative value. The content of the spam
| itself is irrelevant when the same message is being pushed out a
| million times and obscuring all other messages. Reducing spam
| through rate-limiting is certainly the easiest and most impactful
| form of moderation.
 
| linuxftw wrote:
| What a bunch of long-winded babble. Incredulously, he's shilling
| an app at the end of this.
| 
| I don't agree that this is an interesting submission, and IMO
| there's no new information here.
 
| gist wrote:
| > No, you can`t solve it by making them anonymous, because then
| you will be accused of having an unaccountable Star Chamber of
| secret elites (especially if, I dunno, you just took the company
| private too). No, no, they have to be public and "accountable!"
| 
| This is bulls... Sorry.
| 
| Who cares what you are accused of doing?
| 
| Why does it matter if people perceive that there is a star
| chamber. Even that reference. Sure the press cares and will make
| it an issue and tech types will care because well they have to
| make a fuss about everything and anything to remain relevant.
| 
| After all what are grand juries? (They are secret). Does the fact
| that people might think they are star chambers matters at all?
| 
| You see this is exactly the problem. Nobody wants to take any
| 'heat'. Sometimes you just have to do what you need to do and let
| the chips fall where they fall.
| 
| The number of people who might use twitter or might want to use
| twitter that would think anything at all about this issue is
| infinitesimal.
 
| blantonl wrote:
| This really was an outstanding read and take on Elon, Twitter,
| and what's coming up.
| 
| But it literally could not have been posted in a worse medium for
| communicating this message. I felt like I had to pat my head and
| rub my tummy at the same time reading through all this, and to
| share it succinctly with colleagues resulted in me spending a
| good 15 minutes cutting and pasting the content.
| 
| I've never understood people posting entire blog type posts
| to.... Twitter.
 
  | threeseed wrote:
  | It was incoherent rambling and none of really works for
  | Twitter.
  | 
  | Twitter is ultimately at the behest of its advertisers who are
  | constantly on a knife edge about whether to bother using it or
  | not. We have already seen GM and L'Oreal pull ad spend and many
  | more will follow if their moderation policies are not in-line
  | with community standards.
  | 
  | If Musk wants to make Twitter unprofitable then sure relax the
  | moderation otherwise might want to keep it the same.
 
    | filoleg wrote:
    | L'Oreal didn't pull their twitter ad spend[0].
    | 
    | 0. https://www.reuters.com/business/retail-consumer/loral-
    | suspe...
 
      | threeseed wrote:
      | FT disagrees: https://www.ft.com/content/17281b81-b801-4a8f
      | -8065-76d3ffb40...
 
| 0xbadcafebee wrote:
| It seemed interesting but after the 400th tweet I lost interest
| and went to do something productive
 
| dariusj18 wrote:
| Does anyone else think it's brilliant that he put advertisements
| inside his own thread?
 
  | drewbeck wrote:
  | If it was for some random app or gadget I'd be mad but it's
  | literally trying to save humanity so I give it a pass. We need
  | to be talking about mitigating and surviving catastrophic
  | climate change more, not less.
 
  | LegitShady wrote:
  | more like "oh never click on yishan threads ever again, this
  | guy wants to put ads in twitter threads, who has time and
  | patience for that? not me."
  | 
  | Brilliant? For immediately getting large amounts of readers to
  | click away and discrediting himself into the future, sure that
  | might be brilliant I guess.
  | 
  | It makes him seem desperate for attention and clueless.
 
  | luuuzeta wrote:
  | >brilliant
  | 
  | More like weird and unexpected
 
  | klyrs wrote:
  | I found it to be an interesting exercise of "spam is protected
  | speech." I mean, I hated it, but it really did drive the point
  | home.
 
| TheCapeGreek wrote:
| I like yishan's content and his climate focus, but this "we
| interrupt your tweet thread for sponsored content" style tangent
| is a bit annoying - not directly for doing it or its content, but
| because I can see other thread writers picking this up and we end
| up the same as Youtube with sponsored sections of content that
| you can't ad block _.
| 
| _ FWIW With YT you can block them with Sponsorblock, which works
| with user submitted timestamps of sponsored sections in videos.
| If this tweet technique takes off I'd imagine a similar idea for
| tweets.
 
  | syncmaster913n wrote:
  | > but this "we interrupt your tweet thread for sponsored
  | content" style tangent is a bit annoying
  | 
  | I found this hilarious. I don't use Twitter and so was unaware
  | that these annoying tangents are common on the platform. As a
  | result, I thought Yishan was using them to illustrate how it's
  | not necessarily the content (his climate initiative) but a
  | specific pattern of behavior (saying the 'right' thing at the
  | wrong time, in this case) that should be the target of
  | moderation.
  | 
  | In real life we say: "it's not what you said, it's just the
  | _way_ you said it! " Perhaps the digital equivalent of that
  | could be: "it's not what you said, it's just _when_ you said
  | it. "
 
    | bombcar wrote:
    | And it's funny because if you could "downvote/upvote"
    | individual tweets in that tweet storm, his "off topic" tweets
    | would be downvoted into oblivion.
    | 
    | I think the _fundamental problem_ with the internet today is
    | that _by definition_ almost, ads are unwanted content and
    | have to be forced on the user.
 
  | kybernetyk wrote:
  | While many YouTube videos provide very interesting content most
  | twitter ,,threads" are just inane ramblings by some blue
  | checkmark. So for yt videos I go the extra steps to install an
  | extension. For twitter though? I just close the tab and never
  | return.
  | 
  | How can people who are not totally dopamine deprived zombies
  | find twitter and this terrible ,,thread" format acceptable?
  | Just write a coherent blog post pls.
 
    | klodolph wrote:
    | Dopamine-deprived zombies?
    | 
    | I don't find threads hard to read. There's some extra
    | scrolling, but it's still in linear order.
    | 
    | People post on Twitter because it reaches people, obviously.
 
  | tkk23 wrote:
  | >, but this "we interrupt your tweet thread for sponsored
  | content" style tangent is a bit annoying
  | 
  | It is annoying but it can be seen as part of his argument. How
  | can spam be moderated if even trustworthy creators create spam?
  | 
  | According to him, it's not spam because it doesn't fulfill the
  | typical patterns of spam, which shows that identifying noise
  | does require knowledge of the language.
  | 
  | It could be interesting to turn his argument around. Instead of
  | trying to remove all spam, a platform could offer the tools to
  | handle all forms of spam and let its users come up with clever
  | ways to use those tools.
 
| Fervicus wrote:
| > Our current climate of political polarization makes it easy to
| think...
| 
| Stopped reading there. Reddit I think is one of the biggest
| offenders of purposely cultivating a climate of political
| polarization.
 
  | crummy wrote:
  | how come you stopped reading there?
 
    | Fervicus wrote:
    | My second statement answers that question. I don't want
    | moderation advice from someone who was involved in a platform
    | that purposely sets moderation policies to create political
    | polarization. A comment by someone below sums it up nicely.
    | 
    | > ...and it's triply not true on yishan's reddit which both
    | through administrative measures and moderation culture
    | targets any and all communities that do not share the
    | favoured new-left politics.
    | 
    | In yishan's defense however, I am not sure if those problems
    | with reddit started before or after he left.
 
      | hackerlight wrote:
      | > favoured new-left politics.
      | 
      | Citation needed. r/ChapoTrapHouse was banned, and there's
      | many large alt-right subreddits in existence right now that
      | haven't been banned (like r/tucker_carlson).
 
| csours wrote:
| Is there a better name than "rational jail" for the following
| phenomenon:
| 
| We are having a rational, non-controversial, shared-fact based
| discussion. Suddenly the first party in the conversation goes off
| on a tangent and starts saying values or emotions based
| statements instead of facts. The other party then gets angry and
| or confused. The first party then gets angry and or confused.
| 
| The first party did not realize they had broken out of the
| rational jail that the conversation was taking place in; they
| thought they were still being rational. The second party detected
| some idea that did not fit with their rational dataset, and
| detected a jailbreak, and this upset them.
 
| cansirin wrote:
| trying out.
 
| UI_at_80x24 wrote:
| I've always thought that slashdot handled comment moderation the
| best. (And even that still had problems.)
| 
| In addition to that these tools would help:
| 
| (1)Client-side: Being able to block all content from specific
| users and the replies to specific users.
| 
| (2)Server-side: If userA always 'up votes' comments from userB
| apply a negative weighting to that upvote (so it only counts as
| 0.01 of a vote). Likewise, with 'group-voting'; if userA, userB,
| and userC always vote identically down-weight those votes. (this
| will slow the 'echo chamber' effect)
| 
| (3)Account age/contribution scale: if userZ has been a member of
| the site since it's inception, AND has a majority of their posts
| up-voted, AND contributes regularly, then give their votes a
| higher weighted value.
| 
| Of course these wouldn't solve everything, as nothing ever will
| address every scenerio; but I've often thought that these things
| combined with how slashdot allowed you to score between -1 to 5,
| AND let you set the 'post value' to 2+, 3+, or 4+ would help
| eliminate most of the bad actors.
| 
| Side note: Bad Actors, and "folks you don't agree with" should
| not be confused with each other.
 
  | sbarre wrote:
  | One thing that's easy to forget is that super-complex weighted
  | moderation/voting systems can get computationally expensive at
  | the scale of something like Twitter or Facebook etc..
  | 
  | Slashdot had a tiny population, relatively speaking, so could
  | afford to do all that work.
  | 
  | But when you're processing literally millions of posts a
  | minute, it's a different order of magnitude I think.
 
    | qsort wrote:
    | > Slashdot had a tiny population...
    | 
    | Tiny, specific, non-generalist population. As soon as that
    | changed, /. went down the drain like everything else. I still
    | like /.'s moderation system better than most, but the
    | specifics of how the system works are a second order concern
    | at best.
 
      | bombcar wrote:
      | The real problem with ALL moderation systems is Eternal
      | September.
      | 
      | Once the group grows faster than some amount, the new
      | people never get assimilated into the group, and the group
      | dies.
      | 
      | "Nobody goes there, it's too popular."
 
  | haroldp wrote:
  | > I've always thought that slashdot handled comment moderation
  | the best.
  | 
  | A limited number of daily moderation "points". A short list of
  | moderation reasons. Meta-moderation.
 
  | com2kid wrote:
  | > Server-side: If userA always 'up votes' comments from userB
  | apply a negative weighting to that upvote
  | 
  | This falls down, hard, in expert communities. There are a few
  | users who are extremely knowledgeable that are always going to
  | get upvoted by long term community members who acknowledge that
  | expert's significant contributions.
  | 
  | > Being able to block all content from specific users and the
  | replies to specific users.
  | 
  | This is doable now client side, but when /. was big, it had to
  | be done server side, which is where I imagine all the limits
  | around friend/foes came from!
  | 
  | The problem here is, trolls can create gobs of accounts easily,
  | and malevolent users group together to create accounts and
  | upvote them, so they have plenty of spare accounts to go
  | through.
 
  | joemi wrote:
  | I wonder about your (2) idea... If the goal is to reduce the
  | effect of bots that vote exactly the same, then ok, sure.
  | (Though if it became common, I'm sure vote bots wouldn't have a
  | hard time being altered to add a little randomness to their
  | voting.) But I'm not sure how much it would help beyond that,
  | since if it's not just identifying _exact_ same voting, then
  | you're going to need to fine tune some percentage-the-same or
  | something like that. And I'm not sure the same fine-tuned
  | percentage is going to work well everywhere, or even across
  | different threads or subforums on the same site. I also feel
  | like (ignoring the site-to-site or subforum-to-subforum
  | differences) that it would be tricky to fine tune correctly to
  | a point where upvotes still matter. (Admittedly I have nothing
  | solid to base this on other than just a gut feeling about it.)
  | 
  | It's an interesting idea, and I wonder what results people get
  | when trying it.
 
| hunglee2 wrote:
| "there will be NO relation between the topic of the content and
| whether you moderate it, because it`s the specific posting
| behavior that`s a problem"
| 
| some interesting thoughts from Yishan, a novel way to look at the
| problem.
 
  | mikkergp wrote:
  | I thought this point was overstated, twitter certainly has some
  | controversial content related rules and while as the CEO of
  | Reddit he may have been mostly fighting macro battles, there
  | are certainly content related things that both networks censor.
 
    | bink wrote:
    | Reddit's content policy has also changed a LOT since he was
    | CEO. While the policy back then may not have been as loose as
    | "is it illegal?" it was still far looser than what Reddit has
    | had to implement to gain advertisers.
 
    | hunglee2 wrote:
    | I think the unstated caveat would be 'anything illegal' but
    | yes, the point was _overstated_ , though, I think, still
    | stands
 
| greenie_beans wrote:
| that digression into plugging his start-up was gross!
 
| quadcore wrote:
| I think tiktok is doing incredibly well in this regards and in
| almost every social network aspect. Call me crazy but I now
| prefer the discussions there as HN's most of the time. I find
| high-quality comments (and there is still good jokes in the
| middle). The other day I felt upon a video about physics which
| had the most incredibly deep and knowlegeable comments Ive ever
| seen ( _edit: found the video, it is not as good as I remembered
| but still close to HN level imo_ ). It's jaw dropping how well it
| works.
| 
| There is classical content moderation (the platform follows local
| laws) but mostly it kind of understand you so well that it put
| you right in the middle of like minded people. At least it feels
| that way.
| 
| I dont have insider hinsights on how it trully works I can only
| guess but the algorithm feels like a league or two above
| everything I have seen so far. It feels like it understand people
| so well that it prompted deep thought experiments on my end. Like
| let say I want to know someone I could simple ask "show me your
| tiktok". It's just a thought experiments but it feels like tiktok
| could tell how good of a person you are or more precisely what is
| your level of personal development. Namely, it could tell if
| youre racist, it could tell if youre a bully, a manipulator or
| easily manipulated, it could tell if youre smart (in the sense of
| high IQ), if you have fine taste, if you are a leader or a
| loner... And on and on.
| 
| Anyway, this is the ultimate moderation: follow the law and
| direct the user to like minded people.
 
  | ProjectArcturis wrote:
  | >mostly it kind of understand you so well that it put you right
  | in the middle of like minded people
  | 
  | Doesn't this build echo chambers where beliefs get more and
  | more extreme? Good from a business perspective (nobody gets
  | pissed off and leaves because they don't see much that they
  | object to). But perhaps bad from a maintaining-democracy
  | perspective?
 
    | [deleted]
 
| motohagiography wrote:
| I've had to give this some thought for other reasons, and after a
| couple decades solving analogous problems to moderation in
| security, I agree with yishan about signal to noise over the
| specific content, but what I have effectively spent a career
| studying and detecting with data is a single factor: malice.
| 
| It's something every person is capable of, and it takes a lot of
| exercise and practice with higher values to reach for something
| else when your expectations are challenged, and often it's an
| active choice to recognize the urge and act differently. If there
| were a rule or razor I would make on a forum or platform, it's
| that all content has to pass the bar of being without malice.
| It's not "assume good intent," it's recognizing that there are
| ways of having very difficult opinions without malice, and one
| can have conventional views that are malicious, and
| unconventional ones that are not. If you have ever dealt with a
| prosecutor or been on the wrong side of a legal dispute, these
| are people fundamentally actuated by malice, and the similar
| prosecution of ideas and opinions (and ultimately people) is what
| wrecks a forum.
| 
| It's not about being polite or civil, avoiding conflict, or even
| avoiding mockery and some very funny and unexpected smackdowns
| either. It's a quality that in being universally capable of it, I
| think we're also able to know it when we see it. "Hate," is a
| weak substitute because it is so vague we can apply it to
| anything, but malice is ancient and essential. Of course someone
| malicious can just redefine malice the way they have done other
| things and use it as an accusation because words have no meaning
| other than as a means in struggle, but really, you can see when
| someone is actuated by it.
| 
| I think there is a point where a person decides, consciously or
| not, that they will relate to the world around them with malice,
| and the first casulty of that is an alignment to honesty and
| truth. What makes it useful is that you can address malice
| directly and restore an equillibrium in the discourse, whereas
| accusations of hate and others are irrevocable judgments. I'd
| wonder if given it's applicability, this may be the tool.
 
  | creeble wrote:
  | I think it's two things: the power of malice, and popularity
  | measurement. Malice and fame.
  | 
  | Social networks are devices for measuring popularity; if you
  | took the up/down arrows off, no one would be interested in
  | playing. And we have proven once again that nothing gets up
  | arrows like being mean.
  | 
  | HN has the unusual property that you can't (readily) see
  | others' score, just your own. That doesn't really make it any
  | less about fame, but maybe it helps.
  | 
  | When advertising can fund these devices to scale to billions,
  | it's tough to be optimistic about how it reflects human nature.
 
  | dang wrote:
  | I might be misunderstanding what you mean by malice, but in
  | that case it's probably not the best word for what you're
  | describing. I'd be interested in a different description if you
  | want to write one. I definitely don't think that malice is
  | something you can just-see and make accurate judgments about,
  | let alone detect with data.
  | 
  | For me, malice relates to intent. Intent isn't observable. When
  | person X makes a claim about Y's intent, they're almost always
  | filling in invisible gaps using their imagination. You can't
  | moderate on that basis. We have to go by effects, not intent (h
  | ttps://hn.algolia.com/?dateRange=all&page=0&prefix=true&que...)
  | .
  | 
  | It took me a long time to (partially) learn that if I tell a
  | user "you were being $foo" where $foo relates to their intent,
  | then (1) they can simply deny it and no one can prove
  | otherwise, making the moderation position a weak one; and (2)
  | mostly they will deny it sincerely because they never had such
  | an intent, not consciously at least. Now you've given them a
  | reason to feel entirely in the right, and if you moderate them
  | anyway, they will feel treated unjustly. This is a way to
  | generate bad blood, make enemies, and lose the high ground.
  | 
  | The reverse strategy is much better: describe the _effects_ of
  | someone 's posts and explain why they are bad. When inevitably
  | they respond with "but my intent was ABC", the answer is "I
  | believe you [what else can you say about something only that
  | person could know?], but nonetheless the effects were XYZ and
  | we have to moderate based on effects. Intent doesn't
  | communicate itself--the burden is on the commenter to
  | disambiguate it." (https://hn.algolia.com/?dateRange=all&page=0
  | &prefix=true&que...)
  | 
  | Often when people get moderated in this way, they respond by
  | writing the comment they originally had in their head, as a
  | sort of defense of what they actually meant. It's astonishing
  | what a gap there often is between the two. Then you can respond
  | that if they had posted that in the first place, it would have
  | been fine, and that while they know what they have in their
  | head when posting, the rest of us have no access to that--it
  | needs to be spelled out explicitly.
  | 
  | Being able to tell someone "if you had posted that in the first
  | place, it would have been fine" is an _extremely_ strong
  | moderation position, because it takes off the table the idea
  | "you're only moderating me because you dislike my opinions",
  | which is otherwise practically ubiquitous.
 
    | japhyr wrote:
    | Kate Manne, in Down Girl, writes about the problems with
    | using intent as the basis for measuring misogyny. Intent is
    | almost always internal; if we focus on something internal, we
    | can rarely positively identify it. The only real way to
    | identify it is capturing external expressions of intent:
    | manifestos, public statements, postings, and sometimes things
    | that were said to others.
    | 
    | If you instead focus on external effects, you can start to
    | enforce policies. It doesn't matter about a person's intent
    | if their words and actions disproportionately impact women.
    | The same goes for many -isms and prejudice-based issues.
    | 
    | A moderator who understands this will almost certainly be
    | more effective than one who gets mired in back-and-forths
    | about intent.
    | 
    | https://bookshop.org/p/books/down-girl-the-logic-of-
    | misogyny...
 
    | motohagiography wrote:
    | This aspect of people writing what they meant again after
    | being challenged and it being different - I'd assert that
    | when there _is_ malice (or another intent) present, they
    | double down or use other tactics toward a specific end other
    | than improving the forum or relationship they are
    | contributing to. When there is none, you get that different
    | or broader answer, which is really often worth it. However,
    | yes it is intent, as you identify.
    | 
    | I have heard the view that intent is not observable, and I
    | agree with the link examples that the effect of a comment is
    | the best available heuristic. It is also consistent with a
    | lot of other necessary and altruistic principles to say it's
    | not knowable. On detecting malice from data, however, the
    | security business is predicated on detecting intent from
    | network data, so while it's not perfect, there are precedents
    | for (more-) structured data.
    | 
    | I might refine it to say that intent is not _passively_
    | observable in a reliable way, as if you interrogate the
    | source, we get revealed intent. On the intent taking place in
    | the imagination of the observer, that 's a deep question.
    | 
    | I think I have reasonably been called out on some of my views
    | being the artifacts of the logic of underlying ideas that may
    | not have been apparent to me. I've also challenged authors
    | with the same criticism, where I think there are ideas that
    | are sincere, and ones that are artifacts of exogenous intent
    | and the logic of other ideas, and that there is a way of
    | telling the difference by interrogating the idea (via the
    | person.)
    | 
    | I even agree with the principle of not assuming malice, but
    | professionally, my job has been to assess it from indirect
    | structured data (a hawkish, is this an attack?) - whereas I
    | interpret the moderator role as assessing intent directly by
    | its effects, but from unstructured data (is this
    | comment/person causing harm?).
    | 
    | Malice is the example I used because I think it has persisted
    | in roughly its same meaning since the earliest writing, and
    | if that subset of effectively 'evil' intent only existed in
    | the imaginations of its observers, there's a continuity of
    | imagination and false consciousness about their relationship
    | to the world that would be pretty radical. I think it's right
    | to not assume malice, but fatal to deny it.
    | 
    | Perhaps there is a more concrete path to take than my
    | conflating it with the problem of evil, even if on these
    | discussions of global platform rules, it seems like a useful
    | source of prior art?
 
    | bombcar wrote:
    | I would attribute to malice things like active attempts to
    | destroy the very forum - spamming is a form of "malice of the
    | commons".
    | 
    | You will know when you encounter malice because nothing will
    | de-malice the poster.
    | 
    | But if it is not malice; you can even take what they said and
    | _rewrite_ it for them in a way that would pass muster. In
    | debate this is called steelmanning - and it 's a very
    | powerful persuasion method.
 
      | Zak wrote:
      | Spamming is an attempt to promote something. Destroying the
      | forum is a side effect.
      | 
      | It's fair to describe indifference to negative effects of
      | one's behavior as malicious, and it is, indeed almost never
      | possible to transform a spammer into a productive member of
      | a community.
 
        | bombcar wrote:
        | Yeah, most people take the promotion spamming as the main
        | one, but you can also refer to some forms of shitposting
        | as spamming (join any twitch chat and watch whatever the
        | current spam emoji is flood by) - but the second is more
        | almost a form of cheering perhaps.
        | 
        | If you wanted to divide it further I guess you could
        | discuss "in-group spamming" and "out-group spamming"
        | where almost all of the promotional stuff falls in the
        | second but there are still some in the first group.
 
        | Zak wrote:
        | I guess I'd describe repeatedly posting the same emoji to
        | a chat as flooding rather than spamming. Even then, your
        | mention of _cheering_ further divides it into two
        | categories of behavior:
        | 
        | 1. Cheering. That's as good a description as any. This is
        | intended to express excitement or approval and rally the
        | in-group. It temporarily makes the chat useless for
        | anything else, but that isn't its purpose.
        | 
        | 2. Flooding. This is an intentional denial of service
        | attack intended to make the chat useless for as long as
        | possible, or until some demand made by the attacker is
        | met.
 
    | kashyapc wrote:
    | Hi, dang! I wonder if it makes sense to add a summarized
    | version of your critical point on "effects, not intent" to
    | the HN guidelines. Though, I fear there might be undesirable
    | _ill effects_ of spelling it out that way.
    | 
    | Thanks (an understatement!) for this enlightening
    | explanation.
 
      | dang wrote:
      | There are so many heuristics like that, and I fear making
      | the guidelines so long that no one will read them.
      | 
      | I want to compound a bunch of those explanations into a
      | sort of concordance or whatever the right bibliographic
      | word is for explaining and adding to what's written else
      | where (so, not concordance!)
 
        | kashyapc wrote:
        | Fair enough. Yeah your plans of "compounding" and
        | "bibliographic concordance" (thanks for the new word)
        | sound good.
        | 
        | I was going to suggest this (but scratch it, your above
        | idea is better): A small section called "a note on
        | moderation" (or whatever) with hyperlinks to "some
        | examples that give a concrete sense of how moderation
        | happens here". There are many _excellent_ explanations
        | buried deep in the the search links that you post here.
        | Many of them are a valuable riffing on [internet] human
        | nature.
        | 
        | As a quick example, I love your lively analogy[1] of a
        | "boxer showing up at a dance/concert/lecture" for
        | resisting flammable language here. It's funny and a
        | cutting example that is impossible to misunderstand. It
        | (and your other comment[2] from the same thread) makes so
        | many valuable _reminders_ (it 's easy to forget!). An
        | incomplete list for others reading:
        | 
        | - how to avoid the "scorched earth" fate here;
        | 
        | - how "raw self-interest is fine" (if it gets you to
        | curiosity);
        | 
        | - why you can't "flamebait others into curiosity";
        | 
        | - why the "medium" [of the "optionally anonymous internet
        | forum"] matters;
        | 
        | - why it's not practical to replicate the psychology of
        | "small, cohesive groups" here;
        | 
        | - how the "burden is on the commenter";
        | 
        | - "expected value of a comment" on HN; and much more
        | 
        | It's a real shame that these useful heuristics are buried
        | so deep in the comment history. Sure, you do link to them
        | via searches whenever you can; that's how I discovered
        | 'em. But it's hard to stumble upon otherwise. Making a
        | sampling of these easily accessible can be valuable.
        | 
        | [1] 3rd paragraph here:
        | https://news.ycombinator.com/item?id=27166919
        | 
        | [2] https://news.ycombinator.com/item?id=27162386
 
        | bombcar wrote:
        | Commentary or gloss on the text, I believe, is sometimes
        | used.
 
    | AnimalMuppet wrote:
    | But the difference between the original post and the revised
    | post often _is_ malice (or so I suspect). The ideas are the
    | same, though they may be developed a bit more in the second
    | post. The difference is the anger /hostility/bitterness
    | coloring the first post, that got filtered out to make the
    | second post.
    | 
    | I think that maybe the observable "bad effects" and the
    | unobservable "malice" may be almost exactly the same thing.
 
    | Eliezer wrote:
    | This exchange ought to be a post in its own right. It seems
    | to me that malice, hate, Warp contamination, whatever you
    | want to call it, _is_ very much a large part of the modern
    | problem; and also it 's a true and deep statement that you
    | should moderate based on effects and not tell anyone what
    | their inner intentions were, because you aren't sure of those
    | and most users won't know them either.
 
  | nostrebored wrote:
  | I find most forums that advocate against behavior they view as
  | malicious wind up becoming hugboxes as people skirt this
  | arbitrary boundary. I will never, never come back to platforms
  | or groups after I get this feeling.
  | 
  | Hugbox environments wind up having a loose relationship with
  | the truth and a strong emphasis on emotional well-being.
  | 
  | Setting your moderation boundaries determines the values of
  | your platform. I'd much rather talk to someone who wants to
  | hurt my feelings than someone who is detached from reality or
  | saying what they think.
 
  | danans wrote:
  | > "Hate," is a weak substitute because it is so vague we can
  | apply it to anything
  | 
  | That is a big stretch. Hate can't be applied to many things,
  | including disagreements like this comment.
  | 
  | But it can be pretty clearly applied to statements that, if
  | carried out in life, would deny another person or peoples'
  | human rights. Another is denigration or mocking someone on the
  | basis of things that can't or shouldn't have to change about
  | themselves, like their race or religion. There is a pretty
  | bright line there.
  | 
  | Malice (per the conventional meaning of something bad intended,
  | but not necessarily revealed or acted out) is a much lower bar
  | that includes outright hate speech.
  | 
  | > but really, you can see when someone is actuated by it.
  | 
  | How can you identify this systematically (vs it being just your
  | opinion), but not identify hate speech?
 
    | Manuel_D wrote:
    | Hate absolutely can, and is, applied to disagreements: Plenty
    | of people consider disagreement around allowing natal males
    | in women's sport is hateful. Plenty of people consider
    | opposition to the abolishment of police is hateful. Plenty of
    | people immigration enforcement hateful. I could go on...
 
      | bombcar wrote:
      | Any disagreement is classified as _hate_ now; the word is
      | empty and worthless.
      | 
      | We cannot regulate the internal forum, only the actions we
      | perceive.
 
      | danans wrote:
      | > Plenty of people consider disagreement around allowing
      | natal males in women's sport is hateful. Plenty of people
      | consider opposition to the abolishment of police is
      | hateful. Plenty of people immigration enforcement hateful.
      | 
      | Those things aren't deemed hate speech, but they might be
      | disagreed with and downvoted on some forums (i.e. HN), and
      | championed on others (i.e. Parler) but that has nothing to
      | do with them being hate speech. They are just unpopular
      | opinions in some places, and I can understand how it might
      | bother you if those are your beliefs and you get downvoted.
      | 
      | Actual hate speech based on your examples is: promoting
      | violence/harassment against non-cisgender people, promoting
      | violence/harassment by police, and promoting
      | violence/harassment by immigration authorities against
      | migrants.
      | 
      | Promoting violence and harassment is a fundamentally
      | different type of speech than disagreeing with the
      | prevailing local opinion on a controversial subject that
      | has many shades of gray (that your examples intentionally
      | lack).
 
| antod wrote:
| For some reason, this makes me wonder how Slashdot's moderation
| would work in the current age. Too nerdy? Would it get
| overwhelmed by today's shit posters?
 
  | bombcar wrote:
  | People don't care enough about the "community" anymore. It
  | might work on a smallish-scale but the reality is everything is
  | shitposting, even here.
  | 
  | Even in Slashdot's heyday the number of metamoderators was
  | vanishingly small. The best thing it had was the ability to
  | filter anonymous cowards and the ability to browse from -5 to
  | +5 if you wanted to.
 
| nkotov wrote:
| Is anyone else having a hard time following along? Can someone
| provide a tl;dr?
 
  | Consultant32452 wrote:
  | The public thinks about moderation in terms of content. Large
  | social networks think in terms of behavior. Like let's say I
  | get a chip on my shoulder about... the Ukraine war, one
  | direction or another. And I start finding a way to insert my
  | opinion on every thread. My opinion on the Ukraine war is fine.
  | Any individual post I might make is fine and contextual to the
  | convo. But I'm bringing down the over-all quality of discussion
  | by basically spamming every convo with my personal grievance.
  | 
  | Some kinds of content also gets banned, like child abuse
  | material and other obvious things. But the hard part is the
  | "behavior" type bans.
 
    | oceanplexian wrote:
    | > Any individual post I might make is fine and contextual to
    | the convo. But I'm bringing down the over-all quality of
    | discussion by basically spamming every convo with my personal
    | grievance.
    | 
    | Isn't this how a healthy society functions?
    | 
    | Political protests are "spam" under your definition. When
    | people are having a protest in the street, it's inconvenient,
    | people don't consent to it, it brings down the quality of the
    | discussion (Protestors are rarely out to have a nuanced
    | conversation). Social Media is the public square in the 21st
    | century, and people in the public square should have a right
    | to protest.
 
| jchw wrote:
| The commentary is interesting, but it does unfortunately gloss
| over the very real issue of actually controversial topics. Most
| platforms don't typically set out to ban controversial stuff from
| what I can tell, but the forces that be (advertisers, government
| regulators, payment processors, service providers, etc.) tend to
| be quite a bit more invested in such topics. Naughty language on
| YouTube and porn on Twitter are some decent examples; these are
| _not_ and never have been signal to noise ratio problems. While
| the media may be primarily interested in the problem of content
| moderation as it impacts political speech, I 'd literally filter
| all vaguely politically charged speech (even at the cost of
| missing plenty of stuff I'd rather see) if given the option.
| 
| I think that the viewpoints re: moderation are very accurate and
| insightful, but I honestly have always felt that it's been more
| of a red herring for the actual scary censorship creep happening
| in the background. Go find the forum threads and IRC logs you
| have from the 2000s and think about them for a little while. I
| think that there are many ways in which I'd happily admit the
| internet has improved, but looking back, I think that a lot of
| what was discussed and how it was discussed would not be
| tolerated on many of the most popular avenues for discourse today
| --even though there's really nothing particularly egregious about
| them.
| 
| I think this is the PoV that one has as a platform owner, but
| unfortunately it's not the part that I think is interesting. The
| really interesting parts are always off on the fringes.
 
  | wwweston wrote:
  | It's hard for me to imagine what "scary actual censorship" is
  | happening -- that is, to identify topics or perspectives that
  | cannot be represented in net forums. If such
  | topics/perspectives exist, then the effectiveness must be near
  | total to the point where I'm entirely unaware of them, which I
  | guess would be scary if people could provide examples. But
  | usually when I ask, I'm supplied with topics which I have
  | indeed seen discussed on Twitter, Reddit, and often even HN,
  | so...
 
    | cvwright wrote:
    | Nobody wants to answer this, because to mention a
    | controversial topic is to risk being accused of supporting
    | it.
    | 
    | You could look at what famous people have gotten into trouble
    | over. Alex Jones or Kanye West. I assume there have been
    | others, but those two were in the news recently.
 
    | jchw wrote:
    | The problem is that it's not really about censorship the way
    | that people think about it; it's not about blanket banning
    | the discussion of a topic. You can clearly have a discussion
    | about extremely heated debate topics like abortion,
    | pedophilia, genocide, whatever. However, in some of these
    | topics there are pretty harsh chilling effects that prevent
    | people from having very open and honest discussion about
    | them. The reason why I'm being non-specific is twofold: one
    | is because I am also impacted by these chilling effects, and
    | another is because making it specific makes it seem like it's
    | about a singular topic when it is about a recurring pattern
    | of behaviors that shift topics over time.
    | 
    | If you really don't think there have been chilling effects, I
    | put forth two potential theories: one is that you possibly
    | see this as normal "consequences for actions" (I do not
    | believe this: I am strictly discussing ideas and opinions
    | that are controversial even in a vacuum.) OR: perhaps you
    | genuinely haven't really seen the fringes very much, and
    | doubt their existence. I don't really want to get into it,
    | because it would force me to pick specific examples that
    | would inextricably paint me into those arguments, but I guess
    | maybe it's worth it if it makes the point.
 
      | wwweston wrote:
      | > The problem is that it's not really about censorship the
      | way that people think about it; it's not about blanket
      | banning the discussion of a topic.
      | 
      | Then we're far away enough from the topic of censorship
      | that we should be using different language for what we're
      | discussing. It's bad enough that people use the term
      | "censorship" colloquially when discussing private refusal
      | to carry content vs state criminalization. It's definitely
      | not applicable by the time we get to soft stakes.
      | 
      | As someone whose life & social circle is deeply embedded in
      | a religious institution which makes some claims and
      | teachings I find objectionable, I'm pretty familiar with
      | chilling effects and other ways in which social stakes are
      | held hostage over what topics can be addressed and how. And
      | yet I've found these things:
      | 
      | (1) It's taught me a lot about civil disagreement and
      | debate, including the fact that more often than not, there
      | are ways to address _even literally sacred topics_ without
      | losing the stakes. It takes work and wit, but it 's
      | possible. Those lessons have been borne out later when I've
      | chosen to do things like try to illuminate merits in pro-
      | life positions while in overwhelmingly pro-choice forums.
      | 
      | (2) It's made me appreciate the value of what the courts
      | have called time/place/manner restrictions. Not every venue
      | is or should be treated the same. Church services are the
      | last time/place to object to church positions, and when one
      | _does_ choose that it 's best to take on the most
      | obligation in terms of manner, making your case in the
      | terms of the language, metaphors, and values of the church.
      | 
      | (3) Sometimes you have to risk the stakes, and the only
      | world in which it would actually be possible for there NOT
      | to be such stakes would be one in which people have no
      | values at all
 
| ramblerman wrote:
| Did he begin answering the question, drop some big philosophical
| terms, and then just drift off into here is what I think we
| should do about climate change in 4 steps...?
 
  | cwkoss wrote:
  | I find it surprising and a bit disappointing that so many HN
  | readers find the manic meandering of yishans thread persuasive
 
  | jefftk wrote:
  | He goes back to the main topic after a few tweets on his
  | current climate work. It's actually a super long thread:
  | https://threadreaderapp.com/thread/1586955288061452289.html
  | 
  | (But I agree this is weird)
 
  | permo-w wrote:
  | he does come back to the point after his little side-piece
  | about trees, but after a while I didn't feel he was actually
  | providing any valuable information, so I stopped reading
 
  | dahfizz wrote:
  | Its a Twitter thread. Emotional, incoherent ramblings is the
  | standard.
 
  | snowwrestler wrote:
  | Yes, he is leveraging his audience. This like going to a
  | conference with a big-name keynote, but the lunch sponsor gets
  | up and speaks for 5 minutes first.
  | 
  | We're on the thread to read about content moderation. But since
  | we're there, he's going to inject a few promos about what he is
  | working on now. Just like other ads, I skimmed past them until
  | he got back on track with the main topic.
 
| teddyh wrote:
| So what I'm hearing is that ads are moderated spam. Yeah, I can
| see that.
 
| matai_kolila wrote:
| Yeah well, Yishan failed miserably at topic moderation on Reddit,
| and generally speaking Reddit has notoriously awful moderation
| policies that end up allowing users to run their own little
| fiefdoms just because they name-squatted earliest on a given
| topic. Additionally, Reddit (also notoriously) allowed some
| horrendously toxic behavior to continue on its site (jailbait,
| fatpeoplehate, the_donald, conservative currently) for literal
| years before taking action, so even when it comes to basic admin
| activity I doubt he's the guy we should all be listening to.
| 
| I think the fact that this is absurdly long and wanders at least
| twice into environmental stuff (which _is_ super interesting btw,
| definitely read those tangents) kind of illustrates just how not-
| the-best Yishan is as a source of wisdom on this topic.
| 
|  _Very_ steeped in typical SV  "this problem is super hard so
| you're not allowed to judge failure or try anything simple" talk.
| Also it's basically an ad for Block Party by the end (if you make
| it that far), so... yeah.
 
  | ranger207 wrote:
  | Yeah, it's interesting how much reddit's content moderation at
  | a site-wide level is basically the opposite of what he said in
  | this thread. Yeah, good content moderation should be about
  | policing behavior... so why weren't notorious brigading subs
  | banned?
 
  | pixl97 wrote:
  | Do you have any arguments addressing what he actually said, or
  | is this just a reverse argument to authority?
 
    | matai_kolila wrote:
    | Mostly just a reverse argument to authority, which isn't the
    | fallacy an argument to authority is, AFAIK.
 
    | [deleted]
 
| fazfq wrote:
| When people ask you why you hate twitter threads, show them this
| hodgepodge of short sentences with sandwiched CO2 removal
| advertisements.
 
  | kodon wrote:
  | Too bad he didn't post this on Reddit
 
  | luuuzeta wrote:
  | >this hodgepodge of short sentences with sandwiched CO2 removal
  | advertisements
  | 
  | I had to stop after the tree myths. Was it related to content
  | moderation at all?
 
    | halfmatthalfcat wrote:
    | No it wasn't, pure shilling
 
| RockyMcNuts wrote:
| see also -
| 
| Hey Elon: Let Me Help You Speed Run The Content Moderation
| Learning Curve
| 
| https://www.techdirt.com/2022/11/02/hey-elon-let-me-help-you...
 
| hourago wrote:
| > Our current climate of political polarization makes it easy to
| think it`s about the content of the speech, or hate speech, or
| misinformation, or censorship, or etc etc.
| 
| Are we sure that it is not the other way around? Didn't social
| platforms created or increased polarization?
| 
| I always see this comments from social platforms that take as
| fact that society is polarized and they work hard to fix it, when
| I believe that it is the other way around. Social media has
| created the opportunity to increase polarization and they are not
| able to stop it for technical, social or economic reasons.
 
  | throw0101a wrote:
  | > _Are we sure that it is not the other way around? Didn 't
  | social platforms created or increased polarization?_
  | 
  | The process of polarization (in the US) started decades ago:
  | 
  | * https://en.wikipedia.org/wiki/Why_We%27re_Polarized
  | 
  | In fact it seems that people were always polarized, it's just
  | that the political parties (R & D in the US) didn't really
  | bother sorting themselves on topics until the 1960s: even in
  | the 1970s and early 1980s it was somewhat common to vote for
  | (e.g.) an R president but a D representative (or vice versa).
  | Straight-through one-party voting didn't really become the
  | majority until the late-1980s and 1990s.
  | 
  | There's a chapter or two in the above book describing
  | psychology studies showing that humans form tribes
  | 'spontaneously' for the most arbitrary of reasons. "Us versus
  | them" seems to be baked into the structure of humans.
 
    | somenameforme wrote:
    | It's quite interesting that the USSR collapsed in 1991, which
    | removed the biggest external "us vs them" actor.
    | 
    | But on the other hand there are also countless other factors
    | that are going to affect society at scale: rise internet,
    | rise of pharmaceutical psychotropics, surge in obesity, surge
    | in autism, declines in testosterone, apparent reversal of
    | Flynn effect, and more.
    | 
    | With so many things happening it all feels like a Rorschach
    | test when trying to piece together anything like a meaningful
    | hypothesis.
 
    | raxxorraxor wrote:
    | I think political parties only later began astroturfing on
    | social media and split users in camps. Formerly content on
    | reddit in default subreddits often had low quality, but you
    | still got some nice topics here and there. Now it is a
    | propaganda hellhole that is completely in the hands of pretty
    | polarized users.
    | 
    | > "Us versus them" seems to be baked into the structure of
    | humans.
    | 
    | Not quite, but one of the most effective temptations one can
    | offer is giving people a moral excuse to hate others. Best
    | when see as those as responsible for all evil in the world.
    | It feels good to judge, it distracts from your own faults,
    | flaws, insecurities, fears and problems. This is pretty
    | blatant and has become far, far worse than the formerly
    | perhaps populist content on reddit. We especially see this on
    | political topics, but also the pandemic as an example.
 
      | throw0101a wrote:
      | > _I think political parties only later began astroturfing
      | on social media and split users in camps._
      | 
      | The splitting into camps (in the US) based on particular
      | topics started much earlier than social media:
      | 
      | * https://en.wikipedia.org/wiki/Southern_strategy
      | 
      | > _Not quite, but one of the most effective temptations one
      | can offer is giving people a moral excuse to hate others._
      | 
      | The psychology studies referenced in the book show us-
      | versus-them / in/out-group mentality without getting in
      | moral questions or political topics.
 
    | r721 wrote:
    | Scott Alexander's review is worth reading to get a summary of
    | this book: https://astralcodexten.substack.com/p/book-review-
    | why-were-p...
 
  | cafard wrote:
  | I think that you should look into the history of talk radio, or
  | maybe just radio in general. Then maybe a history of American
  | journalism, from Robert McCormick's Chicago Tribune back to the
  | the party newspapers set up in the first years of the republic.
 
    | Spivak wrote:
    | Yepp, same message different medium. Having someone in your
    | family who "listens to talk radio" was the "they went down
    | the far right YouTube rabbit hole" of old.
    | 
    | I mean the big names in talk radio are still clucking if you
    | want to listen to them today.
 
  | nradov wrote:
  | Real society is not that polarized. If you talk to real people
  | they mostly have moderate political opinions. But they don't
  | tweet about politics.
  | 
  | Twitter is not a real place.
 
    | toqy wrote:
    | I used to think this until several instances of various
    | neighbors getting drunk enough to shed the veil of souther
    | hospitality and reveal how racist they are.
    | 
    | Plenty of people have radical thoughts and opinions, but are
    | smart enough to keep it to themselves IRL
 
    | Spivak wrote:
    | But unfortunately real people are influenced and impacted by
    | the fiction.
    | 
    | If far right political bullshit would stay online and out of
    | my state's general assembly that would be such a positive
    | change.
 
    | count wrote:
    | Society is a closed system, twitter is not outside of
    | society.
    | 
    | The people on twitter are real people (well, mostly,
    | probably), and have real political opinions.
    | 
    | If you talk to people, by and large they'll profess moderate
    | opinions, because _in person discussions still trigger
    | politeness and non-confrontational emotions_ in most people,
    | so the default  'safe' thing to say is the moderate choice,
    | no matter what their true opinion happens to be.
    | 
    | The internet allows people to take the proverbial mask off.
 
      | SXX wrote:
      | I would disagree about proverbial masks. Majority of people
      | in the world including US are simply too preoccupied with
      | their everyday routine, problems and work to end up with
      | extreme political views.
      | 
      | What Internet does have is ease of changing masks and
      | joining diverse groups. Trying something unusual without
      | reprecussions appeal to a lot of people who usually simply
      | dont have time to join such groups offline.
      | 
      | The real problem is that unfortunately propoganda has
      | evolved too with all new research about human phychology,
      | behaviors and fallacies. Abusing weaknesses of monkey brain
      | on scale is relatively easy and profitable.
 
      | nradov wrote:
      | Nah. Even those few accounts on Twitter that are actually
      | run by real people (not bots) are mostly trolling to some
      | extent. It's all a big joke.
 
        | count wrote:
        | I thought that as well, until about Nov 2016...
 
        | bombcar wrote:
        | That was the biggest joke of all!
        | 
        | So far ...
 
  | r721 wrote:
  | Yeah, there were some good articles about this:
  | 
  | >The Making of a YouTube Radical
  | 
  | >I visited Mr. Cain in West Virginia after seeing his YouTube
  | video denouncing the far right. We spent hours discussing his
  | radicalization. To back up his recollections, he downloaded and
  | sent me his entire YouTube history, a log of more than 12,000
  | videos and more than 2,500 search queries dating to 2015.
  | 
  | https://www.nytimes.com/interactive/2019/06/08/technology/yo...
  | (2019)
 
| belorn wrote:
| > Moderating spam is very interesting: it is almost universally
| regarded as okay to ban (i.e. CENSORSHIP) but spam is in no way
| illegal.
| 
| Interesting, in my country spam is very much illegal and I would
| hazard a guess that it is also illegal in the US, similar to how
| littering, putting up posters on peoples buildings/cars/walls,
| graffiti (a form of spam), and so on is also illegal. If I
| received the amount of spam I get in email as phone calls I would
| go as far as calling it harassment, and of course robot phone
| calls are also illegal. Unsolicited email spam is also again the
| law.
| 
| And if spam is against the service agreement on twitter then that
| could be a computer crime. If the advertisement is fraudulent (as
| is most spam), it is fraud. Countries also have laws about
| advertisement, which most spam are unlikely to honor.
| 
| So I would make the claim that there is plenty of principled
| reasons for banning spam, all backed up by laws of the countries
| that the users and the operators live in.
 
  | snowwrestler wrote:
  | Nudity and porn are other examples of legal speech that have
  | broad acceptance among the public (at least the U.S. public) to
  | moderate or ban on social media platforms.
  | 
  | Yishan's point is, most people's opinions on how well a
  | platform delivers free speech vs censorship will index more to
  | the content of the speech, rather than the pattern of behavior
  | around it.
 
  | thesuitonym wrote:
  | Unsolicited phone calls are somewhat illegal, but it's
  | dependent on circumstances. It's the same with email spam and
  | mail spam. One person's spam is another person's cold call.
  | Where do you draw the line? Is mailing a flier with coupons
  | spam? Technically yes, but some people find value in it.
  | 
  | In the US, spam is protected speech, but as always, no company
  | is required to give anybody a platform.
 
    | myself248 wrote:
    | > In the US, spam is protected speech
    | 
    | [citation needed]
    | 
    | Doesn't the CAN-SPAM act explicitly declare otherwise?
 
      | toqy wrote:
      | I was under the impression that CAN-SPAM applies to email,
      | not user generated content on the internet at large
 
        | belorn wrote:
        | It is both yes, and no. CAN-SPAM do only apply to
        | electronic mail messages, usually shorten down to email.
        | However...
        | 
        | In late March, a federal court in California held that
        | Facebook postings fit within the definition of
        | "commercial electronic mail message" under the
        | Controlling the Assault of Non-Solicited Pornography and
        | Marketing Act ("CAN-SPAM Act;" 15 U.S.C. SS 7701, et
        | seq.). Facebook, Inc. v. MAXBOUNTY, Inc., Case No.
        | CV-10-4712-JF (N.D. Cal. March 28, 2011).
        | 
        | There is also two other court cases: MySpace v. The
        | Globe.com and MySpace v. Wallace.
        | 
        | In the later, the court concluded that "[t]o interpret
        | the Act in the limited manner as advocated by [d]efendant
        | would conflict with the express language of the Act and
        | would undercut the purpose for which it was passed." Id.
        | This Court agrees that the Act should be interpreted
        | expansively and in accordance with its broad legislative
        | purpose.
        | 
        | The court defined "electronic mail address" as meaning
        | nothing more specific than "a destination . . . to which
        | an electronic mail message can be sent, and the
        | references to local part and domain part and all other
        | descriptors set off in the statute by commas represent
        | only one possible way in which a destination can be
        | expressed.
        | 
        | Basically, in order to follow the spirit of the law the
        | definition of "email" expanded, with traditional email
        | like user@example.invalid being just one example of many
        | forms of "email".
 
    | null0ranje wrote:
    | > In the US, spam is protected speech, but as always, no
    | company is required to give anybody a platform.
    | 
    | Commercial speech in the US is not protected speech and may
    | be subject to a host of government regulation [0]. The
    | government has broad powers to regulate the time, place, and
    | content of commercial speech in ways that it does not for
    | ordinary speech.
    | 
    | [0] See https://www.law.cornell.edu/wex/commercial_speech
 
      | gorbachev wrote:
      | Not all spam is commercial.
      | 
      | In fact, US legislators specifically made political spam
      | legal in the CAN-SPAM bill.
 
    | belorn wrote:
    | It is dependent on circumstances, and the people who would
    | draw the line in the end would be the government followed by
    | the court.
    | 
    | Not all speech is protected speech. Graffiti is speech, and
    | the words being spoken could be argued as protected, but the
    | act of spraying other people properties with it is not
    | protected. Free speech rights does not overwrite other
    | rights. As a defense in a court I would not bet my money on
    | free speech in order to get away with crimes that happens to
    | involves speech.
    | 
    | Historically the US court has defined speech into multiple
    | different categories. One of those are called fraudulent
    | speech which is not protected by free speech rights. An other
    | category is illustrated with the anti-spam law in Washington
    | State, which was found to not be in violation of First
    | Amendment rights because it prevent misleading emails.
    | Washington's statue regulate deceptive _commercial speech_
    | and thus passed the constitutional test. An other court
    | ruling, this one in Maryland, confirmed that commercial
    | speech was less protected than other forms of speech and that
    | commercial speech had no protection when it was demonstrably
    | false.
    | 
    | In theory a spammer could make non-commercial, non-
    | misleading, non-fraudulent speech, and a site like twitter
    | would then actually have to think about questions like first-
    | amendment. I can't say I have ever received or seen spam like
    | that.
 
      | buzer wrote:
      | > In theory a spammer could make non-commercial, non-
      | misleading, non-fraudulent speech, and a site like twitter
      | would then actually have to think about questions like
      | first-amendment. I can't say I have ever received or seen
      | spam like that.
      | 
      | While I don't think I have seen it on Twitter (then again I
      | only read it when it's linked) I have seen plenty of it in
      | some older forums & IRC. Generally it's just nonsense like
      | "jqrfefafasok" or ":DDDDDD" being posted lots of times in
      | quick succession, often to either flood out other things,
      | to draw attention to poster or to show annoyance about
      | something (like being banned previously).
 
        | belorn wrote:
        | You got a point. Demonstration as a form of free speech
        | is an interesting dilemma. Review spam/bombing for
        | example can be non-commercial, non-misleading, non-
        | fraudulent, while still being a bit of a grey-zone.
        | Removing them is also fairly controversial. Outside the
        | web we have a similar problem when demonstrations and
        | strikes are causing disruption in society. Obviously
        | demonstration and strikes should be legal and are
        | protected by free speech, but at the same time there are
        | exceptions when they are not.
        | 
        | I am unsure if one would construct a objective fair model
        | for how to moderate such activity.
 
      | thesuitonym wrote:
      | >a site like twitter would then actually have to think
      | about questions like first-amendment.
      | 
      | I wish people understood that the first amendment does not
      | have anything to do with social media sites allowing people
      | to say anything. Twitter is not a public square, no matter
      | how much you want it to be.
 
| asddubs wrote:
| i love that, this fucking twitter thread has a commercial break
| in the middle of it.
| 
| edit: it has multiple commercial breaks!
 
| mmastrac wrote:
| Unrolled thread: https://mem.ai/p/D0AfFRGYoKkyW5aQQ1En
 
  | top_sigrid wrote:
  | Wait what?
  | 
  | If you want a decend unroll, one example would be
  | threadreaderapp:
  | https://threadreaderapp.com/thread/1586955288061452289.html
 
| datan3rd wrote:
| I think email might be a good system to model this on. In
| addition to an inbox, almost all providers provide a Spam folder,
| and others like Gmail separate items into 'Promotions' and
| 'Social' folders/labels. I imagine almost nobody objects to this.
| 
| Why can't social media follow a similar methodology? There is no
| requirement that FB/Twitter/Insta/etc feeds be a single "unit".
| The primary experience would be a main feed (uncontroversial),
| but additional feeds/labels would be available to view platform-
| labeled content. A "Spam Feed" and a "Controversial Feed" and a
| "This Might Be Misinformation Feed".
| 
| Rather than censoring content, it segregates it. Users are free
| to seek/view that content, but must implicitly acknowledge the
| platform's opinion by clicking into that content. Just like you
| know you are looking at "something else" when you go to your
| email Spam folder, you would be aware that you are venturing off
| the beaten path when going to the "Potential State-Sponsored
| Propaganda Feed". There must be some implicit trust in a singular
| feed which is why current removal/censorship schemas cause such
| "passionate" responses.
 
| wcerfgba wrote:
| I like Yishan's reframing of content moderation as a 'signal-to-
| noise ratio problem' instead of a 'content problem', but there is
| another reframing which follows from that: moderation is also an
| _outsourcing problem_ , in that moderation is about users
| outsourcing the filtering of content to moderators (be they all
| other users through voting mechanisms, a subset of privileged
| users through mod powers, or an algorithm).
| 
| Yishan doesn't define what the 'signal' is, or what 'spam' is,
| and there will probably be an element of subjectivity to these
| which varies between each platform and each user on each
| platform. Thus successful moderation happens when moderators know
| what users want, i.e. what the users consider to be 'good
| content' or 'signal'. This reveals a couple of things about why
| moderation is so hard.
| 
| First, this means that moderation actually _is_ a content
| problem. For example, posts about political news are regularly
| removed from Hacker News because they are off-topic for the
| community, i.e. we don 't consider that content to be the
| 'signal' that we go to HN for.
| 
| Second, moderation can only be successful when there is a shared
| understanding between users and moderators about what 'signal'
| is. It's when this agreement breaks down that moderation becomes
| difficult or fails.
| 
| Others have posted about the need to provide users with the tools
| to do their own moderation in a decentralised way. Since the
| 'traditional'/centralised approach creates a fragile power
| dynamic which requires this shared understanding of signal, I
| completely understand and agree with this: as users we should
| have the power to filter out content we don't like to see.
| 
| However, we have to distinguish between general and topical
| spaces, and to determine which communities live in a given space
| and what binds different individuals into collectives. Is there a
| need for a collective understanding of what's on-topic? HN is not
| Twitter, it's designed as a space for particular types of people
| to share particular types of content. Replacing 'traditional' or
| centralised moderation with fully decentralised moderation risks
| disrupting the topicality of the space and the communities which
| inhabit it.
| 
| I think what we want instead is a 'democratised' moderation, some
| way of moderating that removes a reliance on a 'chosen few', is
| more deliberate about what kinds of moderation need to be
| 'outsourced', and which allows users to participate in a shared
| construction of what they mean by 'signal' or 'on-topic' for
| their community. Perhaps the humble upvote is a good example and
| starting point for this?
| 
| Finally in the interest of technocratic solutions, particularly
| around spam (which I would define as repetitive content), has
| anyone thought about rate limits? Like, yeah if each person can
| only post 5 comments/tweets/whatever a day then you put a cap on
| how much total content can be created, and incentivise users to
| produce more meaningful content. But I guess that wouldn't allow
| for all the _sick massive engagement_ that these attention
| economy walled garden platforms need for selling ads...
 
| [deleted]
 
| bravura wrote:
| Yishan's points are great, but there is a more general and
| fundamental question to discuss...
| 
| Moderation is the act removing content. i.e. of assigning a score
| of 1 or 0 to content.
| 
| If we generalize, we can assign a score from 1 to 0 to all
| content. Perhaps this score is personalized. Now we have a user's
| priority feed.
| 
| How should Twitter score content using personalization? Filter
| bubble? Expose people to a diversity of opinions? etc. Moderation
| is just a special case of this.
 
  | panarky wrote:
  | One size does not fit all.
  | 
  | Some people want to escape the filter bubble, to expose their
  | ideas to criticism, to strengthen their thinking and arguments
  | through conflict.
  | 
  | Other people want a community of like-minded people to share
  | and improve ideas and actions collectively, without trolls
  | burning everything down all the time.
  | 
  | Some people want each of those types of community depending on
  | the topic and depending on their mood at the time.
  | 
  | A better platform would let each community decide, and make it
  | easy to fork off new communities with different rules when a
  | subgroup or individual decides the existing rules aren't
  | working for them.
 
| rongopo wrote:
| Imagine there would be many shades of up and down voting in HN,
| according to your earned karma points, and to your interactions
| outside of your regular opinion echo Chambers.
 
| lawrenceyan wrote:
| You can tell this guy is a genius at marketing.
| 
| Smart to comment on his current pursuits in environmental
| terraforming knowing he's going to get eyeballs on any thread he
| writes.
 
  | yamazakiwi wrote:
  | I commented on another comment discussing this and they thought
  | the opposite. I also thought it was relatively a good idea,
  | albeit distracting.
 
| DelightOne wrote:
| Can there be a moderation bot that detects flamewars and steps
| in? It could enforce civility by limiting discussion to only go
| through the bot and by employing protocols like "each side
| summarize issues", "is this really important here", or "do you
| enjoy this".
| 
| Engaging with the bot is supposed to be a rational barrier, a
| tool to put unproductive discussions back on track.
 
| e40 wrote:
| Easier to read this:
| 
| https://threadreaderapp.com/thread/1586955288061452289.html
 
| wackget wrote:
| Anyone got a TL;DR? I don't feel like trudging through 100
| sentences of verbal diarrhea.
 
| goatcode wrote:
| > you`ll end up with a council of third-rate minds and
| politically-motivated hacks, and the situation will be worse than
| how you started.
| 
| Wow, surprising honesty from someone affiliated with Reddit. I'm
| sad that I wasn't on the site during the time of the old guard.
 
  | commandlinefan wrote:
  | > I'm sad that I wasn't on the site during the time of the old
  | guard.
  | 
  | It really was great - I probably wouldn't care how horrible
  | it's become if not for the fact that I remember how it used to
  | be.
 
| anigbrowl wrote:
| Reposting this paper yet again, to rub in the point that social
| media platforms play host to _communities_ and communities are
| often very good at detecting interlopers and saboteurs and
| pushing them back out. And it turns out the most effective
| approach is to let people give bad actors a hard time. Moderation
| policies that require everyone to adhere to high standards of
| politeness in all circumstances are trying to reproduce the
| dynamics of kindergartens, and are not effective because the
| moderators are easily gamed.
| 
| https://arxiv.org/pdf/1803.03697.pdf
| 
| Also, if you're running or working for a platform and dealing
| with insurgencies, you will lose if you try to build any kind of
| policy around content analysis. Automated context analysis is
| generally crap because of semantic overloading (irony, satire,
| contextual humor), and manual context analysis is labor-intensive
| and immiserating, to the point that larger platforms like
| Facebook are legitimately accused of abusing their moderation
| staff by paying them peanuts to wade through toxic sludge and
| then dumping them as soon as they complain or ask for any kind of
| support from HR.
| 
| To get anywhere you need to look at patterns of behavior and to
| scale you need to do feature/motif detection on dynamic systems
| rather than static relationships like friend/follower selections.
| However, this kind of approach is fundamentally at odds with many
| platforms' goal of maximizing engagement as means to the end of
| selling ad space.
 
| aerovistae wrote:
| These random detours into climate-related topics are insanely
| disruptive of an otherwise interesting essay. I absolutely hate
| this pattern. I see what he's trying to do - you don't want to
| read about climate change but you want to read this other thing
| so I'm going to mix them together so you can't avoid the one if
| you want the other - but it's an awful dark pattern and makes for
| a frustrating and confusing reading experience. I kept thinking
| he was making an analogy before realizing he was just changing
| topics at random again. It certainly isn't making me more
| interested in his trees project. If anything I'm less interested
| now.
 
  | IncRnd wrote:
  | Since the argument was so well-structured, the interim detour
  | to climate related topics was odd. The very argument was that
  | spam can be detected by posting behaviors, yet the author
  | engaged in those for his favored topic.
 
| incomingpain wrote:
| This CEO did the same thread 6 months ago and was blasted off the
| internet. You can see his thread here:
| https://twitter.com/yishan/status/1514938507407421440
| 
| edit/ Guess it is working now?
| 
| The most important post in his older thread:
| https://twitter.com/yishan/status/1514939100444311560
| 
| He never ever justifies this point. The world absolutely has not
| changed in the context of censorship. Censorship apologetics
| notwithstanding.
| 
| The realization is the world changed is a reveal. He as CEO
| learnt about where the censorship is coming from.
 
  | Spivak wrote:
  | What's wrong with this thread? It seems really level headed and
  | exactly accurate to the people I know IRL who are insane-but-
  | left and insane-but-right who won't shut up about censorship
  | while if you look at their posts it's just "unhinged person
  | picks fights with and yells at strangers."
  | 
  | HN in general is convinced that social media is censoring right
  | ideas because it skews counterculture and "grey tribe" and
  | there have been a lot of high profile groups who claim right
  | views while doing the most vile depraved shit like actively
  | trying to harass people into suicide and celebrating it or
  | directing massive internet mobs at largely defenseless not
  | public figures for clout.
 
  | mikkergp wrote:
  | > The world absolutely has not changed in the context of
  | censorship.
  | 
  | Citation needed
 
    | incomingpain wrote:
    | >Citation needed
    | 
    | As I said in my post, he never justifies this point. To then
    | turn it upon me to prove a negative?
    | 
    | Devils advocating against myself: I do believe the parler
    | deplatforming is the proof for what he says. The world has
    | indeed changed, but anyone who knows the details sure isn't
    | saying why. Why? Because revealing how the world has changed,
    | in the usa, would have some pretty serious consequences.
    | 
    | I don't know. I wish I could have a closed door, off record,
    | tell me everything, conversation with yishan to have him tell
    | me why he believes the world changed, in the context of
    | social media censorship.
    | 
    | In terms of public verified knowledge, nothing at all has
    | changed in the context of censorship. I stand by the point.
    | Elon obviously stands by this as well. Though elon's sudden
    | multiweek delays on unbanning... im expecting he suddenly
    | knows as well.
    | 
    | >You're posting too fast. Please slow down. Thanks.
    | 
    | Guess I'm not allowed to reply again today. No discussion
    | allowed on HN.
    | 
    | I do find it funny they say 'you're posting too fast' but I
    | haven't been able to post on HN or reply to you for an hour.
    | How "fast" am I really going. I expect it will be a couple
    | more hours before I am allowed to post again. How dare I
    | discuss a forbidden subject.
 
  | lrm242 wrote:
  | Huh? What do you mean unavailable? I see it just fine.
 
    | r721 wrote:
    | I can confirm - I saw "this tweet is unavailable" message or
    | something similarly worded on first click too. Reloading
    | fixed that.
 
| kmeisthax wrote:
| This is a very good way to pitch your afforestation startup
| accelerator in the guise of a talk on platform moderation. /s
| 
| I'm pretty sure I've got some bones to pick with yishan from his
| tenure on Reddit, but everything he's said here is pretty
| understandable.
| 
| Actually, I would like to develop his point about "censoring
| spam" a bit further. It's often said that the Internet "detects
| censorship as damage and routes around it". This is propaganda,
| of course; a fully censorship-resistant Internet is entirely
| unusable. In fact, the easiest way to censor someone online is
| through harassment, or DDoS attacks - i.e. have a bunch of people
| shout at you until you shut up. Second easiest is through doxing
| - i.e. make the user feel unsafe until they jump off platform and
| stop speaking. Neither of these require content removal
| capability, but they still achieve the goal of censorship.
| 
| The point about old media demonizing moderation is something I
| didn't expect, but it makes sense. This _is_ the same old media
| that gave us cable news, after all. Their goal is not to inform,
| but to allure. In fact, I kinda wish we had a platform that
| explicitly refused to give them the time of day, but I 'm pretty
| sure it's illegal to do that now[0], and even back a decade ago
| it would be financial suicide to make a platform only catering to
| individual creators.
| 
| [0] For various reasons:
| 
| - The EU Copyright Directive imposes an upload filtering
| requirement on video platforms that needs cooperation with old
| media companies in order to implement. The US is also threatening
| similar requirements.
| 
| - Canada Bill C-11 makes Canadian content (CanCon) must-carry for
| all Internet platforms, including ones that take user-generated
| content. In practice, it is easier for old media to qualify as
| CanCon than actual Canadian individuals.
 
  | nullc wrote:
  | I've often pointed out that the concept of censorship as being
  | only or primarily through removal of speech is an antiquated
  | concept from a time before pervasive communications networks
  | had almost effortlessly connected most of the world.
  | 
  | Censorship in the traditional sense is close to impossible
  | online today.
  | 
  | Today censorship is often and most effectively about
  | suppressing your ability to be heard, often by flooding out the
  | good communications with nonsense, spam, abuse, or discrediting
  | it by association (e.g. fill the forums of a political
  | opponents with apparent racists). This turns the neigh
  | uncensorability of modern communications methods on its head
  | and makes it into a censorship tool.
  | 
  | And, ironically, anyone trying to use moderation to curb this
  | sort of censorious abuse is easily accused of 'censorship'
  | themselves.
  | 
  | I remain convinced that the best tool we have is topicality:
  | When a venue has a defined topic you can moderate just to stay
  | onto the topic without a lot of debatable value judgements (or
  | bruised egos-- no reason to feel too bad about a post being
  | moved or removed for being offtopic). Unfortunately, the
  | structure of twitter pretty much abandons this critical tool.
  | 
  | (and with reddit increasingly usurping moderation from
  | subreddit moderators, it's been diminished there)
  | 
  | Topicality doesn't solve all moderation issues, but once an
  | issue has become too acrimonious it will inherently go off-
  | topic: e.g. if your topic is some video game well participants
  | calling each other nasty names is clearly off-topic. Topicality
  | also reduces the incidence of trouble coming in from divisive
  | issues that some participants just aren't interested in
  | discussing-- If I'm on a forum for a video game I probably
  | don't really want to debate abortion with people.
  | 
  | In this thread we see good use of topicality at the top with
  | Dang explicitly marking complaints about long form twitter
  | offtopic.
  | 
  | When it comes to spam scaling considerations mean that you need
  | to be able to deal with much of it without necessarily
  | understanding the content. I don't think this should be
  | confused with content blindness being desirable in and of
  | itself. Abusive/unwelcoming interactions can occur both in the
  | form (e.g. someone stalking some around from thread to thread
  | or repeating an argument endlessly) and and in the content
  | (continually re-litigating divisive/flame-bate issues that no
  | one else wants to talk about, vile threatening messages, etc.)
  | 
  | Related to topicality is that some users just don't want to
  | interact with each other. We don't have to make a value
  | judgement about one vs the other if we can provide space so
  | that they don't need to interact. Twitter's structure isn't
  | great for this either, but more the nature of near-monopoly
  | mega platforms isn't great for it. Worse, twitter actively make
  | it hard-- e.g. if you've not followed someone who is network-
  | connected to other people you follow twitter continually
  | recommends their tweets (as a friend said: "No twitter, there
  | is a reason I'm not following them") and because blocking is
  | visible using it often creates drama.
  | 
  | There are some subjects on HN where I might otherwise comment
  | but I don't because I'd prefer to avoid interacting with a Top
  | Poster who will inevitably be active in those subjects.
  | Fortunately, there are plenty of other places where I can
  | discuss those things where that poster isn't active.
  | 
  | Even a relatively 'small' forum can easily have as many users
  | as many US states populations at the founding of the country. I
  | don't think that we really need to have mega platforms with
  | literally everyone on them and I see a fair amount of harm from
  | it (including the effects of monoculture moderation gone
  | wrong).
  | 
  | In general, I think the less topic constrained you can make a
  | venue the smaller it needs to be-- a completely topic-less
  | social venue probably should have no more than a few dozen
  | people. Twitter is both mega-topicless and ultra-massive-- an
  | explosive mixture which will inevitably disappoint.
  | 
  | Another tool I think many people have missed the value of is
  | procedural norms including decorum. I don't believe that using
  | polite language actually makes someone polite (in fact, the
  | nastiest and most threatening remarks I've ever received were
  | made with perfectly polite language)-- but some people are just
  | unable to regulate their own behavior. When there is an easily
  | followed set of standards for conduct you gain a bright line
  | criteria that makes it easier to eject people who are too
  | unable to control themselves. Unfortunately, I think the value
  | of a otherwise pointless procedural conformity test is often
  | lost on people today, though they appear common in historical
  | institutions. (Maybe a sign of the ages of the creators of
  | these things: As a younger person I certainly grated against
  | 'pointless' conformity requirements, as an older person I see
  | more ways that their value can pay for their costs: I'd rather
  | not waste my time on someone who can't even manage to go
  | through the motions to meet the venue's standards)
  | 
  | Early on in Wikipedia I think we got a lot of mileage out of
  | this: the nature of the site essentially hands every user a
  | loaded gun (ability to edit almost everything, including
  | elements on the site UI) and then tells them not to do use it
  | abusively rather than trying to technically prevent them from
  | using it abusively. Some people can't resist and are quickly
  | kicked out without too much drama. Had those same people been
  | technically prevented they would have hung around longer and
  | created trouble that was harder to kick them out over (and I
  | think as the site added more restrictions on new/casual users
  | the number of issues from poorly behaved users increased).
 
  | mountainriver wrote:
  | I love that he's for flame wars, go figure that's all Reddit is
 
| saurik wrote:
| > there will be NO relation between the topic of the content and
| whether you moderate it, because it`s the specific posting
| behavior that`s a problem
| 
| I get why Yishan wants to believe this, but I also feel like the
| entire premise of this argument is then in some way against a
| straw man version of the problem people are trying to point to
| when they claim moderation is content-aware.
| 
| The issue, truly, isn't about what the platform moderates so much
| as the bias between when it bothers to moderate and when it
| doesn't.
| 
| If you have a platform that bothers to filter messages that "hate
| on" famous people but doesn't even notice messages that "hate on"
| normal people--even if the reason is just that almost no one sees
| the latter messages and so they don't have much impact and your
| filters don't catch it--you have a (brutal) class bias.
| 
| If you have a platform that bothers to filter people who are
| "repetitively" anti large classic tech companies for the evil
| things they do trying to amass money and yet doesn't filter
| people who are "repetitively" anti crypto companies for the evil
| things _they_ do trying to amass money--even if it feels to you
| as the moderator that the person seems to have a point ;P--that
| is another bias.
| 
| The problem you see in moderation--and I've spent a LONG time
| both myself being a moderator and working with people who have
| spent their lives being moderators, both for forums and for live
| chat--is that moderation and verification of everything not only
| feels awkward but simply _doesn 't scale_, and so you try to
| build mechanisms to moderate _enough_ that the forum seems to
| have a high _enough_ signal-to-noise ratio that people are happy
| and generally stay.
| 
| But the way you get that scale is by automating and triaging: you
| build mechanisms involving keyword filters and AI that attempt to
| find and flag low signal comments, and you rely on reports from
| users to direct later attention. The problem, though, is that
| these mechanisms inherently have biases, and those biases
| absolutely end up being inclusive of biases that are related to
| the content.
| 
| Yishan seems to be arguing that perfectly-unbiased moderation
| might seem biased to some people, but he isn't bothering to look
| at where or why moderation often isn't perfect to ensure that
| moderation actually works the way he claims it should, and I'm
| telling you: it never does, because moderation isn't omnipresent
| and cannot be equally applied to all relevant circumstances. He
| pays lip service to it in one place (throwing Facebook under the
| bus near the end of the thread), and yet fails to then realize
| _this is the argument_.
| 
| At the end of the day, real world moderation is certainly biased.
| _And maybe that 's OK!_ But we shouldn't pretend it isn't biased
| (as Yishan does here) or even that that bias is always in the
| public interest (as many others do). That bias may, in fact, be
| an important part of moderating... and yet, it can also be
| extremely evil and difficult to discern from "I was busy" or "we
| all make mistakes" as it is often subconscious or with the best
| of intentions.
 
| karaterobot wrote:
| There were indeed some intelligent, thoughtful, novel insights
| about moderation in that thread. There were also... two
| commercial breaks to discuss his new venture? Eww. While
| discussing how spam is the least controversial type of noise you
| want to filter out? I appreciate the good content, I'm just not
| used to seeing product placement wedged in like that.
 
  | yamazakiwi wrote:
  | I thought it was simultaneously annoying and interesting so it
  | sort of cancelled itself out.
 
| zcombynator wrote:
| Spam is unwelcommed for a simple reason: there is no real person
| behind it.
 
  | kodt wrote:
  | Not always true. In fact often spam is simply self-promotion by
  | the person posting it.
 
    | bombcar wrote:
    | In fact that type of spam is more annoying than the BUY @#$@$
    | NOW generic bot-spam, as it is way more insidious.
 
  | hackerlight wrote:
  | If I was behind the exact same spam, would it be welcomed? Come
  | on.
 
| mcguire wrote:
| Correct me if I'm wrong, but this sounds very much like what dang
| does here.
 
| dang wrote:
| All: this is an interesting submission--it contains some of the
| most interesting writing about moderation that I've seen in a
| long time*. If you're going to comment, please make sure you've
| read and understand his argument and are engaging with it.
| 
| If you dislike long-form Twitter, here you go:
| https://threadreaderapp.com/thread/1586955288061452289.html - and
| please _don 't_ comment about that here. I know it can be
| annoying, but so is having the same offtopic complaints upvoted
| to the top of every such thread. This is why we added the site
| guideline: " _Please don 't complain about tangential annoyances
| --e.g. article or website formats_" (and yes, this comment is
| also doing this. Sorry.)
| 
| Similarly, please resist being baited by the sales interludes in
| the OP. They're also offtopic and, yes, annoying, but this is why
| we added the site guideline " _Please don 't pick the most
| provocative thing in an article to complain about--find something
| interesting to respond to instead._"
| 
| https://news.ycombinator.com/newsguidelines.html
| 
| * even more so than
| https://news.ycombinator.com/item?id=33446064, which was also
| above the median for this topic.
 
| rglover wrote:
| A fun idea that I'm certain no one has considered with any level
| of seriousness: don't moderate anything.
| 
| Build the features to allow readers to self-moderate and make it
| "expensive" to create or run bots (e.g., make it so API access is
| limited without an excessive fee, limit screen scrapers, etc).
| The "pay to play" idea will eliminate an insane amount of the
| junk, too. Any free network is inherently going to have problems
| of chaos. Make it so you can only follow X people with a free
| account, but upgrade to follow more. Limit tweets/replies/etc
| based on this. Not only will it work, but it will remove the need
| for all of the moderation and arguments around bias.
| 
| As for advertisers (why any moderation is necessary in the first
| place beyond totalitarian thought control): have different tiers
| of quality. If you want a higher quality audience, pay more. If
| you're more concerned about broad reach (even if that means
| getting junk users), pay less. Beyond that, advertisers/brands
| should set their expectations closer to reality: randomly
| appearing alongside some tasteless stuff on Twitter does not mean
| you're _vouching_ for those ideas.
 
  | munificent wrote:
  | _> Build the features to allow readers to self-moderate_
  | 
  | This is effectively impossible because of the bullshit
  | asymmetry principle[1]. It's easier to create content that
  | needs moderation than it is to moderate it. In general, there
  | is a fundamental asymmetry to life that it takes less effort to
  | destroy than it does to create, less work to harm than heal.
  | With a slightly sharpened piece of metal and about a newton of
  | force, you can end a life. No amount of effort can resurrect
  | it.
  | 
  | It simply doesn't scale to let bad actors cause all the harm
  | they want and rely on good actors to clean up their messes
  | after the fact. The harm must be prevented before it does
  | damage.
  | 
  |  _> make it  "expensive" to create or run bots (e.g., make it
  | so API access is limited without an excessive fee, limit screen
  | scrapers, etc)._
  | 
  | The simplest approach would be no API at all, but that won't
  | stop scammers and bad actors. It's effectively impossible to
  | prohibit screen scrapers.
  | 
  |  _> Make it so you can only follow X people with a free
  | account, but upgrade to follow more. Limit tweets /replies/etc
  | based on this. Not only will it work, but it will remove the
  | need for all of the moderation and arguments around bias._
  | 
  | This is, I think, the best idea. If having an identity and
  | sharing content costs actual money, you can at least make
  | spamming not be cost effective. But that still doesn't
  | eliminate human bad actors griefing others. Some are happy to
  | pay to cause mayhem.
  | 
  | There is no simple technical solution here. Fundamentally, the
  | value proposition of a community is the other good people you
  | get to connect to. But some people are harmful. They may not
  | always be harmful, or may be harmful only to some people. For a
  | community to thrive, you've got to encourage the good behaviors
  | and police the bad ones. That takes work and human judgement.
  | 
  | [1]: https://en.wikipedia.org/wiki/Brandolini%27s_law
 
    | rglover wrote:
    | > But some people are harmful. They may not always be
    | harmful, or may be harmful only to some people.
    | 
    | This is a fundamental reality of life that cannot be avoided.
    | There is no magical solution (technical or otherwise) to
    | prevent this. At best, you can put in some basic safeguards
    | (like what you/I have stated above) but ultimately people
    | need to learn to accept that you can't make everything 100%
    | safe.
    | 
    | Also, things like muting/blocking work but the ugly truth is
    | that people love the negative excitement of fighting online
    | (it's an outlet for life's pressure/disappointments).
    | Accepting _that_ reality would do a lot of people a lot of
    | good. A staggering amount of the negativity one encounters on
    | social media is self-inflicted by either provoking or
    | engaging with being provoked.
 
  | etchalon wrote:
  | 1. There are plenty of places online that "don't moderate
  | anything". In fact, nearly all of the social networks started
  | off that way.
  | 
  | The end result is ... well, 4Chan.
  | 
  | 2. "Self-moderation" doesn't work, because it's work. User's
  | don't want to have constantly police their feeds and block
  | people, topics, sites, etc. It's also work that never ends. Bad
  | actors jump from one identity to the next. There are no
  | "static" identifiers that are reliable enough for a user to
  | trust.
  | 
  | 3. Advertisers aren't going to just "accept" that their money
  | is supporting content they don't want to be associated with.
  | And they're also not interested in spending time white-listing
  | specific accounts they "know" are good.
 
    | rglover wrote:
    | > The end result is ... well, 4Chan.
    | 
    | And? Your opinion of whether that's bad is subjective, yet
    | the people there are happy with the result (presumably, as
    | they keep using/visiting it).
    | 
    | > Self-moderation" doesn't work, because it's work.
    | 
    | So in other words: "I'm too lazy to curate a non-threatening
    | experience for myself which is my responsibility because the
    | offense being taken is my own." Whether or not you're willing
    | to filter things out that upset you is a personal problem,
    | not a platform problem.
    | 
    | > Advertisers aren't going to just "accept" that their money
    | is supporting content they don't want to be associated with.
    | 
    | It's not. Twitter isn't creating the content nor are they
    | financing the content (e.g. like a Netflix type model). It's
    | user-generated which is completely random and subject to
    | chaos. If they can't handle that, they shouldn't advertise
    | there (hence why a pay-to-play option is best as it prevents
    | a revenue collapse for Twitter). E.g., if I I'm selling
    | crucifixes, I'm not going to advertise on slutmania.com
    | 
    | ---
    | 
    | Ultimately, people need to quit acting like everything they
    | come into contact with needs to be respectful of every
    | possible issue or disagreement they have with it. It's
    | irrational, entitled, and childish.
 
      | etchalon wrote:
      | 1. I didn't imply whether it was good or bad, just that the
      | product you're describing already exists.
      | 
      | 2. It's a platform problem. If you make users do work they
      | don't want to do in order to make the platform pleasant to
      | use, they won't do the work, the platform will not be
      | pleasant to use, and they'll use a different platform that
      | doesn't make them do that work.
      | 
      | 3. "If they can't handle it, they shouldn't advertise
      | there." Correct! They won't advertise there. That's the
      | point.
      | 
      | There are already unmoderated, "you do the work, not us",
      | "advertisers have to know what they're getting into"
      | platforms, and those platforms are niche, with small
      | audiences, filled with low-tier/scam ads and are generally
      | not profitable.
 
  | lambic wrote:
  | It's a problem of scale.
  | 
  | Usenet and IRC used to be self-moderated. The mods in each
  | group or channel would moderate their own userbase, ban people
  | who were causing problems, step in if things were getting too
  | heated. At a broader level net admins dealt with the spam
  | problem system wide, coordinating in groups in the news.admin
  | hierarchy or similar channels in IRC.
  | 
  | This worked fine for many years, but then the internet got big.
  | Those volunteer moderators and administrators could no longer
  | keep up with the flood of content. Usenet died (yes, it's still
  | around, but it's dead as any kind of discussion forum) and IRC
  | is a shell of its former self.
 
    | rglover wrote:
    | Right, which is solved by the pay to play limits. This would
    | essentially cut the problem off immediately and it would be
    | of benefit to everyone. If it actually cost people to "do bad
    | stuff" (post spam, vitriol, etc), they're far less-likely to
    | do it as the incentives drop off.
    | 
    | The dragon folks seem to be chasing is that Twitter should be
    | free but perfect (which is a have your cake and eat it too
    | problem). That will never happen and it only invites more
    | unnecessary strife between sociopolitical and socioeconomic
    | factions as they battle for narrative control.
 
| invalidusernam3 wrote:
| Just add a dislike button and put controversial tweets collapsed
| at the bottom. It works well for reddit. Let the community
| moderate themselves.
 
  | threeseed wrote:
  | Reddit tried to just let communities moderate themselves.
  | 
  | It resulted in rampant child pornography, doxxing, death
  | threats, gory violence etc. It epitomised the worst of
  | humanity.
  | 
  | Now the Reddit admins keep a watch on moderators and if their
  | subreddits do not meet site-wide standards they are replaced.
 
    | pessimizer wrote:
    | > It resulted in rampant child pornography, doxxing, death
    | threats, gory violence etc. It epitomised the worst of
    | humanity.
    | 
    | It resulted in reddit. That style of moderation is how reddit
    | became reddit; so it should also get credit for whatever you
    | think is good about reddit. The new (half-decade old) reddit
    | moderation regime was a new venture that was hoping to retain
    | users who were initially attracted by the old moderation
    | regime.
 
      | threeseed wrote:
      | This is revisionist history.
      | 
      | My Reddit account is 16 years old. I was there in the very
      | early days of the site well before the Digg invasion and
      | well before it gained widespread popularity.
      | 
      | It was never because it allowed anything. It was because it
      | was a much more accessible version of Slashdot. And it was
      | because Digg did their redesign and it ended up with a
      | critical mass of users. Then they started opening up the
      | subreddits and it exploded from there.
      | 
      | The fact that Reddit is growing without that content shows
      | that it wasn't that important to begin with.
 
      | pixl97 wrote:
      | You mean it resulted in the place that couldn't pay the
      | bills and goes around asking for VC money to keep the
      | servers on?
      | 
      | Unmoderated hell holes tend to have to survive on
      | questionable funding and rarely grow to any size.
 
        | SV_BubbleTime wrote:
        | If there is one thing I know about tech companies in the
        | last 20 years, it's that they never want VC money unless
        | they are in trouble... right?
 
    | thrown_22 wrote:
    | It resulted in people _saying_ all those things happened, but
    | never did.
 
      | threeseed wrote:
      | You mean like this list of banned subreddits:
      | 
      | https://en.wikipedia.org/wiki/Controversial_Reddit_communit
      | i...
      | 
      | "The community (Beatingwomen), which featured graphic
      | depictions of violence against women, was banned after its
      | moderators were found to be sharing users' personal
      | information online"
      | 
      | "According to Reddit administrators, photos of gymnast
      | McKayla Maroney and MTV actress Liz Lee, shared to 130,000
      | people on popular forum r/TheFappening, constitute child
      | pornography"
 
        | thrown_22 wrote:
        | You mean like the people who are telling us that happened
        | also said:
        | 
        | > CNN is not publishing "HanA*holeSolo's" name because he
        | is a private citizen who has issued an extensive
        | statement of apology, showed his remorse by saying he has
        | taken down all his offending posts, and because he said
        | he is not going to repeat this ugly behavior on social
        | media again. In addition, he said his statement could
        | serve as an example to others not to do the same.
        | 
        | >CNN reserves the right to publish his identity should
        | any of that change.
        | 
        | https://edition.cnn.com/2017/07/04/politics/kfile-reddit-
        | use...
        | 
        | Yeah, I totally trust these people to not lie.
 
    | tick_tock_tick wrote:
    | And Reddit was still a better site back then.
    | 
    | Their anything goes policy is also a huge part of what made
    | them successful back in the day.
 
    | P_I_Staker wrote:
    | I think it's sad that this seems to be getting so many
    | downvotes. You don't have to agree, but this was helpful
    | commentary.
    | 
    | Reddit definitely had all of these issues, and they were
    | handled horribly.
 
      | rchaud wrote:
      | Digg handled things horribly. Reddit seems to have done
      | just fine.
 
    | dncornholio wrote:
    | So instead of moderating users, you moderate moderators,
    | still seems like a net win.
 
      | [deleted]
 
| tjoff wrote:
| > _Machine learning algorithms are able to accurate identify
| spam_
| 
| Nope. Not even close.
| 
| > _and it`s not because they are able to tell it`s about Viagra
| or mortgage refinancing_
| 
| Funny, because they can't even tell that.
| 
| Which is why mail is being ruined by google and microsoft. Yes
| you could argue that they have incentives to do just that. But
| that doesn't change the fact that they can't identify spam.
 
  | sbarre wrote:
  | Do you have more info on why you believe this?
  | 
  | My experience has been that Google successfully filters spam
  | from my Inbox, consistently.
  | 
  | I get (just looked) 30-40 spam messages a day. I've been on
  | Gmail since the invite-only days, so I'm in a lot of lists I
  | guess..
  | 
  | Very Very rarely do they get through the filter.
  | 
  | I also check it every couple of days to look for false-
  | positives, and maybe once a month or less I find a newsletter
  | or automated promo email in there for something I was actually
  | signed up for, but never anything critical.
 
    | [deleted]
 
    | tjoff wrote:
    | Just see what gets through and more importantly which valid
    | mails are being marked as spam. It is evident that they
    | haven't got a clue.
    | 
    | So, how do they "solve" it? By using the "reputation" of your
    | IP addresses and trust that more than the content of the
    | email.
 
      | zimpenfish wrote:
      | I've got 6 mails in my gmail spam (in the last month) - 2
      | of them are not spam which is about normal for what I see
      | (30-40% non-spam.)
 
        | tjoff wrote:
        | Yet most people don't seem to ever look in their spam
        | folder. Conclusion: gmail has a great spam-filter! :(
 
      | ketzo wrote:
      | You're talking way too hyperbolically to take seriously.
      | 
      | Yes, GMail does, in fact, "have a clue." They do pretty
      | well. They're not perfect, and I have specific complaints,
      | but to pretend they're totally clueless and inept
      | discredits anything else you're saying.
 
        | tjoff wrote:
        | Just as saying that machine learning can identify spam
        | discredits anything else ex-reddit CEO says.
        | 
        | I'm sure gmail have a clue from their point of view, but
        | those doesn't align with mine (nor, I'd argue, most of
        | their users). Their view also as a coincidence happens to
        | strengthen their hold on the market but who cares?
 
| fulafel wrote:
| There seem to be no mention of (de)centralization or use of
| reputation in the comments here or in the twitter thread.
| 
| Everyone is discussing a failure mode of a centralized and
| centrally moderated system and aren't questioning those
| properties, but it's really counter to traditional internet based
| communication platforms like email, usenet, irc etc.
 
| excite1997 wrote:
| He frames this as a behavior problem, not content problem. The
| claim is that your objective as a moderator should to get rid of
| users or behaviors that are bad for your platform, in the sense
| that they may drive users away or make them less happy. And that
| if you do that, you supposedly end up with a fundamentally robust
| and apolitical approach to moderation. He then proceeds to blame
| others for misunderstanding this model when the outcomes appear
| politicized.
| 
| I think there is a gaping flaw in this reasoning. Sometimes, what
| drives your users away or makes them less happy _is_ challenging
| the cultural dogma of a particular community, and at that point,
| the utilitarian argument breaks down. If you 're on Reddit, go to
| /r/communism and post a good-faith critique of communism... or go
| to /r/gunsarecool and ask a pro-gun-tinged question about self-
| defense. You will get banned without any warning. But that ban
| passes the test outlined by the OP: the community does not want
| to talk about it precisely because it would anger and frustrate
| people, and they have no way of telling you apart from dozens of
| concern trolls who show up every week. So they proactively
| suppress dissent because they can predict the ultimate outcome.
| They're not wrong.
| 
| And that happens everywhere; Twitter has scientifically-sounding
| and seemingly objective moderation criteria, but they don't lead
| to uniform political outcomes.
| 
| Once you move past the basics - getting rid of patently malicious
| / inauthentic engagement - moderation becomes politics. There's
| no point in pretending otherwise. And if you run a platform like
| Twitter, you will be asked to do that kind of moderation - by
| your advertisers, by your users, by your employees.
 
  | Atheros wrote:
  | > Challenging the cultural dogma [doesn't work]
  | 
  | That is a byproduct of Reddit specifically. With 90s style
  | forums, this kind of discussion happens just fine because it
  | ends up being limited to a few threads. On Reddit, all
  | community members _must_ interact in the threads posted in the
  | last day or two. After two days they are gone and all previous
  | discussion is effectively lost. So maybe this can be fixed by
  | having sub-reddits sort topics by continuing engagement rather
  | than just by age and upvotes.
  | 
  | A good feature would be for Reddit moderators to be able to set
  | the desired newness for their subreddit. /r/aww should strive
  | for one or two days of newness (today's status quo). But
  | /r/communism can have one year of newness. That way the
  | concerned people and concern trolls can be relegated to the
  | yearly threads full of good-faith critiques of communism and
  | the good-faith responses and everyone else can read the highly
  | upvoted discussion. Everything else could fall in-between.
  | /r/woodworking, which is now just people posting pictures of
  | their creations, could split: set the newness to four months
  | and be full of useful advice; set the newness for
  | /woodworking_pics to two days to experience the subreddit like
  | it is now. I feel like that would solve a lot of issues.
 
    | bombcar wrote:
    | The whole idea of "containment threads" is a powerful one
    | that works very well in older-style forums, but not nearly as
    | well on Reddit. "containment subs" isn't the same thing at
    | all, and the subs that try to run subsubs dedicated to the
    | containment issues usually find they die out.
 
| rootusrootus wrote:
| Having read everything he wrote, it makes it interesting to see
| how the discussion on HN matches.
 
| cwkoss wrote:
| Yishan could really benefit from some self editing. There are
| like 5 tweets worth of interesting content in this hundred tweet
| meandering thread.
 
  | bruce343434 wrote:
  | It might just be an effect of the medium.
 
    | bombcar wrote:
    | I mean it's clearly obviously designed to get people to read
    | the ads he has in it.
 
| [deleted]
 
| MichaelZuo wrote:
| There are some neat ideas raised by Yishan.
| 
| One is 'put up or shutup' for appeals of moderator decisions.
| 
| That is anyone who wishes to appeal needs to also consent to have
| all their activities on the platform, relevant to the decision,
| revealed publicly.
| 
| It definitely could prevent later accusations of secretiveness or
| arbitrariness. And it probably would also make users think more
| in marginal cases before submitting.
 
  | wyldberry wrote:
  | This also used to be relatively popular in the early days of
  | League of Legends, people requesting a "Lyte Smite". Players
  | would make inflammatory posts on the forums saying they were
  | banned wrongly, and Lyte would come in with the chatlogs,
  | sometimes escalating to perma-ban. I did always admire this
  | system and thought it could be improved.
  | 
  | There's also a lot of drama around Lyte in his personal life,
  | should you choose to go looking into that.
 
  | cloverich wrote:
  | It is expensive to do, because you have to ensure the content
  | being made public doesn't dox / hurt someone other than the
  | poster. But I think you could add two things to the recipe. 1 -
  | real user validation. So the banned user can't easily make
  | another account. Obviously not easy and perhaps not even
  | possible, but essential. 2 - increased stake. Protest a short
  | ban, and if you lose, you get an even longer ban.
 
  | TulliusCicero wrote:
  | I've never understood that idea that PM's on a platform must be
  | held purely private by the platform even in cases where:
  | 
  | * There's some moderation dispute that involves the PM's
  | 
  | * At least one of the parties involved consents to release the
  | PM's
  | 
  | The latter is the critical bit, to me. When you send someone a
  | chat message, or an email, obviously there's nothing actually
  | stopping them from sharing the content of the message with
  | others if they feel that way, either legally or technically. If
  | an aggrieved party wants to share a PM, everyone knows they can
  | do so -- the only question mark is that they may have faked it.
  | 
  | To me the answer here seems obvious: allow users to mark a
  | PM/thread as publicly visible. This doesn't make it more public
  | than it otherwise could be, it just lets other people verify
  | the authenticity, that they're not making shit up.
 
  | whitexn--g28h wrote:
  | This is something that occurs on twitch streams sometimes.
  | While it can be educational for users to see why they were
  | banned, some appeals are just attention seeking. Occasionally
  | though it exposes the banned user's or worse a victim users
  | personal information, (eg mental health issues, age, location)
  | and can lead to both users being targeted and bad behaviour by
  | the audience. For example Bob is banned for bad behaviour
  | towards Alice (threats, doxxing), by making that public you are
  | not just impacting Bob, but could also put Alice at risk.
 
  | etchalon wrote:
  | I think this idea rests on the foundation of "shame."
  | 
  | But there are entire groups of users that not only don't feel
  | shame about their activities, but are proud of them.
 
    | codemonkey-zeta wrote:
    | But those users would be left alone in their pride in the
    | put-up-or-shut-up model, because everybody else would see the
    | mistakes of that user and abandon them. So the shame doesn't
    | have to be effective for the individual, it just has to
    | convince the majority that the user is in the wrong.
 
      | kelnos wrote:
      | Right. To put it another way, this "put up or shut up"
      | system, in my mind, isn't even really there to convince the
      | person who got moderated that they were in the wrong. It's
      | to convince the rest of the community that the moderation
      | decision was unbiased and correct.
      | 
      | These news articles about "platform X censors people with
      | political views Y" are about generating mass outrage from a
      | comparatively small number of moderation decisions. While
      | sure, it would be good for the people who are targeted by
      | those moderation decisions to realize "yeah, ok, you're
      | right, I was being a butthole", I think it's much more
      | important to try to show the reactionary angry mob that
      | things are aboveboard.
 
        | etchalon wrote:
        | The most high profile, and controversial, "moderation"
        | decisions made by large platforms recently have generally
        | been for obvious, and very public, reasons.
 
| shashanoid wrote:
 
  | kahrl wrote:
  | Yishan Wong is an American engineer and entrepreneur who was
  | CEO of Reddit from March 2012 until his resignation in November
  | 2014.
  | 
  | Did you need help looking that up? Or were you just being edgy?
 
| ilyt wrote:
| It's kinda funny that many of the problems he's mentioning is
| exactly how moderation on reddit currently works.
| 
| Hell, newly revamped "block user" mode got extra gaslighting as a
| feature, now person blocked can't reply to _anyone_ under the
| comment of person that blocked them, not just the person that
| blocked them so anyone that doesn 't like people discussing how
| they are wrong can just ban the people that disagree with them
| and they will not be able to answer to any of their comments.
 
  | Ztynovovk wrote:
  | Seems reasonable to me. IRL I can walk away from a voluntary
  | discussion when I want. If people want to continue talking
  | after I've left they can form their own discussion group and
  | continue with the topic.
  | 
  | Think this is good because it usually stops a discussion from
  | dissolving into a meaningless flame war.
  | 
  | It allows the power of moderation to stay within the power of
  | those in the discussion.
 
    | scraptor wrote:
    | Now imagine if some random other people in the group who
    | happen to have posts higher in the tree were able to silently
    | remove you without anyone knowing.
 
      | Ztynovovk wrote:
      | Meh, it's the most reasonable bad solution imo. I've had
      | some pretty heated convos on reddit and have only ever been
      | blocked once.
 
| chinchilla2020 wrote:
| The tweetstorm format is such a horrible way to consume articles.
| I cannot wait for twitter to collapse so I never have to read
| another essay composed of 144-word paragraphs.
 
| swarnie wrote:
| Twitter has to be the worst possible medium for reading an essay.
 
  | joemi wrote:
  | You're far from the only person who thinks this, but please see
  | dang's stickied comment at the top of the thread.
 
| ItsBob wrote:
| Here's a radical idea: let me moderate my own shit!
| 
| Twitter is a subscription-based system (by this, I mean that I
| have to subscribe to someone's content) so if I subscribe to
| someone and don't like what they say then buh-bye!
| 
| Let me right click on a comment/tweet (I don't use social media
| so not sure of the exact terminology the kids use these days)
| with the options of:
| 
| - Hide this comment
| 
| - Hide all comments in this thread from 
| 
| - Block all comments in future from  (you can undo this in
| settings).
| 
| That would work for me.
 
  | threeseed wrote:
  | You're not Twitter's customer. Advertisers are.
  | 
  | And they don't want their brands to be associated with
  | unpleasant content.
 
    | q1w2 wrote:
    | To quote the article...
    | 
    | > MAYBE sometimes an advertiser will get mad, but a backroom
    | sales conversation will usually get them back once the whole
    | thing blows over.
 
    | lettergram wrote:
    | People producing products don't actually care. I'd love to
    | see stats on this made public (I've seen internal metrics).
    | Facebook and Twitter don't even show you what your ad is
    | near. You fundamentally just have to "trust" them.
    | 
    | If you have a billboard with someone being raped beneath it
    | and a photo goes viral, no one would blame the company
    | advertising on the billboard. Frankly, no one will associate
    | the two to change their purchasing habits.
    | 
    | The reason corporations care are the ESG scores and activist
    | employees.
    | 
    | Also these brands still advertise in places where public
    | executions will happen (Saudi Arabia). No one is complaining
    | there.
 
      | pfisch wrote:
      | People do care. If you openly associate your brand with
      | strong support for a pro pedophile or pro rape position
      | customers will care about that.
      | 
      | The idea that they won't seems pretty ridiculous.
 
      | threeseed wrote:
      | > Facebook and Twitter don't even show you what your ad is
      | near
      | 
      | But their customers complain about it, media picks it up
      | and it becomes an outrage story.
      | 
      | That's what brands are scared of.
 
    | Spivak wrote:
    | Like I can't believe that this reasoning doesn't resonate
    | with people even outside of advertisers. Who wants to be on a
    | social network where if one of your posts breaks containment
    | you spend the next few weeks getting harassed by people who
    | just hurl slurs and insults at you. This is already right now
    | a problem on Twitter and opening the floodgates is the
    | opposite of helping.
 
      | etchalon wrote:
      | This reasoning is generally lost on people whom are
      | generally not a target for slurs and harassment.
 
      | tick_tock_tick wrote:
      | > few weeks getting harassed by people who just hurl slurs
      | and insults at you
      | 
      | Just ignore it or block them. The only time it's an issue
      | is when you engage. Seriously the only people with this
      | issue can't let shit go.
 
        | fzeroracer wrote:
        | I feel like you don't understand the issue here at all.
        | 
        | Blocking them requires first engaging with their content.
        | This is what people always miss in the discussion. To
        | know if you need to block someone or not involves parsing
        | their comment and then throwing it in the bin.
        | 
        | The same goes for ignoring it. And eventually people get
        | tired of the barrage of slurs and just leave because the
        | brainpower required to sift through garbage isn't worth
        | it anymore. That's how you end up with places like Voat.
 
    | tick_tock_tick wrote:
    | Most customers don't care the only reason it's a real issue
    | is Twitter users run the marketing department at a lot of
    | companies and they are incorrectly convinced people care.
 
      | spaced-out wrote:
      | >Most customers don't care
      | 
      | How much is "most"? What data do you have? Plus, even if
      | ~20% of customers care and only half will boycott, that's
      | still going to have an impact on the company's bottom line.
 
      | threeseed wrote:
      | > Twitter users run the marketing department
      | 
      | If that were the case why is Twitter ad spend in the low
      | single digits for most companies.
 
  | dbbk wrote:
  | This exists.
 
  | spaced-out wrote:
  | Maybe you might like that, but I personally don't want to wade
  | through dozens of "MAKE BIG $$$$ WORKING FROM HOME!!!" every
  | morning on my feed.
 
    | int_19h wrote:
    | This is solved by allowing people to "hire" others as their
    | moderators.
 
      | tedunangst wrote:
      | Why can't I "hire" (join) a social network that
      | preemptively mods according to my preferences?
 
        | int_19h wrote:
        | Because there are too few, due to market dominance of
        | existing players?
 
  | AhmedF wrote:
  | Try moderating 100+ hateful messages an hour.
 
  | AceJohnny2 wrote:
  | You've never been targeted for harassment, obviously.
  | 
  | Blocking a comment, or even blocking a user for a comment is
  | useless on platforms that allow free and endless user accounts.
  | 
  | Mail spam/scam folders of everyone's email accounts are proof
  | that "let me moderate it myself" does not work for the majority
  | of people.
  | 
  | And remember "It is harder to police bad behavior than it is to
  | automate it."
 
    | commandlinefan wrote:
    | > let me moderate it myself
    | 
    | More like "let us moderate it ourselves". Reddit users
    | already do this - there are extensions you can install that
    | allow you to subscribe to another group of user's ban list.
    | So you find a "hivemind" that you mostly agree with, join
    | their collective moderation, and allow that to customize the
    | content you like. The beauty is that _you_ get to pick the
    | group you find most reasonable.
 
  | pjc50 wrote:
  | > - Block all comments in future from  (you can undo this
  | in settings).
  | 
  | This is what the existing block feature does?
 
| dimva wrote:
| His argument makes no sense. If this is indeed why they are
| banning people, why keep the reasoning a secret? Honestly, every
| ban should come with a public explanation from the network, in
| order to deter similar behavior. The way things are right now,
| it's unclear if, when, and for what reason someone will be
| banned. People get banned all the time with little explanation or
| explanations that make no sense or are inconsistent. There is no
| guidance from Twitter on what behavior or content or whatever
| will get you banned. Why is some rando who never worked at
| Twitter explaining why Twitter bans users?
| 
| And how does Yishan know why Twitter bans people? And why should
| we trust that he knows? As far as I can tell, bans are almost
| completely random because they are enacted by random low-wage
| contract workers in a foreign country with a weak grasp of
| English and a poor understanding of Twitter's content policy (if
| there even is one).
| 
| Unlike what Yishan claims, it doesn't seem to me like Twitter
| cares at all about how pleasant an experience using Twitter is,
| only that its users remain addicted to outrage and calling-out
| others, which is why most Twitter power-users refer to it as a
| "hellsite".
 
  | sangnoir wrote:
  | > Honestly, every ban should come with a public explanation
  | from the network, in order to deter similar behavior
  | 
  | This only works on non-adversarial systems. Anywhere else, it
  | will be like handing over to bad actors (i.e. people whose
  | interests will _never_ align with operator 's) a list of
  | blindspots
 
    | noasaservice wrote:
    | "You have been found guilty of crimes in $State. Please
    | submit yourself to $state_prison on the beginning of the next
    | month. We're sorry, but we cannot tell you what you are
    | guilty of."
 
      | vkou wrote:
      | "Look, I'd like you to stop being a guest in my house,
      | you're being an asshole."
      | 
      | "PLEASE ENUMERATE WHICH EXACT RULES I HAVE BROKEN AND
      | PROVIDE ME WITH AN IMPARTIAL AVENUE FOR APPEAL."
      | 
      | ---
      | 
      | When you're on a platform, you are a guest. When you live
      | in society, you don't have a choice about following the
      | rules. That's why most legal systems provide you with clear
      | avenues for redress and appeal in the latter, but most
      | private property does not.
 
    | 10000truths wrote:
    | Imagine for a moment what would happen if this rationale were
    | extended to the criminal justice system. Due process is
    | sanctified in law for a good reason. Incontestable
    | assumptions of adversarial intent are the slow but sure path
    | to the degradation of any community.
    | 
    | There will _always_ be blind spots and malicious actors, no
    | matter how you structure your policies on content moderation.
    | Maintaining a thriving and productive community requires
    | active, human effort. Automated systems can be used to
    | counteract automated abuses, but at the end of the day, you
    | need _human_ discretion /judgement to fill those blind spots,
    | adjust moderation policies, proactively identify
    | troublemakers, and keep an eye on people toeing the line.
 
      | Spivak wrote:
      | Being cagey about the reasons for bans is
      | 
      | 1. To keep people from cozying up to the electric fence. If
      | you don't know where the fence is you'll probably not risk
      | a shock trying to find it. There are other ways one can
      | accomplish this like bringing the banhammer down on
      | everyone near the fence every so often very publicly but
      | point 2 kinda makes that suck.
      | 
      | 2. To not make every single ban a dog and pony show when
      | it's circulated around the blogspam sites.
      | 
      | I'm not gonna pass judgement as to whether it's a good
      | thing or not but it's not at all surprising that companies
      | plead the 5th in the court of public opinion.
 
        | bink wrote:
        | Sorta related to (1) but not really: there are also more
        | "advanced" detection techniques that most sites use to
        | identify things like ban evasion and harassment using
        | multiple accounts. If they say "we identified that you
        | are the same person using this other account and have
        | reason to believe you've created this new account solely
        | to evade that ban" then people will start to learn what
        | techniques are being used to identify multiple accounts
        | and get better at evading detection.
 
      | sangnoir wrote:
      | > Imagine for a moment what would happen if this rationale
      | were extended to the criminal justice system.
      | 
      | It already is!
      | 
      | The criminal justice system is a perfect example of why
      | total information transparency is a terrible idea: _never
      | talk to the cops_ even if they just want to  "get one thing
      | cleared up" - your intentions don't matter, you're being
      | given more rope to hang yourself with.
      | 
      | It's an adversarial system where transparency gets you
      | little, but gains your adversary a whole lot. You should
      | not ever explain your every action and reasoning to the
      | cops without your lawyer telling you when to STFU.
      | 
      | Due process is sanctified, but the criminal justice system
      | is self-aware enough to recognize that self-incrimination
      | is a hazard, and rightly does not place the burden on the
      | investigated/accused, why should other adversarial system
      | do less?
 
  | ascv wrote:
  | Honestly it seems like you didn't read the thread. He's not
  | talking about how Twitter itself works but about problems in
  | moderation more generally based on his experience at Reddit.
  | Also, he specifically advocates public disclosure on ban
  | justifications (though acknowledges it is a lot of work).
 
    | dang wrote:
    | He also makes an important and little-understood point about
    | asymmetry: the person who posts complaints about being
    | treated unfairly can say whatever they want about how they
    | feel they were treated, whereas the moderation side usually
    | can't disclose everything that happened, even when it would
    | disprove what that user is saying, because it's operating
    | under different constraints (e.g. privacy concerns).
    | Ironically, sometimes those constraints are there to protect
    | the very person who is making false and dramatic claims. It
    | sucks to be on that side of the equation but it's how the
    | game is played and the only thing you can really do is learn
    | how to take a punch.
 
  | roblabla wrote:
  | From my understanding, he's not claiming this is how twitter
  | currently works. He's offering advice about how to solve
  | content moderation on twitter.
 
    | dontknowwhyihn wrote:
    | He's offering advice that differs from what Reddit does in
    | practice. They absolutely ban content rather than behavior.
    | Try questioning "the science" and it doesn't matter how
    | considerate you are, you will be banned.
 
      | CountHackulus wrote:
      | He covers that further down in the tweets, near the end of
      | the thread. He doesn't necessarily agree with the Reddit
      | way of doing things, but it has interesting compromises wrt
      | privacy.
 
      | pixl97 wrote:
      | Because no one has developed a moderation framework based
      | on behavior. Content is (somewhat) easy, a simple regex can
      | capture that. Behavior is far more complicated and even
      | more subject to our biases.
 
  | bitxbitxbitcoin wrote:
  | He's specifically referring to Reddit's content moderation
  | which actually has two levels of bans. Bans by mods from a
  | specific subreddit are done by mods from that specific
  | subreddit and having an explanation isn't required but is
  | sometimes given - these bans apply to just the subreddit and
  | are more akin to a block by the community. Bans by admins
  | happen to people that have been breaking a site rule, not a
  | subreddit rule.
  | 
  | Both types of bans have privacy issues that result in lack of
  | transparency with bans.
 
| matchagaucho wrote:
| tldr; Many posts on social media are "spam". Nobody objects to
| spam filters.
| 
| Therefore, treat certain types of content as spam (based on
| metadata, not moderators).
 
| ufo wrote:
| In the US, where Twitter & Facebook are dominant, the current
| consensus in the public mind is that political polarization and
| radicalization are driven by the social media algorithms.
| However, I have always felt that this explanation was lacking.
| Here in Brazil we have many of the same problems but the dominant
| social media are Whatsapp group chats, which have no algorithms
| whatsoever (other than invisible spam filters). I think Yishan is
| hitting the nail on the head by focusing the discussion on user
| behavior instead of on the content itself.
 
  | MikePlacid wrote:
  | > I think Yishan is nailing the nail on the head by focusing
  | the discussion on user behavior instead of on the content
  | itself.
  | 
  | But user behavior problem can be solved cheaply, easily and in
  | a scalable way:
  | 
  | Give each user an ability to form the personal filter.
  | Basically, all what I need is:
  | 
  | 1. I want to read what person A writes - always.
  | 
  | 2. I want to read what person B writes - always, except when
  | talking to person C.
  | 
  | 3. I want to peek through a filter of the person I like - to
  | discover more interesting to me persons.
  | 
  | 4. Show me random people posts like 3-4 (configurable) times
  | per day.
  | 
  | This is basically how my spinal brain worked in unmoderated
  | Fido-over-Usenet groups. Some server help will be great, sure,
  | but there is nothing here that is expensive or not scalable.
  | PS: centralized filtering is needed only when you are going
  | after some content, not noise.
 
    | ufo wrote:
    | I disagree, we can't frame this discussion on only the
    | content. My whatsapp "feed" is doing just fine. The problem
    | are all the other whatsapp groups that I'm not in, which are
    | filled with hateful politics. It hurts to when you meet in
    | real life a friend that you haven't met in a while, and then
    | find out that they've been radicalized.
    | 
    | The radical Bolsonaro whatsapp groups are a mix of top down
    | and grass roots content. On one end there is the central
    | "propaganda office", or other top political articulators. On
    | the bottom are the grassroots group chats: neightborhoods,
    | churches, biker communities, office mates, etc. Memes and
    | viral content flow in both directions. The contents and
    | messages that ressonate in the lower levels get distributed
    | by the central articulators, which have a hierarchy of group
    | chats to circulate new propaganda as widely as possible. You
    | can see this happen in real time when a political conundrum
    | happens,e.g a corruption scandal. The central office will A-B
    | test various messages in their group chats and then the one
    | that resonates better with their base gets amplified and
    | suddenly they manage to "change the topic" on the news. The
    | end result is that we just had 4 years of political chaos,
    | where the modus operandi of the goverment was to put out
    | fires by deflecting the public discourse whenever a new
    | crisis emerged. It's not a problem of content itself, that
    | could be solved by a better filtration algorithm. It's a
    | deeper issue having to do with how quickly memetic ideas can
    | spread in this new world of social media.
 
  | originalvichy wrote:
  | I actually went into a deep dive of any statistical efforts
  | that showed bans on twitter based on American political
  | leanings.
  | 
  | Apparently in both studies I found the most statistically
  | significant user behavior for bans was if the user had a
  | tendency to post articles from low quality online "news" sites.
  | 
  | So essentially even the political controversy around moderation
  | boils down to the fact that one side, the right, is happily
  | posting low quality news/fake news that they either get banned
  | for disinformation or other rule-breaking behavior.
  | 
  | https://psyarxiv.com/ay9q5
 
    | leereeves wrote:
    | The first concern is identifying "fake news".
    | 
    | One of the biggest political news stories of 2020, Hunter
    | Biden's laptop, was falsely declared misinformation, and the
    | NY Post was accused of being a low quality site. Now we know
    | it's true.
    | 
    | On the other hand, the Steele Dossier was considered
    | legitimate news at the time and "many of the dossier's most
    | explosive claims...have never materialized or have been
    | proved false."[1].
    | 
    | So I'd like to know exactly what the study's authors
    | considered low-quality news, but unfortunately I couldn't
    | find a list in the paper you linked. In my experience, most
    | people tend to declare sources "high-quality" or "low-
    | quality" based on whether they share the same worldview.
    | 
    | 1: https://www.nytimes.com/2021/05/15/business/media/spooked-
    | pr...
 
      | haberman wrote:
      | I had the same skepticism as you, but the study authors did
      | attempt to be fair by letting a set of politically balanced
      | laypeople (equal number Democrat and Republican) adjudicate
      | the trustworthiness of sites. They also had journalists
      | rate the sites, but they present the results of both
      | results (layperson-rated and journalist-rated).
      | 
      | I wish they had included the list so we could see for
      | ourselves. It's still possible that there are flaws in the
      | study. But the it appears to take the concern of fairness
      | more seriously than most.
 
    | nfgivivu wrote:
 
  | lifeisstillgood wrote:
  | This to me is a vital point.
  | 
  | One of the things rarely touched on about Twitter / FB et al
  | are that they are transmission platforms and then a discovery /
  | recommendation layer on top.
  | 
  | The "algorithm" is this layer on top and it is assumed that
  | this actively sorts people into their bubbles and people
  | passively follow - there is much discussion about splitting the
  | companies AT&T style to improve matters.
  | 
  | But countries where much of the discourse is on WhatsApp do not
  | have WhatsApp to do this recommendation - it is done IRL
  | (organically) - and people actively sort themselves.
  | 
  | The problem is not (just) the social media companies. It lies
  | in us.
  | 
  | The solution if we are all mired in the gutter of social media,
  | is to look up and reach for the stars.
 
| monksy wrote:
| > No, what`s really going to happen is that everyone on council
| of wise elders will get tons of death threats, eventually quit...
| 
| Yep if you can't stand being called an n* (or other racial slurs)
| don't be a reddit moderator. Also I've been called a hillary boot
| licker and a trump one.
| 
| Being a reddit moderator isn't for the thin of skinned.I hosted
| social meetups so this could have run out in the real
| world..Luckily I had a strong social support in the group where
| that would have been taken care of real quick. I've only had one
| guy that tried to threaten to come and be disruptive at one of
| the meetups. He did come out. He did meet me.
| 
| ----
| 
| > even outright flamewars are typically beneficial for a small
| social network:
| 
| He's absolutely correct. It also helps to define community
| boundries and avoid extremism. A lot of this "don't be mean"
| culture only endorses moderators stepping in and dictating how a
| community talks and how people who disagree are officially
| bullied.
 
| fuckHNtho wrote:
| tldr tangential babbling that HN protects and wants us to
| admire...because reddit YC darlings. it almost makes me feel
| nostalgic.
| 
| Why are we to take yishan as an authority on content moderation,
| have you BEEN to reddit?! the kind of moderation of repetitive
| content he's referring to is clearly not done AT ALL.
| 
| He does not put forth any constructive advice. be "operationally
| excellent". ok, thanks. you're wrong about spam. you're wrong
| about content moderation. ok, thanks. who is his audience? he's
| condescending the people who are dialed into online discourse
| inbetween finding new fun ways to plant trees and design an
| indulgent hawaiian palace. i expected more insight, to be honest.
| but time and time again we find the people at the top of internet
| companies are disappointingly common in their perspective on the
| world. they just happened to build something great once and it
| earned them a lifetime soapbox ticket.
| 
| ok, thanks.
 
| P_I_Staker wrote:
| Key word here: ex (joking)... but seriously I'm absolutely
| baffled why someone would look to a former reddit exec for advice
| on moderation.
| 
| I guess you could say that they have experience, having made all
| the mistakes, and figured it out through trial and error! This
| seems to be his angle.
| 
| What I got from the whole reddit saga is how horrible the
| decision making was, and won't be looking to them for sage
| advice. These people are an absolute joke.
 
  | mikkergp wrote:
  | Who is doing a good job at scale? Is there really anyone we can
  | look to other than people who "have experience, having made all
  | the mistakes, and figured it out through trial and error"?
 
    | P_I_Staker wrote:
    | Sorry if this wasn't clear, but that's just his perspective.
    | Mine is that they're a bunch of clowns with little to offer
    | anyone. Who cares what this person thinks more than you, I,
    | or a player for the Miami Dolphins.
    | 
    | Twitter is going to have to moderate at least exploitative
    | and a ton of abusive content, eventually. I don't understand
    | how this rant is helpful in the slightest. Seemed like a
    | bunch of mumbo jumbo.
    | 
    | You do have a good point about there not being very many good
    | actors, if any.
 
      | dbbk wrote:
      | Who cares what this person thinks? They actually have
      | experience tackling the problem. You or I have never been
      | in a position of tackling the problem. Of course I am
      | interested in the experience of someone who has seen this
      | problem inside and out.
 
| armchairhacker wrote:
| I wonder if the problems the author describes can be solved by
| artifically downvoting and not showing spam and flamewar content,
| not banning people.
| 
| - Spam: don't show it to anyone, since nobody wants to see it.
| Repeatedly saying the same thing will get your posts heavily
| downvoted or just coalesced into a single post.
| 
| - Flamewars: again, artifically downvote them so that your
| average viewer doesn't even see them (if they aren't naturally
| downvoted). And also discourage people from participating, maybe
| by explicitly adding the text "this seems like a stupid thing to
| argue about" onto the thread and next to the reply button. The
| users who persist in flaming each other and then get upset, at
| that point you don't really want them on your platform anyways
| 
| - Insults, threats, etc: again, hide and reword them. If it
| detects someone is sending an insult or threat, collapse it into
| "" or "" so that people know the content of
| what's being sent but not the emotion (though honestly, you
| probably should ban threats altogether). You can actually do this
| for all kinds of vitriolic, provocative language. If someone
| wants to hear it, they can expand the "" bubble, the
| point is that most people probably don't.
| 
| It's an interesting idea for a social network. Essentially,
| instead of banning people and posts outright, down-regulate them
| and collapse what they are saying while remaining the content. So
| their "free speech" is preserved, but they are not bothering
| anyone. If they complain about "censorship", you can point out
| that the First Amendment doesn't require anyone to hear you, and
| people _can_ hear you if they want to, but the people have
| specified and algorithm detects that they don 't.
| 
| EDIT: Should also add that Reddit actually used to be like this,
| where subreddits had moderators but admins were very hands-off
| (actually just read about this yesterday). And it resulted in
| jailbait and hate subs (and though this didn't happen, could have
| resulted in dangerous subs like KiwiFarms). I want to make clear
| that I still think that content should be banned. But that
| content isn't what the author is discussing here: he is
| discussing situations where "behavior" gets people banned and
| then they complain that their (tame) content is being censored.
| Those are the people who should be down-regulated and text
| collapsed instead of banned.
 
| pluc wrote:
| Reddit uses an army of free labour to moderate.
 
| ConanRus wrote:
 
| jamisteven wrote:
| How about, dont moderate it? Just, let it be.
 
| jameskilton wrote:
| Every single social media platform that has ever existed makes
| the same fundamental mistake. They believe that they just have to
| remove or block the bad actors and bad content and that will make
| the platform good.
| 
| The reality is _everyone_ , myself included, can be and will be a
| bad actor.
| 
| How do you build and run a "social media" product when the very
| act of letting anyone respond to anyone with anything is itself
| the fundamental problem?
 
  | [deleted]
 
  | onion2k wrote:
  | _The reality is everyone, myself included, can be and will be a
  | bad actor._
  | 
  | Based on this premise we can conclude that the best way to
  | improve Reddit and Twitter is to block everyone.
 
    | madeofpalk wrote:
    | To be honest, I would not disagree with that. Very 'the only
    | winning move is not to play'.
 
      | PM_me_your_math wrote:
      | To be honest, and maybe this will be panned, but the real
      | answer is for people to grow thicker skin and stop putting
      | one's feelings on a pedestal above all.
 
        | mikkergp wrote:
        | Interesting, that wasn't my interpretation of the twitter
        | thread, it was more that spam and not hurtful content was
        | the real tricky thing about moderating social media.
 
        | chipotle_coyote wrote:
        | Spam was more of an example than the point, I think --
        | the argument Yishan is making is that moderation isn't
        | for _content,_ it 's for _behavior._ The problem is that
        | if bad behavior is tied to partisan and /or controversial
        | content, which it often is, people react as if the
        | moderation is about the content.
 
        | pwinnski wrote:
        | When _I_ make comments about _other_ people, they need to
        | grow thicker skin.
        | 
        | When _other_ people attack _me_ personally, it 's a deep
        | and dangerous violation of social norms.
 
        | madeofpalk wrote:
        | I don't block people because they hurt my feelings, i
        | block people because im just not interested in seeing
        | bird watching content on my timeline. No one deserves my
        | eyeballs.
 
        | dang wrote:
        | That's asking human nature to change, or at least asking
        | almost everyone to work on their trauma until they don't
        | get so activated. Neither will happen soon, so this can't
        | be the real answer.
 
        | horsawlarway wrote:
        | Look - I don't even particularly disagree with you, but I
        | want to point out a problem with this approach.
        | 
        | I'm 33. I grew up playing multiplayer video games
        | (including having to run a db9 COM cable across the house
        | from one machine to another to play warcraft 2
        | multiplayer, back when you had to explicitly pick the
        | protocol for the networking in the game menu)
        | 
        | My family worked with computers, so I had DSL since I
        | have memories. I played a ton of online games. The
        | communities are _BRUTAL_. They are insulting, abusive,
        | misogynistic, racist, etc... the spectrum of unmonitored
        | teenage angst, in all it 's ugly forms (and to be fair,
        | some truly awesome folks and places).
        | 
        | As a result - I have a really thick skin about basically
        | everything said online. But a key difference between the
        | late 90s and today, is that if I wanted it to stop, all I
        | had to do was close the game I was playing. Done.
        | 
        | Most social activities were in person, not online. I
        | could walk to my friend's houses. I could essentially
        | tune out all the bullshit by turning off my computer, and
        | there was plenty of other stuff to go do where the
        | computer wasn't involved at all.
        | 
        | I'm not convinced that's enough anymore. The computer is
        | in your pocket. It's always on. Your social life is
        | probably half online, half in person. Your school work is
        | online. Your family is online. your reputation is online
        | (as evidenced by those fucking blue checkmarks). The
        | abuse is now on a highway into your life, even if you
        | want to turn it off.
        | 
        | It's like the school bully is now waiting for you
        | everywhere. He's not waiting at school - he's stepping
        | into the private conversations you're having online. He's
        | talking to your friends. He's hurling abuse at you when
        | you look at your family photos. He's _in_ your life in a
        | way that just wasn 't possible before.
        | 
        | I don't think it's fair to say "Just grow a thicker skin"
        | in response to that. I think growing a thicker skin is
        | desperately needed, but I don't think it's sufficient.
        | The problem is deeper.
        | 
        | We have a concept for people who do the things these
        | users are doing on twitter in person - They're called
        | fighting words, and most times, legally (even in the US)
        | there is _zero_ assumption of protected speech here. You
        | say bad shit about someone with the goal of riling them
        | up and no other value? You have no right of free speech,
        | because you aren 't "speaking" - you're trying to start a
        | fight.
        | 
        | I'm not protecting your ability to bully someone. Full
        | stop. If you want to do that, do it with the clear
        | understanding that you're on your own, and regardless of
        | how thick my skin is - I think you need a good slap
        | upside the head. I'd cheer it on.
        | 
        | In person - this resolves itself because the fuckwads who
        | do this literally get physically beaten. Not always - but
        | often enough we have a modicum of civil discussion we
        | accept, and a point where no one is going to defend you
        | because you were a right little cunt, and the beating was
        | well deserved.
        | 
        | I don't know how you simulate the same constraint online.
        | I'm not entirely sure you can, but I think the answer
        | isn't to just stop trying.
 
        | ryandrake wrote:
        | > The computer is in your pocket. It's always on. Your
        | social life is probably half online, half in person. Your
        | school work is online. Your family is online. your
        | reputation is online (as evidenced by those fucking blue
        | checkmarks). The abuse is now on a highway into your
        | life, even if you want to turn it off.
        | 
        | It is still a choice to participate online. I'm not on
        | Twitter or Facebook or anything like that. It doesn't
        | affect my life in the slightest. Someone could be on
        | there right this minute calling me names, and it can't
        | bother me because I don't see it, and I don't let it into
        | my life. This is not a superpower, it's a choice to not
        | engage with social media and all the ills it brings.
        | 
        | Have I occasionally gotten hate mail from an HN post?
        | Sure. I even got a physical threat over E-mail (LOL good
        | luck, guy). If HN ever became as toxic as social media
        | can be, I could just stop posting and reading. Problem
        | solved. Online is not real if you just ignore it.
 
        | Sohcahtoa82 wrote:
        | The attitude of "If you don't like it, leave!" is
        | allowing the bullies to win.
        | 
        | Minorities, both racial and gender, should be able to use
        | social media without having vitriol spewed at them
        | because they're guilty of being a minority.
 
        | pjc50 wrote:
        | Paul Pelosi's skin wasn't thick enough to deflect
        | hammers.
        | 
        | (OK, that wasn't a twitter problem, but the attack on him
        | was 100% a product of unmoderated media in general)
 
        | PM_me_your_math wrote:
        | I respectfully disagree. Beyond the reason that there is
        | no way you can be 100% certain 'unmoderated media' was
        | the primary motivator. Nobody can presume to know his
        | motivations or inner dialogue. A look at that mans
        | history shows clear mental health issues and self-
        | destructive behavior so we can infer some things but
        | never truly know.
        | 
        | Violence exists outside of mean tweets and political
        | rhetoric. People, even crazy ones, almost always have
        | their own agency even if it runs contrary to what most
        | consider to be normal thoughts and behavior. They choose
        | to act, regardless of others and mostly without concern
        | or conscious. There are crazy people out there and
        | censoring others wont ever stop bad people from doing bad
        | things. If so, then how do we account for the evils done
        | by those prior to our inter-connected world?
 
        | krtzulo wrote:
        | We really don't know much. The attacker used drugs, his
        | ex partner said the he went away for year and came back a
        | changed person.
        | 
        | He lived in a community with BLM signs and a rainbow
        | flag. He did hemp jewellry.
        | 
        | He registered a website three months ago and only
        | recently filled it with standard extreme right garbage.
        | 
        | This is all so odd that for all we know someone else
        | radicalized him offline, the old fashioned way.
 
    | rchaud wrote:
    | You hit the nail on the head, but maybe the other way around.
    | 
    | "Block" and "Mute" are the Twitter user's best friends. They
    | keep the timeline free of spam, be it advertisers, or the
    | growth hackers creating useless threads of Beginner 101 info
    | and racking up thousands of likes.
 
      | gorbachev wrote:
      | After using several communications tools over the past
      | couple of decades (BBSes, IRC, Usenet, AIM, plus the ones
      | kids these days like), I'm convinced blocking and/or muting
      | is required for any digital mass communication tool anyone
      | other than sociopaths would use.
 
    | MichaelZuo wrote:
    | Doesn't Twitter give the option of a whitelist (Just who you
    | follow + their retweets) already?
 
      | mikkergp wrote:
      | Not really, even the it still does recommended tweets and I
      | don't want to see retweets or likes and you have to turn
      | that off per person.
 
        | fknorangesite wrote:
        | Yes, really. I never see 'recommended' or 'so-and-so
        | liked...' in my feed.
 
        | the_only_law wrote:
        | I had to create a new account after losing mine, and
        | without following many people it seems like easily 30-50%
        | of my feed is recommended content.
 
        | fknorangesite wrote:
        | There are "home" and "newest" feeds. I agree it's shitty
        | that the default shows this stuff, but you just have to
        | switch it over to "newest."
 
        | leephillips wrote:
        | Yes really: you can get this experience with Twitter:
        | https://lee-phillips.org/howtotwitter/
 
        | mikkergp wrote:
        | Edit: I apologize that my original post was dismissive of
        | your effort to help people use twitter.
 
        | leephillips wrote:
        | Don't worry about it. I didn't exactly understand your
        | original comment, but I don't have a problem with people
        | having opinions.
 
        | MichaelZuo wrote:
        | That's unnecessarily dismissive of someone trying their
        | best to share some tips. It's not like they charged you
        | to read it.
 
        | mikkergp wrote:
        | Fair enough it wasn't meant to be a commentary on them,
        | but I will edit with an apology
 
  | thrown_22 wrote:
 
  | Invictus0 wrote:
  | I'm not a bad actor, I only have 3 tweets and they're all
  | reasonable IMO. So your premise is wrong.
 
  | phillipcarter wrote:
  | > The reality is everyone, myself included, can be and will be
  | a bad actor.
  | 
  | But you likely aren't, and most people likely aren't either.
  | That's the entire premise behind removing bad actors and spaces
  | that allow bad actors to grow.
 
    | pessimizer wrote:
    | > But you likely aren't, and most people likely aren't
    | either.
    | 
    | Is there any evidence of this? 1% bad content can mean that
    | 1% of your users are bad actors, or it can mean that 100% of
    | your users are bad actors 1% of the time (or anything in
    | between.)
 
      | munificent wrote:
      | I assume all of us have evidence of this in our daily
      | lives.
      | 
      | Even the best people we know have bad days. But you have
      | probably also encountered people in your life who have
      | consistent patterns of being selfish, destructive, toxic,
      | or harmful.
 
        | pessimizer wrote:
        | > you have probably also encountered people in your life
        | who have consistent patterns of being selfish,
        | destructive, toxic, or harmful.
        | 
        | This is not evidence that most bad acts are done by bad
        | people. This is evidence that I've met people who've
        | annoyed or harmed me at one or more points, and projected
        | my personal annoyance into my fantasies of their internal
        | states or of their _essence._ Their  "badness" could
        | literally have only consisted of the things that bothered
        | me, and during the remaining 80% of the time (that I
        | wasn't concerned with) they were tutoring orphans in
        | math.
        | 
        | Somebody who is "bad" 100% of the time on twitter could
        | be bad 0% of the time off twitter, and vice-versa. Other
        | people's personalities aren't reactions to our values and
        | feelings; they're as complex as you are.
        | 
        | As the OP says: our definitions of "badness" in this
        | context are of _commercial_ badness. Are they annoying
        | our profitable users?
        | 
        | edit: and to add a bit - if you have a diverse userbase,
        | you should expect them to annoy each other at a pretty
        | high rate with absolutely no malice.
 
    | simple-thoughts wrote:
    | Your logic makes sense but is not how these moderation
    | services actually work. When I used my own phone number to
    | create a Twitter, I was immediately banned. So instead I
    | purchased an account from a service with no issues. It's
    | become impossible for me at least to use large platforms
    | without assistance from an expert who runs bot farms to build
    | accounts that navigate the secret rules that govern bans.
 
    | cwkoss wrote:
    | Spam is a behavior, not a fundamental trait of the actor.
    | 
    | Would be interesting to make a service where spammers have to
    | do recaptcha-like spam flagging to get their account
    | unlocked.
 
      | fragmede wrote:
      | Which definition of spam are you operating under? I think
      | it _is_ a fundamental trait of the actor.
 
        | pessimizer wrote:
        | So you would expect a spammer to only ever post spam,
        | even on their own personal account? Or a spam emailer to
        | never send a personal email?
 
        | fragmede wrote:
        | I know sales bros who live their live by their ABCs -
        | always be closing, but that's besides the point. if the
        | person behind the spam bot one day wakes up and decides
        | to do turn over a new leaf and something else with their
        | life, they're not going to use the buyC1alis@vixagra.com
        | email address they use for sending spam as the basis for
        | their new persona. thus sending spam is inherit to the
        | buyC1alis@vixagra.com identity that we see - of course
        | there's a human being behind it, but as we'll never know
        | them in ant other context, that is who they are to us.
 
    | stouset wrote:
    | > and spaces that allow bad actors to grow
    | 
    | I believe that's GP's point! Any of us has the potential to
    | be the bad actor in some discussion that gets us irrationally
    | worked up. Maybe that chance is low for you or I, but it's
    | never totally zero.
    | 
    | And _even if_ the chance is zero for you or I specifically,
    | there 's no way for the site operators to a priori know that
    | fact or to be able to predict which users will suddenly
    | become bad actors and which discussions will trigger it.
 
    | pixl97 wrote:
    | I think the point is that anyone and/or everyone can be a bad
    | actor in the right circumstances, and moderations job is to
    | prevent those circumstances.
 
    | edgyquant wrote:
    | We have laws around mobs and peaceful protest for a reason.
    | Even the best people can become irrational as a group. The
    | groupmind is what we need controls for: not good and bad
    | people.
 
  | dgant wrote:
  | This is something Riot Games has spoken on, the observation
  | that ordinary participants can have a bad day here or there,
  | and that forgiving corrections can preserve their participation
  | while reducing future incidents.
 
    | synu wrote:
    | Did Riot eventually sort out the toxic community? If so that
    | would be amazing, and definitely relevant. I stopped playing
    | when it was still there, and it was a big part of the reason
    | I stopped.
 
      | BlargMcLarg wrote:
      | The only success I've seen in sorting out random vitriol is
      | cutting chat off entirely and minimizing methods of passive
      | aggressive communication. But Nintendo's online services
      | haven't exactly scaled to the typical MOBA size to see how
      | it actually works out
 
      | ketzo wrote:
      | Anecdotally: outright flame / chat hate is a bit better
      | than it used to be, but not much.
 
      | zahrc wrote:
      | I've been playing very very active from 2010 to 2014 and
      | since then on-off, sometimes skipping a season.
      | 
      | Recently picked it up again and I noticed that I didn't had
      | to use /mute all anymore. I've got all-chat disabled by
      | default so I've got no experience there, but overall I'd
      | say it has come a long way.
      | 
      | But I'd also say it depends which mode and MMR you are in.
      | I mostly play draft pick normals or ARAMs in which I both
      | have a lot of games played - I heard from a mate that chat
      | is unbearable in low level games.
 
  | gambler wrote:
  | It's not a mistake. It's a PR strategy. Social media companies
  | are training people to blame content and each other for the
  | effects that are produced by design, algorithms and moderation.
  | This reassigns blame away from things that those companies
  | control (but don't want to change) to things that aren't
  | considered "their fault".
 
  | stcredzero wrote:
  | The original post is paradoxical in the very way it talks about
  | social media being paradoxical.
  | 
  | He observes that social media moderation is about signal to
  | noise. Then he goes on about introducing off-topic noise. Then,
  | he comes to conclusions that seem to ignore his original
  | conclusion about it being a S/N problem.
  | 
  | Chiefly, he doesn't show how a "council of elders" is necessary
  | to solve S/N problems.
  | 
  | Strangely enough, Slashdot seems to have a system which worked
  | pretty well back in the day.
 
    | bombcar wrote:
    | I think the key is that no moderation can withstand _outside_
    | pressure. A community can be entirely consistent and happy
    | but the moment outside pressure is applied it folds or falls.
 
      | stcredzero wrote:
      | Slashdot moderation is largely done by the users
      | themselves, acting anonymously as "meta-moderators." I
      | think they were inspired by Plato's ideas around partially
      | amnesiac legislators who forget who they are while
      | legislating.
 
  | paul7986 wrote:
  | Having a a verified public Internet /Reputation ID system for
  | those who want to be bad or good publicly is one way!
  | 
  | All others are just trolls not backed up by their verified
  | public Internet / Reputation ID.
 
  | P_I_Staker wrote:
  | At the very least you could be susceptible overreacting because
  | of an emotionally charged issue. Eg. Reddit's boston marathon
  | bomber disaster, when they started trying to round up brown
  | people (actual perp "looked white")
  | 
  | Maybe that wouldn't be your crusade and maybe you would think
  | you were standing up for an oppressed minority. You get overly
  | emotional, and you could be prone to making some bad decisions.
  | 
  | People act substantially differently on reddit vs. hackernews;
  | honestly I have to admit to being guilty of it. Some of the
  | cool heads here are probably simultaneously engaged in
  | flamewars on reddit/twitter.
 
  | esotericimpl wrote:
  | Charge them $10 to create an account (anonymous, real, parody
  | whatever), then if they break a rule give them a warning, 2
  | rule breaks, a 24 hour posting suspension, 3 strikes and
  | permanently ban the account.
  | 
  | Let them reregister for $10.
  | 
  | Congrats, i just solved spam, bots, assholes and permanent line
  | steppers.
 
    | etchalon wrote:
    | You solved bots, but destroyed the product.
 
      | lancesells wrote:
      | I don't even know if it solved bots. Rich countries, rich
      | organizations, rich people could do a lot. $100M would buy
      | you 10M bots.
 
        | cvwright wrote:
        | I think the idea is that it shifts the incentives. Sure,
        | a rich nation state could buy tons of bot accounts at $10
        | a pop. But is that still the most rational path to their
        | goal? Probably not, because there are lots of other
        | things you can do for $100M.
 
      | trynewideas wrote:
      | I mean, who here remembers app.net? Love the Garry Tan
      | endorsement! https://web.archive.org/web/20120903182620/htt
      | ps://join.app....
      | 
      | EDIT: Lol Dalton's PART of YC now. Hey dude, why not pitch
      | it then
 
    | DeanWormer wrote:
    | This was the strategy at the SomethingAwful forums. They
    | seemed pretty well moderated, but definitely never hit the
    | scale of Reddit or Twitter.
 
      | v64 wrote:
      | Having posted there in its heyday, it made for an
      | interesting self-moderation dynamic for sure. Before I
      | posted something totally offbase that I knew I'd be
      | punished for, I had to think "is saying this stupid shit
      | really worth $10 to me?". Many times that was enough to get
      | me to pause (but sometimes you also can't help yourself and
      | it's well worth the price).
 
    | pfortuny wrote:
    | The problem is _the meaning of those rules_. Any rule looks
    | reasonable when it is written down. And after some time it
    | becomes a weapon.
    | 
    | For instance, the three deletions (forget the exact term)
    | rule in wikipedia. It is now a tool used by "the first to
    | write"...
 
    | dwater wrote:
    | This is how the SomethingAwful forums operated when they
    | started charging for accounts. Unfortunately it probably
    | wouldn't be useful as a test case because it was/is, at it's
    | core, a shitposting site.
 
      | rsync wrote:
      | I think metafilter still does this ?
 
      | nebqr wrote:
      | And twitter isn't?
 
    | pixl97 wrote:
    | Unless you generate more than $10 from the account. For
    | example in presidential election years in the US billions is
    | spent in advertising the elections. A few PACs would gladly
    | throw cash at astroturf movements on social media even at the
    | risk of being banned.
 
      | pessimizer wrote:
      | Sounds good to me. That would mean that your energy in
      | moderation would directly result in income. If superpacs
      | are willing to pay $3.33 a message, that's a money-spinner.
 
    | pclmulqdq wrote:
    | This kind of thing worked for a few forums that tried it
    | before FB/Twitter came around.
 
  | Covzire wrote:
  | Give the user exclusive control over what content they can see.
  | The platform should enforce legal actions against users only,
  | as far as bans are concerned.
  | 
  | Everything else, like being allowed to spam or post too
  | quickly, is a bug, and bugs should be addressed in the open.
 
  | visarga wrote:
  | > The reality is everyone, myself included, can be and will be
  | a bad actor.
  | 
  | Customised filters for anyone, but I am talking about filters
  | completely under the control of the user. Maybe running
  | locally. We can wrap ourselves in a bubble but better that than
  | having a bubble designed by others.
  | 
  | I think AI will make spam irrelevant over the next decade by
  | switching from searching and reading to prompting the bot. You
  | don't ever need to interface with the filth, you can have your
  | polite bot present the results however you please. It can be
  | your conversation partner and you get to control its biases as
  | well.
  | 
  | Internet <-> AI agent <-> Human
  | 
  | (the web browser of the future, the actual web browser runs in
  | a sandbox under the AI)
 
    | swayvil wrote:
    | I'll raise you a forum-trained AI spambot to defeat the AI
    | spamfilter. It'll be an entirely automated arms race.
 
  | Melatonic wrote:
  | Not true at all - everyone has the capacity for bad behaviour
  | in the right circumstances but most people are not, in my
  | opinion, there intentionally to be trolls.
  | 
  | There are the minority who love to be trolls and get any big
  | reaction out of people (positive or negative). Those people are
  | the problem. But they are also often very good at evading
  | moderation or laying in wait and toeing the line between
  | bannable offences and just every so slightly controversial
  | comments.
 
  | bnralt wrote:
  | Some people are much more likely to engage in bad behavior than
  | others. The thing is, people who engage in bad behavior are
  | also much more likely to be "whales," excessive turboposters
  | who have no life and spend all day on these sites.
  | 
  | Someone who has a balanced life, who spends time at work, with
  | family, in nature, only occasionally goes online, uses most of
  | their online time for edification, spends 30 minutes writing a
  | reply if they decide one is warranted - that type of person is
  | going to have a minuscule output compared to the whales. The
  | whales are always online, thoughtlessly writing responses and
  | upvoting without reading articles or comments. They have a
  | constant firehouse of output that dwarfs other users.
  | 
  | Worth reading "Most of What You Read on the Internet is Written
  | by Insane People"[1].
  | 
  | If you actually saw these people in real life, chances are
  | you'd avoid interacting with them. People seeing a short
  | interview with the top mod of antiwork almost destroyed that
  | sub (and lead to the mod stepping down). People say the
  | internet is a bad place because people act badly when they're
  | not face to face. That might be true to some extent, but we're
  | given online spaces where it's hard to avoid "bad actors" (or
  | people that engage in excessive bad behavior) the same way we
  | would in person.
  | 
  | And these sites need the whales, because they rely on a
  | constant stream of low quality content to keep people engaged.
  | There are simple fixes that could be done, like post limits and
  | vote limits, but sites aren't going to implement them. It's
  | easier to try to convince people that humanity is naturally
  | terrible than to admit they've created an environment that
  | enables - and even relies on - some of the most unbalanced
  | individuals.
  | 
  | [1]
  | https://www.reddit.com/r/slatestarcodex/comments/9rvroo/most...
 
  | ajmurmann wrote:
  | It sounds like a insurmountable problem. What makes this even
  | more interesting to me is that HN seems to have this working
  | pretty well. I wonder how much of it has to do with clear
  | guidelines of what should be valued and what shouldn't and
  | having a community that buys in to that. For example one learns
  | quickly that Reddit-style humor comments are frowned upon
  | because the community enforces it with downvotes and frequently
  | explanations of etiquette.
 
    | etchalon wrote:
    | I suspect HN succeeds due to heavy moderation, explicit
    | community guidelines and a narrow topic set.
 
      | blep_ wrote:
      | Some areas of reddit do similar things with similar
      | results. AskHistorians and AskScience are the first two to
      | come to mind.
      | 
      | This may be a lot easier in places where there's an
      | explicit _point_ to discussion beyond the discussion itself
      | - StackOverflow is another non-Reddit example. It 's easier
      | to tell people their behavior is unconstructive when it's
      | clearly not contributing to the goal. HN's thing may just
      | be to declare a particular type of conversation to be the
      | goal.
 
    | vkou wrote:
    | HN works very well, because it's about as far from free
    | speech as you can get on the internet, short of dang
    | personally approving every post.
 
      | swayvil wrote:
      | It's proof of the old adage : the best possible government
      | is a benign dictator.
 
      | theGnuMe wrote:
      | I think most posts are short lived so they drop off quickly
      | and people move on to new content. I think a lot of folks
      | miss a lot of activity that way. I know I miss a bunch. And
      | if you miss the zeitgeist it doesn't matter what you say
      | cause nobody will reply.
      | 
      | The twitter retweet constantly amplifies and the tweets are
      | centered around an account vs a post.
      | 
      | Reddit should behave similarly but I think subreddit topics
      | stick longer.
 
        | luckylion wrote:
        | Very good point about the "fog of war". If HN had a
        | reply-notification feature, it would probably look
        | differently. Every now and then someone builds a
        | notification feature as an external service. I wonder if
        | you can measure change in the behavior of people before
        | and after they've started using it?
        | 
        | Of course, that also soft-forces everyone to move on.
        | Once a thread is a day or two old, you can still reply,
        | but the person you've replied to will probably not read
        | it.
 
        | rjbwork wrote:
        | There's also the fact that there's no alerts about people
        | replying to you or commenting on your posts. You have to
        | explicitly go into your profile, click comments, and
        | _then_ you can see _if_ anyone has said anything to you.
        | 
        | This drastically increases time between messages on a
        | topic, lets people cool off, and lets a topic naturally
        | die down.
 
      | prox wrote:
      | What kind of free speech is not allowed? What can't you say
      | right now that you feel should?
 
        | ajmurmann wrote:
        | Category 1 from Yishan's thread, spam, obviously isn't
        | allowed. But also thinking about house general framework
        | of it all coming down to signal vs noise, most "noise"
        | gets heavily punished on here. Reddit-style jokes
        | frequently end in the light greys or even dead. I had my
        | account shadow-banned over a decade ago because I made a
        | penis joke and thought people didn't get the joke.
 
        | kevin_thibedeau wrote:
        | Free speech doesn't mean you can say whatever, wherever,
        | without any repercussions. It solely means the government
        | can't restrict your expression. On a private platform you
        | abide their rules.
 
        | michaelt wrote:
        | Where are the people arguing about Donald Trump? Where
        | are the people promoting dodgy cryptocurrencies? Where
        | are the people arguing about fighting duck-sized horses?
        | Where's the Ask HN asking for TV show recommendations?
 
    | slg wrote:
    | If we follow the logic of Yishan's thread, HN frowns upon and
    | largely doesn't allow discussion that would fall into group 3
    | which removes most of the grounds for accusations of
    | political and other biases in the moderation. As Yishan says,
    | no one really cares about banning groups 1 and 2, so no one
    | objects to when that is done here.
    | 
    | Plus scale is a huge factor. Automated moderation can have
    | its problems. Human moderation is expensive and hard to keep
    | consistent if there are large teams of individuals that can't
    | coordinate on everything. HN's size and its lack of desire
    | for profit allow for a very small human moderation team that
    | leads to consistency because it is always the same people
    | making the decisions.
 
    | rsync wrote:
    | I have a suspicion that the medium is the message at HN:
    | 
    | No pictures and no avatars.
    | 
    | I wonder how much bad behavior is weeded out by the interface
    | itself ?
    | 
    | A lot, I suspect ...
 
      | jdp23 wrote:
      | Nope. There's been abuse in text-only environments online
      | since forever. And lots of people have left (or rarely post
      | on) HN because of complaints about the enviroment here.
 
      | ChainOfFools wrote:
      | > No pictures and no avatars
      | 
      | This is essentially moderation rule #0. it is unwritten,
      | enforced before violation can occur, and generates zero
      | complaints because it filters complainers out of the user
      | pool from the start.
 
        | luckylion wrote:
        | The no-avatars rule also takes away some of the
        | personalization aspect. If you set your account up with
        | your nickname, your fancy unique profile picture and your
        | favorite quote in the signature, and someone says you're
        | wrong, you're much more invested because you've tied some
        | of your identity to the account.
        | 
        | If you've just arrived on the site, have been given a
        | random name and someone says you're wrong, what do you
        | care? You're not attached to that account at all, it's
        | not "you", it's just a random account on a random
        | website.
        | 
        | I thought that was an interesting point on 4chan (and
        | probably other sites before them), that your identity was
        | set per thread (iirc they only later introduced the
        | ability to have permanent accounts). That removes the
        | possibility of you becoming attached to the random name.
 
  | dfxm12 wrote:
  | _How do you build and run a "social media" product when the
  | very act of letting anyone respond to anyone with anything is
  | itself the fundamental problem?_
  | 
  | This isn't the problem as much as giving bad actors tools to
  | enhance their reach. Bad actors can pay to get a wider reach or
  | get/abuse a mark of authority, like a special tag on their
  | handle, getting highlighted in a special place within the app,
  | gaming the algorithm that promotes some content, etc. Most of
  | these tools are built into the platform. Some though, like sock
  | puppets, can be detected but aren't necessarily built in
  | functionality.
 
  | bambax wrote:
  | You're confusing _bad actors_ with _bad behavior_. Bad behavior
  | is something good people do from time to time because they get
  | really worked up about a specific topic or two. Bad actors are
  | people who act bad all the time. There may be some of those but
  | they 're not the majority by far (and yes, sometimes normal
  | people turn into bad actors because they get upset about a
  | given thing that they can't talk about anything else anymore).
  | 
  | OP's argument is that you can moderate content based on
  | behavior, in order to bring the heat down, and the signal to
  | noise ratio up. I think it's an interesting point: it's neither
  | the tools that need moderating, nor the people, but
  | _conversations_ (one by one).
 
    | rlucas wrote:
    | ++
    | 
    | A giant amount of social quandaries melt away when you
    | realize:
    | 
    | "Good guys" and "Bad guys" is not a matter of identity, it's
    | a matter of activity.
    | 
    | You aren't a "Good guy" because of _who you are_ , but
    | because of _what you do_.
    | 
    | There are vanishingly few people who as a matter of identity
    | are reliably and permanently one way or another.
 
    | dang wrote:
    | I think that's right. One benefit this has: if you can make
    | the moderation about behavior (I prefer the word effects [1])
    | rather than about the person, then you have a chance to
    | persuade them to behave differently. Some people, maybe even
    | most, adjust their behavior in response to feedback. Over
    | time, this can compound into community-level effects (culture
    | etc.) - that's the hope, anyhow. I _think_ I 've seen such
    | changes on HN but the community/culture changes so slowly
    | that one can easily deceive oneself. There's no question it
    | happens at the individual user level, at least some of the
    | time.
    | 
    | Conversely, if you make the moderation about the person
    | (being a bad actor etc.) then the only way they can agree
    | with you is by regarding themselves badly. That's a weak
    | position for persuasion! It almost compels them to resist
    | you.
    | 
    | I try to use depersonalized language for this reason. Instead
    | of saying " _you_ " did this (yeah that's right, YOU), I'll
    | tell someone that their _account_ is doing something, or that
    | their _comment_ is a certain way. This creates distance
    | between their account or their comment and _them_ , which
    | leaves them freer to be receptive and to change.
    | 
    | Someone will point out or link to cases where I did the exact
    | opposite of this, and they'll be right. It's hard to do
    | consistently. Our emotional programming points the other way,
    | which is what makes this stuff hard and so dependent on self-
    | awareness, which is the scarcest thing and not easily added
    | to [2].
    | 
    | [1] https://news.ycombinator.com/item?id=33454968
    | 
    | [2] https://news.ycombinator.com/item?id=33448079
 
      | bombcar wrote:
      | The other tricky thing is a bad actor will work to stay
      | just this side of the rules while still causing damage and
      | destruction to the forum itself.
 
        | dang wrote:
        | Yes. But in our experience to date, this is less common
        | than people say it is, and there are strategies for
        | dealing with it. One such strategy is https://hn.algolia.
        | com/?sort=byDate&dateRange=all&type=comme... (sorry I
        | don't have time to explain this, as I'm about to go
        | offline - but the key word is 'primarily'.) No strategy
        | works in all cases though.
 
      | jimkleiber wrote:
      | > I try to use depersonalized language for this reason.
      | Instead of saying "you" did this (yeah that's right, YOU),
      | I'll tell someone that their account is doing something, or
      | that their comment is a certain way. This creates distance
      | between their account or their comment and them, which
      | leaves them freer to be receptive and to change.
      | 
      | I feel quite excited to read that you, dang, moderating HN,
      | use a similar technique that I use for myself and try to
      | teach others. Someone told my good friend the other day
      | that he wasn't being a very good friend to me, and I told
      | him that he may do things that piss me off, annoy me,
      | confuse me, or whatever, but he will always be a good
      | friend to me. I once told an Uber driver who told me he
      | just got out of jail and was a bad man, I said, "No, you're
      | a good man who probably did a bad thing."
      | 
      | Thank you for your write-up.
 
      | user3939382 wrote:
      | > persuade the user to behave differently
      | 
      | That scares me. Today's norms are tomorrow's taboos. The
      | dangers of conforming and shaping everyone into the least
      | controversial opinions and topics are self evident. It's an
      | issue on this very forum. "Go elsewhere" doesn't solve the
      | problem because that policy still contributes to a self-
      | perpetuating feedback loop that amplifies norms, which
      | often happen to be corrupt and related to the interests of
      | big (corrupt) commercial and political powers.
 
        | dang wrote:
        | I don't mean persuade them out of their opinions on
        | $topic! I mean persuade them to express their opinions in
        | a thoughtful, curious way that doesn't break the site
        | guidelines -
        | https://news.ycombinator.com/newsguidelines.html.
 
        | user3939382 wrote:
        | Sufficiently controversial opinions are flagged,
        | downvoted til dead/hidden, or associated users shadow
        | banned. HN's policies and voting system, both de facto
        | and de jure, discourage controversial opinions and reward
        | popular, conformist opinions.
        | 
        | That's not to pick on HN, since this is a common problem.
        | Neither do I have a silver bullet solution, but the issue
        | remains, and it's a huge issue. Evolution of norms, for
        | better or worse, is suppressed to the extent that big
        | communication platforms suppress controversy. The whole
        | concept of post and comment votes does this by
        | definition.
 
        | dang wrote:
        | That's true to an extent (and so is what ativzzz says, so
        | you're both right). But the reasons for what you're
        | talking about are much misunderstood. Yishan does a good
        | job of going into some of them in the OP, by the way.
        | 
        | People always reach immediately for the conclusion that
        | their controversial-opinion comments are getting
        | moderated because people dislike their controversial
        | opinion--either because of groupthink in the community or
        | because the admins are hostile to their views. Most of
        | the time, though, they've larded their comments pre-
        | emptively with some sort of hostility, snark, name-
        | calling, or other aggression--no doubt because they
        | expect to be opposed and want to make it clear they
        | already know that, don't care what the sheeple think, and
        | so on.
        | 
        | The way the group and/or the admins respond to those
        | comments is often a product of those secondary mixins.
        | Forgive the gross analogy, but it's as if someone serves
        | a shit milkshake and when it's rejected, say, "you just
        | hate dairy products" or "this community is so biased
        | against milkshakes".
        | 
        | If you start instead from the principle that the value of
        | a comment is the expected value of the subthread it forms
        | the root of [1], then a commenter is responsible for the
        | effects of their comments [2] - at least the predictable
        | ones. From that it follows that there's a greater burden
        | on the commenter who's expressing a contrarian view [3].
        | The more contrarian the view--the further it falls
        | outside the community's tolerance--the more
        | responsibility that commenter has for not triggering
        | degenerative effects like flamewars.
        | 
        | This may be counterintuitive, because we're used to
        | thinking in terms of atomic individual responsibility,
        | but it's a model that actually works. Threads are
        | molecules, not atoms--they're a cocreation, like one of
        | those drawing games where each person fills in part of a
        | shared picture [4], or like a dance--people respond to
        | the other's movements. A good dancer takes the others
        | into account.
        | 
        | It may be unfair that the one with a contrarian view is
        | more responsible for what happens--especially because
        | they're already under greater pressure than the one whose
        | views agree with the surround. But fair or not, it's the
        | way communication works. If you're trying to deliver
        | challenging information to someone, you have to take that
        | person into account--you have to regulate what you say by
        | what the listener is capable to hear and to tolerate.
        | Otherwise you're predictably going to dysregulate them
        | and ruin the conversation.
        | 
        | Contrarian commenters usually do the opposite of this--
        | they express their contrarian opinion in a deliberately
        | aggressive and uncompromising way, probably because (I'm
        | repeating myself sorry) they expect to be rejected
        | anyhow, and it's safer to be inside the armor of "you
        | people can't handle the truth!" than it is to really
        | communicate, i.e. to connect and relate.
        | 
        | This model is the last thing that most contrarian-opinion
        | commenters want to adopt, because it's hard and risky,
        | and because usually they have pre-existing hurt feelings
        | from being battered repeatedly with majoritarian opinions
        | already (especially the case when identity is at issue,
        | such as being from a minority population along whatever
        | axis). But it's a model that actually works and it's by
        | far the best solution I know of to the problem of
        | unconventional opinions in forums.
        | 
        | Are there some opinions which are so far beyond the
        | community's tolerance that any mention in any form will
        | immediately blow up the thread, making the above model
        | impossible? Yes.
        | 
        | [1] https://hn.algolia.com/?dateRange=all&page=0&prefix=t
        | rue&sor...
        | 
        | [2] https://hn.algolia.com/?dateRange=all&page=0&prefix=t
        | rue&sor...
        | 
        | [3] https://hn.algolia.com/?dateRange=all&page=0&prefix=t
        | rue&que...
        | 
        | [4[ https://news.ycombinator.com/item?id=6813226
 
        | ativzzz wrote:
        | Completely disagree about HN. Controversial topics that
        | are thought out, well formed, and argued with good intent
        | are generally good sources of discussion.
        | 
        | Most of the time though, people arguing controversial
        | topics phrase them so poorly or include heavy handed
        | emotions so that their arguments have no shot of being
        | fairly interpreted by anyone else.
 
      | pbhjpbhj wrote:
      | You're doing a good job dang.
      | 
      | ... kinda wondering if this is the sort of OT post we're
      | supposed to avoid, it would be class if you chastised me
      | for it. But anyway, glad you're here to keep us in check
      | and steer the community so well.
 
        | dang wrote:
        | For stuff like that I go by what pg wrote many years ago:
        | https://news.ycombinator.com/newswelcome.html
        | 
        |  _Empty comments can be ok if they 're positive. There's
        | nothing wrong with submitting a comment saying just
        | "Thanks." What we especially discourage are comments that
        | are empty and negative--comments that are mere name-
        | calling._
        | 
        | It's true that empty positive comments don't add much
        | information but they have a different healthy role
        | (assuming they aren't promotional)
 
      | ar_lan wrote:
      | You definitely hit the nail on the head.
      | 
      | If someone points out a specific action I did that
      | can/should be improved upon (and especially if they can
      | tell me why it was "bad" in the first place), I'm far more
      | likely to accept that, attempt to learn from it, and move
      | on. As in real life, I might still be heated in the moment,
      | but I'll usually remember that when similar cues strike
      | again.
      | 
      | But if moderation hints at something being wrong with my
      | identity or just me fundamentally, then that points to
      | something that _can't be changed_. If that's the case, I
      | _know they are wrong_ and simply won't respect that they
      | know how to moderate anything at all, because their
      | judgment is objectively incorrect.
      | 
      | Practically at work, this has actually been a good policy
      | you described when I think about bugs and code reviews.
      | 
      | > "@ar_lan broke `main` with this CLN. Reverting."
      | 
      | is a pretty sure-fire way to make me defend my change and
      | believe you are wrong. My inclination, for better or worse,
      | will be to dispute the accusation directly and clear my
      | name (probably some irrational fear that creating a bug
      | will go on a list of reasons to fire me).
      | 
      | But when I'm approached with:
      | 
      | > "Hey, @ar_lan. It looks like pipeline X failed this test
      | after this CLN. We've automatically reverted the commit.
      | Could you please take a second look and re-submit with a
      | verification of the test passing?"
      | 
      | I'm almost never defensive about it, and I almost always go
      | right ahead to reproducing the failure and working on the
      | fix.
      | 
      | The first message conveys to me that I (personally) am the
      | reason `main` is broken. The second conveys that it was my
      | CLN that was problematic, but fixable.
      | 
      | Both messages are taken directly from my companies Slack
      | (ommitting some minor details, of course), for reference.
 
      | camgunz wrote:
      | I think your moderation has made me better at HN, and
      | consequently I'm better in real life. Actively thinking
      | about how to better communicate and create environments
      | where everyone is getting something positive out of the
      | interaction is something I maybe started at HN, and then
      | took into the real world. I think community has a lot to do
      | with it, like "be the change you want to see".
      | 
      | But to your point, yeah my current company has feedback
      | guidelines that are pretty similar: criticize the work, not
      | the worker, and it super works. You realize that action
      | isn't aligned with who you want to be or think you are, and
      | you stop behaving that way. I mean, it's worked on me and
      | I've seen it work on others, for sure.
 
    | mypalmike wrote:
    | I can "behave well" and still be a bad actor in that I'm
    | constantly spreading dangerous disinformation. That
    | disinformation looks like signal by any metadata analysis.
 
      | bambax wrote:
      | Yes, that's probably the limit of the pure behavioral
      | analysis, esp. if one is sincere. If they're insincere it
      | will probably look like spam; but if somebody truly
      | believes crazy theories and is casually pushing them (vs
      | promoting them aggressively and exclusively), that's
      | probably harder to spot.
 
    | afiori wrote:
    | I think you agree with the parent.
    | 
    | They pointed out that everybody can be a bad actor and you
    | will not find a way to get better users.
 
    | whoopdedo wrote:
    | And bad behavior gets rewarded with engagement. We learned
    | this from "reality television" where the more conflict there
    | was among a group of people the more popular that show was.
    | (Leading to producers abandoning the purity of being
    | unscripted in the pursuit of better ratings.) A popular
    | pastime on Reddit is posting someone behaving badly (whether
    | on another site, a subreddit, or in a live video) for the
    | purpose of mocking them.
    | 
    | When the organizational goal is to increase engagement, which
    | will be the case wherever there are advertisers, inevitably
    | bad behavior will grow more frequent than good behavior.
    | Attempts to moderate toward good behavior will be abandoned
    | in favor of better metrics. Or the site will stagnate under
    | the weight of the new rules.
    | 
    | In this I'm in disagreement with Yishan because in those
    | posts I read that engagement feedback is a characteristic of
    | old media (newspapers, television) and social media tries to
    | avoid that. The OP seems to be saying that online moderation
    | is an attempt to minimize controversial engagement because
    | platforms don't like that. I don't believe it. I think social
    | media loves controversial engagement just as much as the old-
    | school "if it bleeds, it leads" journalists from television
    | and newspapers. What they don't want is the (quote/unquote)
    | wrong kind of controversies. Which is to say, what defines
    | bad behavior is not universally agreed upon. The threshold
    | for what constitutes bad behavior will be different depending
    | on who's doing the moderating. As a result the content seen
    | will be influenced by the moderation, even if said moderation
    | is being done in a content-neutral way.
    | 
    | And I just now realize that I've taken a long trip around to
    | come to the conclusion that the medium is the message. I
    | guess we can now say the moderation is the message.
 
    | jonny_eh wrote:
    | > Bad actors are people who act bad all the time
    | 
    | I'd argue that bad actors are people that behave badly "on
    | purpose". Their _goals_ are different than the normal actor.
    | Bad actors want to upset or scare people. Normal actors want
    | to connect with, learn from, or persuade others.
 
| paradite wrote:
| I recently started my own Discord server and had my first
| experience in content moderation. The demographics is mostly
| teenagers. Some have mental health issues.
| 
| It was the hardest thing ever.
| 
| In first incident I chose to ignore a certain user being targeted
| by others for posting repeated messages. The person left a very
| angry message and left.
| 
| Comes the second incident, I thought I learnt my lesson. Once a
| user is targeted, I tried to stop others from targeting the
| person. But this time the people who targeted the person wrote
| angry messages and left.
| 
| Someone asked a dumb question, I replied in good faith. The
| conversation goes on and on and becomes weirder and weirder,
| until the person said "You shouldn't have replied me.", and left.
| 
| Honestly I am just counting on luck at this time that I can keep
| it running.
 
  | derefr wrote:
  | > In first incident I chose to ignore a certain user being
  | targeted by others for posting repeated messages. The person
  | left a very angry message and left.
  | 
  | > Comes the second incident, I thought I learnt my lesson. Once
  | a user is targeted, I tried to stop others from targeting the
  | person. But this time the people who targeted the person wrote
  | angry messages and left.
  | 
  | Makes me think that moderators should have the arbitrational
  | power to take two people or groups, and (explicitly, with
  | notice to both people/groups) make each person/group's public
  | posts invisible to the other person/group. Like a cross between
  | the old Usenet ignore lists, and restraining orders, but
  | externally-imposed without either party actively seeking it
  | out.
 
  | watwut wrote:
  | Imo, some people leaving is not necessary bad thing. Like, some
  | people are looking for someone to bully. Either you allow them
  | bully or they leave. The choice determines overall culture of
  | you community.
  | 
  | And sometimes people are looking for a fight and will search it
  | until they find it ... and then leave.
 
    | Goronmon wrote:
    | _And sometimes people are looking for a fight and will search
    | it until they find it ... and then leave._
    | 
    | I've found the more likely result is that people looking for
    | a fight will find it, and then stay because they've found a
    | target and an audience. Even if the audience is against them
    | (and especially so if moderators are against them), for some
    | people that just feeds their needs even more.
 
  | thepasswordis wrote:
  | How old are you? An adult running a discord server for mentally
  | ill teenagers seems like a cautionary tale from the 1990s about
  | chatrooms.
 
    | paradite wrote:
    | I'm afraid I'm too young to understand that reference or
    | context around chatrooms.
    | 
    | Anyway, the Discord server is purely for business and
    | professional purposes. And I use the same username everywhere
    | including Discord, so it's pretty easy to verify my identity.
 
      | Tenal wrote:
 
    | drekipus wrote:
    | Its in vogue today.
 
    | skissane wrote:
    | > An adult running a discord server for mentally ill
    | teenagers seems like a cautionary tale from the 1990s about
    | chatrooms
    | 
    | It sounds like a potential setup for exploitation, grooming,
    | cult recruitment, etc. (Not saying the grandparent is doing
    | this, for all I know their intentions are entirely above
    | board-but other people out there likely are doing it for
    | these kinds of reasons.)
 
      | [deleted]
 
      | peruvian wrote:
      | Discord is already considered a groomer hotspot, at least
      | in joking. You can join servers based on interests alone
      | and find yourself in a server with very young people.
 
    | TulliusCicero wrote:
    | I doubt it's explicitly for mentally ill teenagers. It could
    | be, say, a video game discord, and so the demographics are
    | mostly teens who play the game, and obviously some subset
    | will be mentally ill.
 
      | strken wrote:
      | It's probably something like this. I'm interested in a
      | specific videogame and have bounced around a lot of
      | discords trying to find one where most of the members are
      | older. We still have some under-18s (including one guy's
      | son), but they're in the minority, and that makes
      | everything easier to moderate. We can just ban (or temp-
      | ban) anyone who's bringing the vibe down and know that the
      | rest will understand and keep the peace.
      | 
      | Teens don't have as much experience with communities going
      | to shit, or with spaces like the workplace where you're
      | collectively responsible for the smooth running of the
      | group. They're hot-headed and can cause one bad experience
      | to snowball where an adult might forgive and forget.
      | 
      | About the only thing that makes mentally healthy adults
      | hard to moderate is when they get drunk or high and do
      | stupid stuff because they've stopped worrying about
      | consequences.
 
        | MichaelCollins wrote:
        | > _Teens don 't have as much experience with communities
        | going to shit, or with spaces like the workplace where
        | you're collectively responsible for the smooth running of
        | the group. They're hot-headed and can cause one bad
        | experience to snowball where an adult might forgive and
        | forget._
        | 
        | Some people, not just teens of course, feel utterly
        | compelled to go tit-for-tat, to retaliate in kind. Even
        | if you can get them to cool down and back off for a
        | while, and have a civil conversation with you about
        | disengaging, they may tell you that they're going to
        | retaliate against the other person anyway at a later
        | date, in cold blooded revenge, because they _have to_.
        | That necessity seems to be an inescapable reality for
        | such people. They feel they have _no choice_ but to
        | retaliate.
        | 
        | When two such people encounter each other and an accident
        | is mispercieved as an offense, what follows is
        | essentially a blood feud. An unbreakable cycle of
        | retaliation after retaliation. Even if you can get to the
        | bottom of the original conflict, they'll continue
        | retaliating against each other for the later acts of
        | retaliation. The only way to stop it is to ban one if not
        | both of them. Moderation sucks, never let somebody talk
        | you into it.
 
    | pr0zac wrote:
    | My interpretation was he ran a discord server for a topic
    | who's demographics happened to include a large number of
    | teenagers and folks with mental illness thus unintentionally
    | resulting in a discord containing a lot of them, not that he
    | was specifically running a discord server targeting mentally
    | ill teens.
 
  | bmitc wrote:
  | I think all this just revolves around humans being generally
  | insane and emotionally unstable. Technology just taps into
  | this, exposes it, and connects it to others.
 
  | themitigating wrote:
  | Wow, and now we all learned that nothing should be censored
  | thanks to this definitely real situation where the same outcome
  | occurred when you censored both the victim and perpetrator
 
  | DoItToMe81 wrote:
  | Mental illness or not, your interactions with users in a
  | service with a block button are all voluntary. Unless someone
  | is going out of their own way to drag drama out of Discord, or
  | god forbid, into real life, it tends to be best to just let it
  | happen, as they are entirely willingly participating in it and
  | the escape is just a button away.
 
    | watwut wrote:
    | Community defined by the most aggressive people that come in
    | tend to be the one where everyone else voluntarily leaves,
    | cause leaving is much better for them.
 
    | TulliusCicero wrote:
    | I see this a fair amount, and yeah, "just let people block
    | others" is really terrible moderation advice.
    | 
    | Besides the very reasonable expectation almost everyone has
    | that assholes will be banned, the inevitable result of not
    | banning assholes is that you get more and more assholes,
    | because their behavior will chase away regular users. Even
    | some regular users may start acting more like assholes,
    | because what do you do when someone is super combative, aside
    | from possibly leaving? You become combative right back, to
    | fight back.
 
  | beezlebroxxxxxx wrote:
  | IME, places (or forums, or social networks, etc.) with good
  | moderation tend to fall into 2 camps of putting that into play:
  | 
  | 1. The very hands-off approach style that relies on the subject
  | matter of the discussion/topic of interest naturally weeding
  | out "normies" and "trolls" with moderation happening "behind
  | the curtain";
  | 
  | 2. The very hands-on approach that relies on explicit clear
  | rules and no qualms about acting on those rules, so moderation
  | actions are referred directly back to the specific rule broken
  | and in plain sight.
  | 
  | Camp 1 begins to degrade as more people use your venue; camp 2
  | degrades as the venue turns over to debate about the rules
  | themselves rather than the topic of interest that was the whole
  | point of the venue itself (for example, this is very common in
  | a number of subreddits where break-off subreddits usually form
  | in direct response to a certain rule or the enforcement of a
  | particular rule).
 
    | derefr wrote:
    | Camp 2 works fine in perpetuity _if_ the community is built
    | as a cult of personality around a central authority figure;
    | where the authority figure is also the moderator (or, if
    | there are other moderators, their authority is delegated to
    | them by the authority figure, and they can always refer
    | arbitration back to the authority figure); where the clear
    | rules are understood to be _descriptive_ of the authority 's
    | decision-tree, rather than _prescriptive_ of it -- i.e.
    | "this is how I make a decision; if I make a decision that
    | doesn't cleanly fit this workflow, I won't be constrained by
    | the workflow, but I will try to change the workflow such that
    | it has a case for what I decided."
 
    | wwweston wrote:
    | Is people leaving and founding a different forum with
    | different rules really a failure/degradation?
 
      | cloverich wrote:
      | It would be cool if such forks were transparent on the
      | original forum / subreddit, and if they also forked on
      | specific rules. I.e. like a diff with rule 5 crossed out /
      | changed / new rule added, etc.
 
        | wutbrodo wrote:
        | I've seen an example of this. The fork is less active
        | than the original, but I wouldn't call it a failure.
        | Rather, it was a successful experiment with a negative
        | result. The original forum was the most high-quality
        | discussion forum I've ever experienced in my life, so
        | this wasn't quite a generalizable experiment.
 
  | krippe wrote:
  | Someone asked a dumb question, I replied in good faith. The
  | conversation goes on and on and becomes weirder and weirder,
  | until the person said "You shouldn't have replied me.", and
  | left.
  | 
  | Haha wtf, why would they do that?
 
  | TulliusCicero wrote:
  | I'm confused, do you think some individual leaving is a failure
  | state? Realistically I don't think you can avoid banning or
  | pissing some people off as a moderator, at least in most cases.
  | 
  | There's a lot of people whose behavior on internet message
  | boards/chat groups can be succinctly summarized as, "they're an
  | asshole." Now maybe IRL they're a perfectly fine person, but
  | for whatever reason they just engage like an disingenuous jerk
  | on the internet, and the latter case is what's relevant to you
  | as a moderator. In some cases a warning or talking-to will
  | suffice for people to change how they engage, but often times
  | it won't, they're just dead set on some toxic behavior.
 
    | shepherdjerred wrote:
    | > I'm confused, do you think some individual leaving is a
    | failure state?
    | 
    | When you are trying to grow something, them leaving is a
    | failure.
    | 
    | I ran a Minecraft server for many years when I was in high
    | school. It's very hard to strike a balance of:
    | 
    | 1. Having players
    | 
    | 2. Giving those players a positive experience (banning
    | abusers)
    | 
    | 3. Stepping in only when necessary
    | 
    | Every player that I banned meant I lost some of my player
    | base. Some players in particular would cause an entire group
    | to leave. Of course, plenty of players have alternate
    | accounts and would just log onto one of those.
 
      | TulliusCicero wrote:
      | I think it _can_ be a failure state, certainly, but
      | sometimes it 's unavoidable, and banning someone can also
      | mean more people in the community, rather than less.
      | 
      | Would HN be bigger if it had always had looser moderation
      | that involved less banning of people? I'm guessing not.
      | 
      | edit: I guess what I was thinking was that often in a
      | community conflict where one party is 'targeted' by another
      | party, banning one of those parties is inevitable. Not
      | always, but often people just cannot be turned away from
      | doing some toxic thing, they feel that they're justified in
      | some way and would rather leave/get banned than stop.
 
    | whatshisface wrote:
    | The person leaving is the least bad part of what happened in
    | the OP's example, try reading this again?:
    | 
    | > _In first incident I chose to ignore a certain user being
    | targeted by others for posting repeated messages. The person
    | left a very angry message and left._
 
      | TulliusCicero wrote:
      | They have three examples, and all of them ended with the
      | person leaving; it just sounded to me like they were
      | implying that the person leaving represented a failure on
      | their part as a moderator. That, had they moderated better,
      | they could've prevented people leaving.
 
        | whatshisface wrote:
        | Each of the examples had something bad happen in the
        | lead-up to the person leaving.
 
        | TulliusCicero wrote:
        | Yes, and? I honestly can't tell what you're getting at
        | here.
 
        | whatshisface wrote:
        | That the bad thing they were talking about was the bad
        | stuff leading up to the person leaving.
 
        | TulliusCicero wrote:
        | That was bad yes, but it sounds like they feel that the
        | outcome each time of someone leaving (and possibly with
        | an angry message) was also bad, and indicative that they
        | handled the situation incorrectly.
 
  | lovehashbrowns wrote:
  | Discord is particularly tough, depending on the type of
  | community. I very briefly moderated a smaller community for a
  | video game, and goodness was that awful. There was some
  | exceptionally egregious behavior, which ultimately made me
  | quit, but even things like small cliques. Any action, perceived
  | or otherwise, taken against a "popular" member of that clique
  | would immediately cause chaos as people would begin taking
  | sides and forming even stronger cliques.
  | 
  | One of the exceptionally egregious things that made me quit
  | happened in a voice call where someone was screensharing
  | something deplorable (sexually explicit content with someone
  | that wasn't consenting to the screensharing). I wouldn't have
  | even known it happened except that someone in the voice call
  | wasn't using their microphone, so I was able to piece together
  | what happened from them typing in the voice chat text channel.
  | I can't imagine the horror of moderating a larger community
  | where various voice calls are happening at all times of the
  | day.
 
    | ChainOfFools wrote:
    | flamebait directed at specific groups: cliquebait
    | 
    | /s
 
  | [deleted]
 
| ojosilva wrote:
| There are so many tangible vectors in content! It makes me feel
| like moderation is a doable, albeit hard to automate, task:
| 
| - substantiated / unsubstantiated - extreme / moderate -
| controversial / anodyne - fact / fun / fiction - legal / unlawful
| - mainstream / niche - commercial / free - individual /
| collective - safe / unsafe - science / belief - vicious / humane
| - blunt / tactful - etc. etc.
| 
| Maybe I'm too techno-utopic, but can't we model AI to detect how
| these vectors combine to configure moderation?
| 
| Ex: Ten years ago masks were _niche_ , therefore
| _unsubstantiated_ news on the drawbacks of wearing masks were
| still considered _safe_ because very few people were paying
| attention and /or could harm themselves, so that it was not
| _controversial_ and did not require moderation. Post-covid, the
| vector values changed, questionable content about masks could be
| flagged for moderation with some intensity indexes, user-
| discretion-advised messages and /or links to rebuttals if
| applicable.
| 
| Let the model and results be transparent and reviewable, and,
| most important, editorial. I think the greatest mistake of
| moderated social networks is that many people (and the network
| themselves) think that these internet businesses are not
| "editorial", but they are not very different from regular news
| sources when it comes to editorial lines.
 
  | raxxorraxor wrote:
  | Not a good idea. Your example already has flaws. An AI could
  | perform on a larger scale, but the result would be worse.
  | Probably far worse.
  | 
  | I specifically don't want any editor for online content. Just
  | don't make it boring or worse turn everything into
  | astroturfing. Masks are a good example already.
 
  | pixl97 wrote:
  | >Maybe I'm too techno-utopic,
  | 
  | https://xkcd.com/1425/
  | 
  | I personally believe this won't be a solvable problem, or at
  | least the problem will grow a long tail. One example would be
  | hate groups co-opting the language of the victim group in an
  | intentional manner. Then as the hate group is moderated for
  | their behaviors, the victim group is caught up in the action by
  | intentional user reporting for similar language.
  | 
  | It's a difficult problem to deal with as at least some portion
  | of your userbase will be adversarial and use external signaling
  | and crowd sourced information to cause issues with your
  | moderation system.
 
| fleddr wrote:
| In the real world, when you're unhinged, annoying,
| intrusive...you face almost immediate negative consequences. On
| social media, you're rewarded with engagement. Social media
| owners "moderate" behavior that maximizes the engagement they
| depend on, which makes it somewhat of a paradox.
| 
| It would be similar to a newspaper "moderating" their journalists
| to bring news that is balanced, accurate, fact-checked, as
| neutral as possible, with no bias to the positive or negative.
| This wouldn't sell any actual news papers.
| 
| Similarly, nobody would watch a movie where the characters are
| perfectly happy. Even cartoons need villains.
| 
| All these types of media have exploited our psychological draw to
| the unusual, which is typically the negative. This attention hack
| is a skill evolved to survive, but now triggered all day long for
| clicks.
| 
| Can't be solved? More like unwilling to solve. Allow me to clean
| up Twitter:
| 
| - Close the API for posting replies. You can have your weather
| bot post updates to your weather account, but you shouldn't be
| able to instant-post a reply to another account's tweet.
| 
| - Remove the retweet and quote tweet buttons. This is how things
| escalate. If you think that's too radical, there's plenty of
| variations: a cap on retweets per day. A dampening of how often a
| tweet can be retweeted in a period of time to slow the network
| effect.
| 
| - Put a cap on max tweets per day.
| 
| - When you go into a polarized thread and rapidly like a hundred
| replies that are on your "side", you are part of the problem and
| don't know how to use the like button. Hence, a cap on max likes
| per day or max likes per thread. So that they become quality
| likes that require thought. Alternatively, make shadow-likes.
| Likes that don't do anything.
| 
| - When you're a small account spamming low effort replies and the
| same damn memes on big accounts, you're hitchhiking. You should
| be shadow-banned for that specific big account only. People would
| stop seeing your replies only in that context.
| 
| - Mob culling. When an account or tweet is mass reported in a
| short time frame and it turns out that it was well within
| guidelines, punish every single user making those reports. Strong
| warning, after repeated abuse a full ban or taking away the
| ability to report.
| 
| - DM culling. It's not normal for an account to suddenly receive
| hundreds or thousands of DMs. Where a pile-on in replies can be
| cruel, a pile-on in DMs is almost always harassment. Quite a few
| people are OK with it if only the target is your (political)
| enemy, but we should reject it by principle. People joining such
| campaigns aren't good people, they are sadists. Hence they should
| be flagged as potentially harmful. The moderation action here is
| not straightforward, but surely something can be done.
| 
| - Influencer moderation. Every time period, comb through new
| influencers manually, for example those breaking 100K followers.
| For each, inspect how they came to power. Valuable, widely loved
| content? Or toxic engagement games? If it's the latter, dampen
| the effect, tune the alghoritm, etc.
| 
| - Topic spam. Twitter has "topics", great way to engage in a
| niche. But they're all engagement farmed. Go through these topics
| manually every once in a while and use human judgement to tackle
| the worst offenders (and behaviors)
| 
| - Allow for negative feedback (dislike) but with a cap. In case
| of a dislike mob, take away their ability to dislike or cap it.
| 
| Note how none of these potential measures address what it is that
| you said, it addresses behavior: the very obvious misuse/abuse of
| the system. In that sense I agree with the author. Also, it
| doesn't require AI. The patterns are incredibly obvious.
| 
| All of this said, the above would probably make Twitter quite an
| empty place. Because escalated outrage is the product.
 
| lm28469 wrote:
| Reading these threads on twitter is like listening to a friend
| having a bad mdma trip replaying his whole emotional life to you
| in a semi incoherent diarrhea like stream of thoughts
| 
| Please write a book, or at the very least an article... posting
| on twitter is like writing something on a piece of paper, showing
| it to your best friend and worst enemy before throwing it in the
| trash
 
  | Canada wrote:
  | If only there was some site that was good for posting longer
  | text, with a really good comment system to discuss it...
 
  | mbesto wrote:
  | And hilariously he starts with "How do you solve the content
  | moderation problems on Twitter?" and never actually answer it.
  | Just rambles on about a dissection of the problem. Guess we
  | know now why content moderation was never "solved" at Reddit,
  | nor will it ever be.
 
    | ilyt wrote:
    | He kinda did in roundabout way; the "perfect" moderation,
    | even if possible, will turn it into nice and cultured place
    | to have discussion and _that doesn 't bring controversy and
    | sell ads_.
    | 
    | You would have way less media "journalists" making a fuss
    | about what someone said on that social network and would have
    | problems just getting it to be popular, let alone displace
    | any of the big ones. It would maybe be possible with existing
    | one but that's a ton of work and someone needs to pay for
    | that work.
    | 
    | And it's entirely possible for smaller community to have
    | that, but the advantage with this is small community about X
    | will also have moderators that care about X so
    | 
    | * any on-topic bollocks can be spotted by mods and it is no
    | longer "unknown language"
    | 
    | * any off-topic bollocks can be just dismissed with "this is
    | a forum about X, if you don't like it go somewhere else
 
      | mbesto wrote:
      | > the "perfect" moderation, even if possible, will turn it
      | into nice and cultured place to have discussion and that
      | doesn't bring controversy and sell ads.
      | 
      | That's not a solution though since every for profit
      | business is generally seeking to maximize profit, and
      | furthermore we already knew this to be the case - nothing
      | he is saying is novel. I guess that's where I'm confused.
 
  | drewbeck wrote:
  | There's a study to be done on the polarization around twitter
  | threads. I have zero problem with them and find overall that
  | lots of great ideas are posted in threads, and the best folks
  | doing it end up with super cogent and well written pieces. I
  | find it baffling how many folks are triggered by them and
  | really hate them!
 
    | rchaud wrote:
    | This is likely because threads are a "high engagement" signal
    | for Twitter and therefore prone to being gamed.
    | 
    | There are courses teaching people how to game the Twitter
    | algo. One of those took off significantly in the past 18
    | months. You can tell by the number of amateurs creating
    | threads on topics far beyond their reach. The purpose of
    | these threads is for it to show up on people's feeds under
    | the "Topic" section.
    | 
    | For example, I often see see random posts from "topics"
    | Twitter thinks I like (webdev, UI/UX, cats, old newspaper
    | headlines). I had to unsubscribe from 'webdev' and "UI/UX"
    | because the recommended posts were all growth hackers. It
    | wasn't always that way.
    | 
    | I'm not the only one, others have commented on it as well,
    | including a well known JS developer:
    | 
    | https://twitter.com/wesbos/status/1587071684539973633
 
      | drewbeck wrote:
      | > This is likely because threads are a "high engagement"
      | signal for Twitter and therefore prone to being gamed.
      | 
      | You mean this is the reason folks respond differently to
      | the form of twitter thread? This is one that is definitely
      | not from a growth hacker but folks here still seem to hate
      | it.
 
  | peruvian wrote:
  | Thing is no one's going to read a blog post that he would've
  | linked to. As bad as they are, Twitter threads guarantee a way
  | larger audience.
 
  | pc86 wrote:
  | These things seem to be fine when it's 5-6 tweets in a coherent
  | thread. There's even that guy who regularly multi-thousand-word
  | threads that are almost always a good read.
  | 
  | This thread in particular is really bad.
 
  | heed wrote:
  | What got me was him weaving in (2-3 times) self promotion
  | tweets of some tree planting company he funds/founded(?). He
  | basically personally embedded ads into his thread, which is
  | actually kind of smart I suppose, but very confusing as a
  | reader.
 
    | pjc50 wrote:
    | Kind of genius to put it in the middle. Most normal people
    | write a tweet that blows up and then have to append "Check
    | out my soundcloud!" on the end. Or an advert for the nightsky
    | lamp.
 
      | rchaud wrote:
      | That was the middle? I stopped reading once it got to
      | 'here's my exciting new gig about trees' bit.
 
        | adharmad wrote:
        | There is one at the end too (if you reach that far)
        | shilling for another company where he is an investor -
        | Block Party.
 
    | mikeryan wrote:
    | I didn't even know he circled around back to the topic. I
    | split when I got to "TREES!" and wondering "that's it?"
    | 
    | After this comment I went back to read the rest.
 
    | toss1 wrote:
    | At the same time (as much as I strongly support climate
    | efforts, and am impressed by his approach, so give him a pass
    | in this instance), that 'genius move' sort of needs to be
    | flagged as his [Category #1 - Spam], which should be
    | moderated. It really is inserting off-topic info into another
    | thread.
    | 
    | The saving grace may be that both small enough volume and
    | sufficiently interesting to his audience to be just below the
    | threshold.
 
      | slowmovintarget wrote:
      | Was he perhaps trying for a Q.E.D. there?
 
  | 2OEH8eoCRo0 wrote:
  | https://news.ycombinator.com/newsguidelines.html
  | 
  | > Please don't complain about tangential annoyances--e.g.
  | article or website formats, name collisions, or back-button
  | breakage. They're too common to be interesting.
 
    | IshKebab wrote:
    | It's not really a tangential annoyance. I literally couldn't
    | read the post because of the insane format.
 
    | miiiiiike wrote:
    | It was a massive stream of tweets, with two long digressions,
    | and several embeds. The only thing that would have made it
    | worse is if every tweet faded in on scroll.
    | 
    | If we're going to pedantically point out rules, why don't we
    | add one that says "No unrolled Twitter threads."?
 
      | Karunamon wrote:
      | It is not pedantic, it is people derailing possibly
      | interesting discussion of the content with completely off-
      | topic noise discussion of the presentation. If you do not
      | like the presentation there are ways to change it.
 
      | drewbeck wrote:
      | If we're going to pedantically point out rules why don't we
      | cook hamburgers on the roof of parliament? Or something
      | else that isn't pedantically point out rules?
 
  | Karawebnetwork wrote:
  | Imagine it is a text and you can mark any paragraph. You can
  | save that paragraph, like it, or even reply to it. So the
  | interaction can grow like tentacles (edit: or rather like a
  | tree).
  | 
  | Right now, I could make a comment on either your first or
  | second paragraph, or on your entire comment. However, there is
  | no way to determine which category my reply falls into until
  | you have read it entirely. On a platform like Twitter, where
  | there can be up to 100,000 comments on a given piece of
  | content, this is very useful.
  | 
  | Better yet, it allows the author himself to dig down into
  | tangent. In theory, someone could create an account and then
  | have all of their interactions stay on the same tree without
  | ever cutting off. Essentially turning their account into an
  | interconnected "wiki" where everyone can add information.
  | 
  | With enough time your brain no longer registers the metadata
  | around the tweet. If you ignore it and read it as an entire
  | text it is not very different from a regular article or long
  | form comment:
  | https://threadreaderapp.com/thread/1586955288061452289.html
 
    | dingaling wrote:
    | I am imaging a normal long-form blog format but with comments
    | collapsed after each paragraph as a compromise between the
    | two current options.
 
    | rakoo wrote:
    | > Right now, I could make a comment on either your first or
    | second paragraph, or on your entire comment. However, there
    | is no way to determine which category my reply falls into
    | until you have read it entirely
    | 
    | That is _exactly_ what quoting does, and is older than the
    | web itself.
 
      | ilyt wrote:
      | The less inept sites also allow you to just select text and
      | click reply to get that quote and your cursor set below,
      | ready for reply
 
    | rcarr wrote:
    | This is brilliant, I had never thought about it like this
    | before. I'd maybe say grow like a tree rather than tentacles
    | although you might have a point in that if you're speaking
    | with the wrong person it could be pretty cthulonic.
 
    | lm28469 wrote:
    | What you're saying is that we should optimise the way we
    | debate things to please the algorithm and maximise user
    | engagement instead of maximising quality content and
    | encouraging deep reflexions
    | 
    | The truth is people can't focus for more than 15 seconds so
    | instead of reading a well researched and deep article or book
    | that might offer sources, nuances, &c. they'll click "like"
    | and "retweet" whoever vomited something that remotely goes
    | their way while ignoring the rest
    | 
    | > If you ignore it and read it as an entire text it is not
    | very different from a regular article or long form comment
    | 
    | It is extremely different as each piece is written as a
    | independent 10s thought ready to be consumed and retweeted.
    | Reading it on threadreaderapp makes it even more obvious,
    | your brain need to work 300% harder to process the semi
    | incoherent flow, some blogs written by 15 years old are more
    | coherent and pleasant to read than this
    | 
    | btw this is what I see on your link, more ads:
    | https://i.imgur.com/rhaXStj.png
 
      | Karawebnetwork wrote:
      | > What you're saying is that we should optimise the way we
      | debate things to please the algorithm and maximise user
      | engagement instead of maximising quality content and
      | encouraging deep reflexions
      | 
      | Not at all, in my opinion being able to interact with every
      | piece of an exchange allows to dig down into specific
      | points of a debate.
      | 
      | There is a soft stop at the end of every tweet because it's
      | a conversation and not a presentation. It's an interactive
      | piece of information and not a printed newspaper. You can
      | interact during the thread and it might change its outcome.
      | 
      | When you are the person interacting, it's similar to a real
      | life conversation. You can cut someone and talk about
      | something else at any time. The focus conversation will
      | shift for a short moment and then come back to the main
      | topic.
      | 
      | For someone arriving after the fact, you have a time
      | machine of the entire conversation.
      | 
      | ---
      | 
      | About the link, it is only the first result on Google
      | because I don't use those services and not me vetting for
      | this specific one. I also use ad blockers at all levels
      | (from pi-hole to browser extension to VPN level blocking),
      | so I don't see ads online.
      | 
      | If I go meta for a second, this is the perfect example of
      | how breaking ideas into different tweets can be useful.
      | 
      | Were I to share your comment on its own, it contains that
      | information about a link that is not useful to anyone but
      | you and I.
      | 
      | For someone reading our comments, they have to go through
      | this interaction on the ads and this product. If instead
      | this were two tweets it would have allowed us to comment on
      | this in parallel. If it was HN, imagine if you had made two
      | replies under my comments and we could have commented under
      | each. However, that's the wrong way on this platform.
 
    | ilyt wrote:
    | > Right now, I could make a comment on either your first or
    | second paragraph, or on your entire comment. However, there
    | is no way to determine which category my reply falls into
    | until you have read it entirely. On a platform like Twitter,
    | where there can be up to 100,000 comments on a given piece of
    | content, this is very useful.
    | 
    | Oh, look, I have managed to reply to your second paragraph
    | without having to use twatter, how quaint!
 
      | Karawebnetwork wrote:
      | There would be a lot of noise if everyone left 5 comments
      | under every comments. This is not the way HN is built.
      | Commenting too quickly even blocks you from interacting.
 
  | polytely wrote:
  | I've never understood this, it's just reading: you start at the
  | beginning of tweet, you read it, then go to the next tweet and
  | read it. How is that different from reading paragraphs?
 
    | lm28469 wrote:
    | idk man
    | 
    | Maybe
    | 
    | we could put more
    | 
    | Words in a single
    | 
    | Please visit my new website and subscribe to my podcast
    | 
    | Line so it would be
    | 
    | More readable
    | 
    | random user comment: yes you're right mark that would be
    | better
    | 
    | and much more user friendly
    | 
    | _ insert latest Musk schizophrenic rant_
    | 
    | Ah and I forgot the "sign up to read the rest" pop up that
    | seemingly appears at random interval
 
      | mikkergp wrote:
      | This is great! It's like modern poetry. I think the
      | research suggests less words per line make it faster
      | reading too.
 
      | adharmad wrote:
      | Like re-inventing Bukowski's writing style with terrible UX
 
      | polytely wrote:
      | are you reading on the computer? maybe thats the
      | disconnect. for me each tweet has around 4 lines of text.
      | 
      | so it reads more like the post you are reading now. where
      | each tweet is a decent chunk of text.
      | 
      | making the reading experience only marginally worse than
      | reading something on hacker news.
 
        | aniforprez wrote:
        | The amount of UI noise around each tweet and how much you
        | have to scroll, coupled with the need to trigger new
        | loads once Twitter has truncated the number of replies
        | and also HOW MUCH YOU HAVE TO SCROLL makes this a
        | _terrible_ experience
        | 
        | I understand why people tweet rather than write blogs.
        | Twitter gives more visibility and is a far lower barrier
        | of entry than sitting down and writing an article or a
        | blog. That Twitter hasn't solved this problem after years
        | of people making long threads and this being a big way
        | that people consume content on the platform is a failure
        | on their part. Things like ThreadReader should be in-
        | built and much easier to use. I think they acquired one
        | of these thread reader apps too
 
        | j33zusjuice wrote:
 
    | abetusk wrote:
    | I think this is important enough to highlight. Tweets are
    | very different from other forms of communication on the
    | internet. You can see it even here on HN in the comments
    | section.
    | 
    | Twitter railroads the discussion into a particular type by
    | the form of discourse. Each tweet, whether it's meant to or
    | not, is more akin to a self contained atomic statement then a
    | paragraph relating to a whole. This steers tweets into short
    | statements of opinion masquerading as humble, genuine
    | statements of fact. Often times each tweet is a simple idea
    | that's given more weight because it's presented in tweet
    | form. An extreme example is the joke thread of listing out
    | each letter of the alphabet [0] [1].
    | 
    | When tweets are responding to another tweet, it comes off as
    | one of the two extreme choices of being a shallow affirmation
    | or a combative "hot take".
    | 
    | Compare this with the comments section here. Responses are,
    | for the most part, respectful. Comments tend to address
    | multiple points at once, often interweaving them together.
    | When text is quoted, it's not meant as a hot take but a
    | refresher on the specific point that they're addressing.
    | 
    | The HN comments section has its problems but, to me, it's
    | night and day from Twitter.
    | 
    | I basically completely avoid responding to most everything on
    | Twitter for this reason. Anything other than a superficial
    | "good job" or "wow" is taken as a challenge and usually gets
    | a nasty response. I also have to actively ignore many tweets,
    | even from people I like and respect, because the format over
    | emphasizes trivial observations or opinions.
    | 
    | [0]
    | https://twitter.com/dancerghoul/status/1327361236686811143
    | 
    | [1]
    | https://twitter.com/ChaikaGaming/status/1270330453053132800
 
    | P_I_Staker wrote:
    | You gotta understand their angle...
    | 
    | ... in the early days of the internet ...
    | 
    | ... comments could be very, very long; the user was given a
    | virtual unbounded battleground to fight their ideological
    | battles ...
    | 
    | ... The public, the rabble, couldn't stop.. The words kept
    | coming; a torrent of consonants and vowels descending upon
    | our eye ba ... (limit exceeded)
    | 
    | ... lls like an avalanche of ideas ...
    | 
    | ... it was too much and twitter was borne ...
    | 
    | ... the people keep their ideas small, like their tiny
    | brains, and non-existent attention spans ...
    | 
    | P.S. I was gonna write this as a comment chain, but
    | HackerNews, in all their wisdom, limits self-replies to only
    | one
 
      | [deleted]
 
  | rchaud wrote:
  | If you want to talk diarrhea, look no further than those "save
  | to Readwise / Notion / Pocket" comments that pollute most of
  | these long threads.
 
  | rideontime wrote:
  | But the twitter thread makes it much easier to pivot into
  | talking about his latest startup that's totally unrelated!
 
  | dariusj18 wrote:
  | It seems apt that the most engaged comment in the thread is a
  | meta comment which derails any conversation about the content
  | of the post itself.
 
    | rakoo wrote:
    | The post says that moderation is first and foremost a signal-
    | to-noise curation. Writing long form content in a Twitter
    | thread greatly reduces the signal-to-noise ratio.
 
    | hypertele-Xii wrote:
    | The medium is the message.
 
      | dariusj18 wrote:
      | But also an example of how moderation or lack therein would
      | help to serve a particular end goal. ex. HackerNews is a
      | pretty well moderated forum, however sometimes the content
      | (PC being related to technology) is within the rules, but
      | the behavior it elicited in the other users is detrimental
      | to the overall experience.
 
  | bongobingo1 wrote:
  | nitter.net has a vaguely more readable view, not good but
  | better.
  | 
  | https://nitter.net/yishan/status/1586955288061452289
 
    | dynm wrote:
    | Just realized there are extensions that will auto-redirect
    | all twitter links to nitter. Why didn't I do this year ago!?
 
  | foobarbecue wrote:
  | https://threadreaderapp.com/thread/1586955288061452289.html
  | 
  | Agreed, though, Twitter threads are a really poor
  | communications medium.
 
    | madsmith wrote:
    | This thread was a delightful read.
    | 
    | Just NOT in twitter. I gave up on twitter and signed out of
    | it years ago and refuse to sign back in.
    | 
    | I spent a good hour of my life looking for ways to read this
    | thread. I personally know Yishan and value the opinions he
    | cares to share so I new this would be interesting if I could
    | just manage to read it.
    | 
    | Replacing the url to nitter.net helped but honestly it was
    | most cohesive in threadreaderapp although it missed some of
    | the referenced sidebar discussions (like the appeal to Elon
    | to not waste his mental energy on things that aren't real
    | atom problems).
 
  | throw7 wrote:
  | agreed, but you can go here as a workaround:
  | 
  | https://threadreaderapp.com/thread/1586955288061452289.html
 
  | PM_me_your_math wrote:
  | It is painful. I got about 11 posts in before giving up. You
  | described it perfectly.
 
  | awb wrote:
  | What's funny is he's arguing that moderation should be based on
  | behavior, not content. And that you could identify spam if it
  | was written in Loren Ipsum.
  | 
  | If this thread and self-referential Tweeting was written in
  | Loren Ipsum, it would definitely look like spam to me.
  | 
  | So I guess I disagree with one of the main points. For me, the
  | content matters much more than the behavior. Pretty sure that's
  | how the Supreme Court interprets 1A rights as well. The
  | frequency and intensity of the speech hasn't played a part in
  | any 1A cases that I can remember, it's exclusively if the
  | content of the speech violates someone's rights and then
  | deciding which outcome leads to bigger problems, allowing the
  | speech or not.
 
  | dubeye wrote:
  | You do sound a bit like my dad complaining about youngsters
  | reading kindle.
  | 
  | I read long twitter threads often, you get used to it
 
    | lm28469 wrote:
    | People also can get used to a diet of stale bread and bad
    | soup, it doesn't mean I'm striving for a stale bread and bad
    | soup diet
 
    | gort19 wrote:
    | I've heard you get used to jail too.
 
  | pjc50 wrote:
  | The core is really:
  | 
  | > https://twitter.com/yishan/status/1586956650455265281
  | 
  | "Here`s the answer everyone knows: there IS no principled
  | reason for banning spam. We ban spam for purely outcome-based
  | reasons:
  | 
  | It affects the quality of experience for users we care about,
  | and users having a good time on the platform makes it
  | successful."
  | 
  | And also this Chinese Room argument: "once again, I challenge
  | you to think about it this way: could you make your content
  | moderation decisions even if you didn`t understand the language
  | they were being spoken in?""
  | 
  | In other words, there are certain kinds of post which trigger
  | escalating pathological behavior - more posts - which destroy
  | the usability platform for _bystanders_ by flooding it. He
  | argues that it doesn 't matter what these posts _mean_ or whose
  | responsibility is it for the escalation, just the simple
  | physics of  "if you don't remove these posts and stop more
  | arriving, your forum will die".
 
    | bitshiftfaced wrote:
    | I would argue that the signal-to-noise ratio outcome-based
    | reason _is_ the principle: it 's off-topic. You could also
    | argue another principle: you're censoring a bot, not a human.
 
  | danwee wrote:
  | I gave up after the 3rd tweet in the thread. I can't understand
  | why Twitter threads are a thing.
 
    | nfin wrote:
    | my guess is that people can like () individual posts.
    | 
    | The positive of that is:
    | 
    | a) possibility to like () just one post, or 2, 3... depending
    | of who good the thread is
    | 
    | b) the fine granular way to like () gives the algorithm way
    | better possibilities to whom to show a thread and even
    | better, to first show just one intereting post out of that
    | thread (also people can mores easily quote or retweet
    | individual parts of a thread)
 
      | Akronymus wrote:
      | Why are you adding () after "like"? Havent seen that
      | convention before, so I am unaware of the meaning.
 
        | KMnO4 wrote:
        | There may have been an emoji or Unicode character between
        | the parens that was stripped by HN.
 
        | Akronymus wrote:
        | That does make quite a lot of sense.
 
        | darrenf wrote:
        | I was assuming there's a trademark symbol (tm) that had
        | been stripped by HN. But since I've managed to post one,
        | I'm apparently wrong!
 
      | fknorangesite wrote:
      | _Or retweet_ individual posts. It makes each one a possible
      | pull-quote.
 
        | jjulius wrote:
        | Yep; everything's a soundbite.
 
  | fknorangesite wrote:
  | At this point, posting this sentiment on HN is more boring and
  | irritating than the tweet-thread format could ever be.
 
    | bee_rider wrote:
    | It is also specifically called out in the guidelines as not
    | helpful.
 
  | sorum wrote:
  | He inadvertently invented the Twitter mid-roll ad and we're all
  | doomed now because of it
 
| shkkmo wrote:
| Let's take the core points at the end in reverse order:
| 
| > 3: Could you still moderate if you can`t read the language?
| 
| Except, moderators do read the language. If think it is pretty
| self-serving to say that users views of moderation decisions are
| biased by content but moderators views are not.
| 
| > 2: Freedom of speech was NEVER the issue (c.f. spam)
| 
| Spam isn't considered a free speech issue because we generally
| accept that spam moderation is done based on behavior in a
| content-blind way.
| 
| This doesn't magically mean that any given moderation team isn't
| impinging free speech. Especially when there are misinformation
| policies in place which are explicitly content-based.
| 
| > 1: It is a signal-to-noise management issue
| 
| Signal-to-noise management is part of why moderation can be good,
| but it doesn't even justify the examples from the twitter thread.
| Moderation is about creating positive experiences on the platform
| and signal-to-noise is just part of that.
| 
| The
 
| modeless wrote:
| It seems like he's arguing that people claiming moderation is
| censoring them are wrong, because moderation of large platforms
| is dispassionate and focused on limiting behavior no one likes,
| rather than specific topics.
| 
| I have no problem believing this is true for the vast majority of
| moderation decisions. But I think the argument fails because it
| only takes a few exceptions or a little bit of bias in this
| process to have a large effect.
| 
| On a huge platform it can simultaneously be true that platform
| moderation is _almost_ always focused on behavior instead of
| content, and a subset of people and topics _are_ being censored.
 
  | mr_toad wrote:
  | Rules against hate speech will disproportionately affect males.
  | Does that mean they're biased against men? If so, is that even
  | a bad thing?
 
  | hackerlight wrote:
  | > On a huge platform it can simultaneously be true that
  | platform moderation is almost always focused on behavior
  | instead of content, and a subset of people and topics are being
  | censored.
  | 
  | He made this exact point in a previous post. Some topics look
  | like they're being censored only because they tend to attract
  | such a high concentration of bad actors who simultaneously
  | engage in bullying type behavior. They get kicked off for that
  | behavior and it looks like topic $X is being censored when it
  | mostly isn't.
 
    | modeless wrote:
    | That's not the same point. Again, I have no problem believing
    | that what you say happens, even often. Even still, some
    | topics may _really_ be censored. They may even be the same
    | topics; just because there 's an angry mob on one side of a
    | topic doesn't mean that everyone on that side of the topic is
    | wrong, and that's the hardest situation to moderate
    | dispassionately. Maybe even impossible. Which is when I can
    | imagine platforms getting frustrated and resorting to topic
    | censorship.
 
  | rootusrootus wrote:
  | Could also be that some objectionable behavior patterns are
  | much more common in some ideological groups than others, which
  | makes it appear as if the moderation is biased against them. It
  | is, just not in the way they think.
 
| ethotool wrote:
| Nobody has the answers. Social media is an experiment gone wrong.
| Just like dating apps and other pieces of software that exist
| that are trying to replace normal human interaction. These first
| generation prototypes have a basic level of complexity and I
| expect by 2030 technology should evolve to the point where better
| solutions exist.
 
  | dna_polymerase wrote:
  | And when, ever in human history, did something improve without
  | intelligent people trying to solve these issues?
 
| sweetheart wrote:
| I'm amazed at the number of people in this thread who are annoyed
| that someone would insert mention of a carbon capture initiative
| into an unrelated discussion. The author is clearly tired of
| answering the same question, as stated in the first tweet, and is
| desperately trying to get people to think more critically about
| the climate crisis that is currently causing the sixth mass
| extinction event in the history of the planet.
| 
| Being annoyed that someone "duped" you into reading about the
| climate crisis is incredibly frustrating to activists because
| it's SO important to be thinking about and working on, and yet
| getting folks to put energy into even considering climate crisis
| is like pulling teeth.
| 
| I wonder if any of the folks complaining about the structure of
| the tweets has stopped to think about why the author feels
| compelled to "trick" us into reading about carbon capture.
 
  | Mezzie wrote:
  | To add another perspective (albeit with politics rather than
  | climate change):
  | 
  | I worked in political communications for a while. Part of the
  | reason it was so toxic to my mental health and I burnt out was
  | that it was nearly impossible to avoid politics online even in
  | completely unrelated spaces. So I'd work 40 hours trying to
  | improve the situation, log in to discuss stupid media/fan shit,
  | and have to wade through a bunch of stuff that reminded me how
  | little difference I was making, stuff assuming I wasn't
  | involved/listening, etc. It was INFURIATING. Yes, I had the
  | option to not go online, but I'm a nerd living in a small city.
  | There isn't enough people here that share my interests to go
  | completely offline.
  | 
  | Staying on topic helps people who are already involved in
  | important causes to step away and preserve their mental health,
  | which in turn makes them more effective.
 
  | rchaud wrote:
  | The simple fact of the matter is that too many people are
  | either resigned to a Hobbesian future of resource wars, or
  | profiting too much from the status quo to go beyond a
  | perfunctory level of concern.
  | 
  | $44bn of real-world cash was just spent on Twitter, and HN
  | users alone have generated tens of thousands of comments on the
  | matter.
  | 
  | How many climate tech related stories will have the same level
  | of interest?
 
| gambler wrote:
| _> No one argues that speech must have value to be allowed (c.f.
| shitposting)._
| 
|  _> Here`s the answer everyone knows: there IS no principled
| reason for banning spam._
| 
| The whole threads seems like it revolves around this line of
| reasoning, which strawmans what free speech advocates are
| actually arguing for. I've never heard of any of them, no matter
| how principled, fighting for the "right" of spammers to spam.
| 
| There is an obvious difference between spam moderation and
| content suppression. No recipient of spam wants to receive spam.
| On the other hand, labels like "harmful content" are most often
| used to stop communication between willing participants by a 3d
| party who doesn't like the conversation. They are fundamentally
| different scenarios, regardless of how much you agree or disagree
| with specific moderation decisions.
| 
| By ignoring the fact that communication always has two parties
| you construct a broken mental model of the whole problem space.
| The model will then lead you stray in analyzing a variety of
| scenarios.
| 
| In fact, this is a very old trick of pro-censorship activists.
| Focus on the speaker, ignore the listeners. This way when you
| ban, say, someone with millions of subscribers on YouTube you can
| disingenuously pretend that it's an action affecting only one
| person. You can then draw false equivalency between someone who
| actually has a million subscribers and a spammer who sent a
| message to million email addresses.
 
| [deleted]
 
___________________________________________________________________
(page generated 2022-11-03 23:00 UTC)