Loomio

CoC DOs are enforceable too?

AW Aaron Wolf Public Seen by 243

I like that there are DOs and DON'Ts in the CoC. But I think there's not been enough consideration of enforcement.

If we want something to be just a general guidance that isn't enforceable, either it should be separated or at least have language like the "strive to" which is present in some areas.

A good CoC should not be vague about which items are hard enough rules that violations could be reported.

The reporting guidelines themselves could clarify how to handle the items that are the "strive to" stuff… (how to handle an image that lacks alt text for example) versus those violations that could result in censorship or other actions.

AW

Aaron Wolf Wed 8 Aug 2018 11:14PM

I'd like to see a separate (or extension) flow-chart for helping people decide the best resolution strategies when they are staying in the self-help part of the process.

I also don't see in the reporting any outcomes that are more qualified. I want this to be common:

  • moderator deletes/hides problem post
  • violator is told that if they repost an improved version that expresses their points without violating the CoC, then no further penalties will occur
  • a process where the reporter gets notified to verify whether a repost is an adequate fix
ED

emi do Wed 8 Aug 2018 12:55PM

@wolftune I am a little confused and uncertain about 'enforcement'. Who would do this 'enforcement' and do you think the CWG or members need (or would want) a list of explicit enforceable items in order to confront or discuss issues that are creating discomfort?

I think some of these are discussed in the conflict resolution guidelines (the soft steps prior to reporting) that allows individuals, on a case by case basis, determine what their needs/wants are in order for a conflicting situation to be resolved.

To go back to your example: Let's say I wasn't assuming good faith and interpreted a comment that someone as being derogatory against women. In our current model with conflict resolution: the person who made the original comment would (if they felt comfortable with doing so) would message me privately to start a conversation, ask questions, understand where I was coming from and perhaps make a plan about how we both can avoid incendiary comments in the future. If they didn't feel comfortable messaging me directly, they could solicit support from a mutually trusted friend or a 'social.coop steward' (to be determined by CWG group) to mediate a conversation to help resolve the issue. All of this before having to resort to a report which would lead to formal action. If we were instead using an enforcement model, I feel like it would encourage members to hide behind a CoC rather than really listening to where the other was coming from and seeking to understand the situation.

I am aware that there are gaps in this type of approach, but am also trying to be pragmatic about where volunteers will want to allocate their time (social.coop stewards vs. social.coop enforcers) and how we can facilitate better communication and fostering/empowering positive interactions.

AW

Aaron Wolf Wed 8 Aug 2018 11:24PM

I like all this. What I see missing is that sometimes the best thing for such a situation is to get the problem post hidden ASAP while the resolution is being worked out.

So, in that case, enforcement means that there is a CoC item to point to for requesting that a post be hidden at least temporarily. Since there's no way technically with our tools (I think) to give the reporter the ability to hide another's posts (which if it existed would be a privilege for those with a certain level of trust), they probably need to request that a moderator do the hiding.

But getting a post hidden can feel really bad if interpreted badly. So, we want a situation where it's as clear as possible that there's no personal attack, no further penalties… it's just a temporary situation as the concern is resolved.

Point is: a derogatory-against-women post shouldn't sit public and get public replies etc. while a private discussion is happening to address the issue.

The whole goal needs to be to block the tensions that arise from such a post before they cause further issues in the community. Most times, it will be a minor slip or misunderstanding or carelessness. So, it should be treated as casually as possible other than the getting hidden.

Right now, most everything online is either stuff staying public and causing all sorts of problems or treated as a real rebuked, finger-wagging, punished serious violation. There's no middle where it's like "this post is likely to cause problems, I already feel tense about it, we need to address it maybe with an edited repost that is better, so we're just hiding it for now — and as soon as we can resolve this we can continue as though nothing went wrong. There's no personal offense here per se…"

But even that situation needs to only happen if there's something from the CoC to point to about why the hiding should happen. Otherwise, we need to leave the hiding to be purely voluntary from the poster if they get a private request. So, that's the distinction for enforcement: is this issue something where a mod might hide a post prior to the poster getting a chance to voluntarily hide it (or decide they don't want to).

ED

emi do Thu 9 Aug 2018 12:59AM

Again more food for thought.

At the moment it feels as though the rate at which toots/re-toots take place, and by the nature of Mastadon design where you can't quote toots (see anti-abuse rules) I don't feel that this type of explosive spread of a harmful thought will take place.

Going back to your example, a derogatory post about women, I think these happen on a gradient. Some feel like they threaten safety (I hate women and I am going to kill them all), while others are less obvious (women aren't as smart as men which is why we shouldn't be expecting equal representation in workplaces). I think the former needs to be taken down immediately, which according to our flow chart, would probably result in a report and action within 24 hours. The latter is one that I hope the community would address in direct dialogue, and feel like if we were to step in and hide it without engaging the member, would actually escalate the situation.

How about adding "If the conflict is in regards to a specific toot or written comment in a social.coop space, consider asking for it to be taken down temporarily while you work on resolving the conflict" under the "invite conversation" section of the Conflict Resolution Guide?

Personally, I am a little uncomfortable with the idea of having to create a task force that is going around hiding toots/comments on behalf of someone else? I fully support the need to support those who are feeling triggered, and I would hope we can come up with strategies to empower those members(such as muting/blocking etc.) and provide avenues where that member can be heard (social.coop stewards) and have their feelings validated. I think sometimes in an effort to help, we want to 'rescue' someone and we do it by putting our superhero costume on and seeking justice and executing punishment when there might be other ways we can facilitate resilience of the community?

AW

Aaron Wolf Thu 9 Aug 2018 2:06AM

if we were to step in and hide it without engaging the member

Misunderstanding. I'm never saying that toots should be hidden without notice. Every time a toot is hidden, it should be in conjunction with saying "this violated the CoC, please just post again in a manner that is aligned with the CoC". It needs to be an invitation to simply repost and continue.

In the case of "women are less smart", we probably don't want a public back-and-forth debating that. In the case of "you're less smart because you're a woman" we definitely don't want that to be discussed in a public back-and-forth debate.

How about adding "If the conflict is in regards to a specific toot or written comment in a social.coop space, consider asking for it to be taken down temporarily while you work on resolving the conflict" under the "invite conversation" section of the Conflict Resolution Guide?

That sounds pretty good!

As to the worry of excessive hiding, that's where (A) we need some clear guidance (but not a strict program, we want human judgment still) about when something deserves to be hidden (and it should be usually a private request to the poster to choose to delete and repost (editing would be far better, but that's not available yet in Mastodon) and (B) clarity from the CoC is part of that (so, the CoC should not say that it's a violation to trigger someone, it says that only a specific list of things are violations, so anything that's not a violation will not be hidden by any task force…)

ED

emi do Thu 9 Aug 2018 2:37PM

I've been looking at our CoC and trying to find ways to address some of the things that you have brought up and I just can't figure out an appropriate place for it.

I then come back to what the objective of the CoC is. I think you're raising really great points, I know I personally don't really feel like engaging/educating/debating misogynistic comments and at the same time, I am still concerned about our capacity and what we are actually able to implement.

I think encouraging the person who notices a post that might need amending to directly address the person to edit the post/take it down is probably the most realistic. I find that templates or examples are sometimes nice and a way to decrease barrier to taking that step? Perhaps we can put a template into the conflict resolution guideline?

Something along the lines of "I noticed your post on social.coop that mentioned __________. I took it to mean ______ which makes me feel ____________ . Would it be possible for you to edit your post/ take down your post/ put a CW on your post to be more reflective of the CoC of social.coop? "

AW

Aaron Wolf Thu 9 Aug 2018 7:15PM

I feel you're thinking about this well.

I think the template should reference the CoC, as in: "I noticed your [post] on social.coop and it seems to be to be problematic per [specified item(s) from CoC]. Would you please [delete and repost an improved version]?"

A final part to the optional part of the template should say "I think if you [suggested improvement], that would address the problem"

If editing is a possibility (as it is on Loomio), then the request can be for an edit. On Mastodon today, it seems to be a delete-and-repost out of necessity.

In my view, it's a responsibility of all members to help maintain our standards. So, it's not okay to leave it to an offended woman to object to a sexist post. The more clear it is on signing up that we have certain standards and we all work to maintain them, the less personal any particular situation is likely to feel.

The final point is that this process should suggest that people can directly contact or can request someone else do it on their behalf. And I don't think even the latter case needs to be a big drawn-out reporting / enforcement process.

Ideally, it would be the case (as it is in Discourse, which we don't use) that an edited or reposted item fixing a problem notifies the reporter so they can check on the fix being okay.

ED

emi do Fri 10 Aug 2018 2:50AM

Ok, so I think we're on the same page. Just some differences in wording. Here is my suggested edit to the template:
"I noticed your [post]. I took it to mean ______ which I feel is not reflective of [specific portion of CoC]. Would you please (to edit your post/ take down your post/ put a CW on your post so that it is more reflective of the CoC of social.coop?)"

Currently our conflict resolution and reporting guidelines don't specify that it has to be someone who identifies as the gender/race/sexuality etc. being addressed in the offending post. In fact, I would argue that it is specified in our CoC in these two points, that people SHOULD be speaking out if they see a problematic post regardless of if they identify as x, y z...:
* Help maintain a culture of inclusivity and open participation
* Show care, consideration, and respect for others

Perhaps your concern can be addressed in the annotated version as an example of how a member can maintain a culture of inclusivity?

AW

Aaron Wolf Fri 10 Aug 2018 3:54AM

I like your updated template wording overall.

Minor detail: I don't like the word "mean" because we don't want to suggest what the poster intended. We want to emphasize the problem with the post, whether it was intended or not. Maybe something more open-ended like "I noticed your [post]. I feel it may not adequately reflect [specific portion(s) of CoC] because ___"?

note portion(s) plural possible. Posts can violate multiple parts of CoC.

Other wordings instead of "took it to mean" could be things like "it could be interpreted as" or something like that.

Example: I assume good faith, give benefit of the doubt that someone did not mean to be sexist and taking it to be that way would be assuming bad faith. But I still want to call it out because it is easy to interpret as sexist. So, I call out the problem regardless of the poster not intending the problem.

Anyway, there's some misunderstanding about identity. I wasn't suggesting and don't want that the identity of any party have any relevance. Women don't get any extra pass on spreading misogyny for example, and anyone who recognizes a problem should do something regardless of their personal identity.

There can be a wide range of reasons people don't speak up or take action. I wasn't assuming how they identify would be part of that necessarily.

The core issue is that I'm not thinking about the worst-case scenarios. I don't want a CoC and process that only deals with horrible bad actors. I want to see troublesome minor slips addressed before they lead to division and tension. I think a good portion of online conflicts come from poorly dealing with misunderstandings or carelessness, and we should address that as well as addressing serious bigotry etc.

MB

Manuela Bosch Fri 10 Aug 2018 6:20AM

Thank you all. I have to admit, I am overwhelmed by the amount of information that is up in the air. I have no other idea how to solve this, than asking now to include the parts that are important to the person who brought them up into the respective document. I suggest we do so by considering that every change we make helps making the text/wording more

+ concise/focused
+ inclusive
+ inspirational
+ memorable
Let’s try not to discuss anything further. Just change the things or leave them for now as a next project/ in the living document we thought of. Does that make sense? I am speaking out of my role helping getting this done. :-)

Load More