(Translated by https://www.hiragana.jp/)
Incident Reporting System/zh: Difference between revisions - Meta Jump to content

Incident Reporting System/zh: Difference between revisions

From Meta, a Wikimedia project coordination wiki
Content deleted Content added
FuzzyBot (talk | contribs)
Updating to match new version of source page
No edit summary
Line 130: Line 130:
{{anchor|New!_21_September_2023_Update:_Sharing_incident_reporting_research_findings}}
{{anchor|New!_21_September_2023_Update:_Sharing_incident_reporting_research_findings}}
<span id="21_September_2023_Update:_Sharing_incident_reporting_research_findings"></span>
<span id="21_September_2023_Update:_Sharing_incident_reporting_research_findings"></span>
=== 2023ねん9がつ21にち更新こうしんぶんとおる事件じけん报告研究けんきゅう成果せいか ===
<div class="mw-translate-fuzzy">
=== {{new}} 2023ねん9がつ21にち更新こうしんぶんとおる事件じけん报告研究けんきゅう成果せいか ===
</div>


[[File:Findings Report Incident Reporting 2023.pdf|thumb|<span lang="en" dir="ltr" class="mw-content-ltr">Research Findings Report on Incident Reporting 2023</span>|alt=|{{dirstart}}]]
[[File:Findings Report Incident Reporting 2023.pdf|thumb|<span lang="en" dir="ltr" class="mw-content-ltr">Research Findings Report on Incident Reporting 2023</span>|alt=|{{dirstart}}]]

Revision as of 18:30, 19 June 2024

事件じけんどおり报系统

    关于

    维基媒体ばいたい基金ききん会期かいきもち改善かいぜん骚扰其他伤害事件じけん受害しゃどおり事件じけんてき方式ほうしき,以为しゃぐん提供ていきょうさら安全あんぜんさら健康けんこうてき环境。

    信任しんにんあずか安全あんぜん产品团队てきにん务是建立こんりゅう事件じけんどおり报系统(IRS)以前いぜんしょう为私じん事件じけんどおり报系统(PIRS))。わが们的标是让用户能够轻まつ安全あんぜんわたしみつどおり有害ゆうがい事件じけん

    专案背景はいけい

    关于骚扰有害ゆうがい事件じけんてきどおり报和处理,一直是维基媒体社群关注的话题。ずいしんてき通用つうようぎょう为准则てき建立こんりゅう,讨论よう户举报系统也是非ぜひつね重要じゅうようてき通用つうようぎょう为准则(UCoC)いち政策せいさく,为维もと媒体ばいたい专案そら间中てき所有しょゆう参与さんよしゃてい义了最低さいてい标准てきぎょう为准则和不可ふか接受せつじゅてきぎょう为。

    处理しょう案件あんけん当行とうこう为和政策せいさく违反事件じけんてき方式ほうしきやめざい不同ふどうてき维基媒体ばいたいそら间和专案さと执行并且ゆうつくえ发展,并在同社どうしゃぐん存在そんざい异。

    まい个维もと媒体ばいたい专案あるしゃぐんゆう自己じこてき管理かんり方式ほうしき事件じけんてきどおり报和处理ゆう种方しき

    • とおる过维もと讨论页
    • とおる通告つうこくばんめん
    • とおる过电邮件
    • とおる过维もと以外いがいてき私人しじん讨论频道(Discord、IRC とう

    对于许多よう户来说,发生事件じけんきさき该怎么处清楚せいそてき哪里、あずか谁交谈、如何いかどおり报、报告ちゅう包含ほうがん哪些资讯、报告如何いか处理、これきさきかい发生什么とう

    よう户必须知どう如何いかどおり报问题以及到哪里どおり报。 关于报告ひさげ交后かい发生什么,以及よう户应该有什么もちてき资讯也很しょう

    よし于通报过ほどてき复杂せい隐私考量こうりょうゆう些用户在どおり事件じけん时会かんいた安全あんぜん

    目前もくぜん还没ゆう标准てき方法ほうほう供用きょうよう户私そこつつみ交报つげ

    专案焦点しょうてん

    此专あんてきだか阶目标是希望きぼうのうさら容易よういかい决骚扰和有害ゆうがい事件じけん,并确这些报告以去いた必须处理它们てき适当实体组织。

    わが希望きぼう确保どおり报者てき隐私安全あんぜんわが们还希望きぼう确保报告具有ぐゆうせい确的讯息,并能导引到必须处理它们てき适当实体组织,どう时不かい给处じん员带らい额外てき压力。

    信任しんにん安全あんぜん产品团队ただしざいしょう事件じけんどおり报系统视为更だいてき维基媒体ばいたいしゃぐん事件じけん管理かんりせい态系统的いち部分ぶぶんわが们正ざい研究けんきゅうかく个系统如なん运作以及一切如何连接整合在一起。

    产品わざ更新こうしん

    Test the Incident Reporting System in Beta – November 2023

    We invite editors to test the initial version for the Incident Reporting System. It makes it possible file a report from the talk page where an incident occurs. This version is for learning about filing reports to a private email address (e.g., emergency(_AT_)wikimedia.org or an admin group). It doesn't cover all scenarios, like reporting to a public noticeboard. We need your opinions to see if this approach is effective.

    To test:

    1. Visit any talk namespace page on Wikipedia in Beta that contains discussions. We have sample talk pages available at User talk:Testing and Talk:African Wild Dog you can use and log in.

    2. Next, click on the overflow button (vertical ellipsis) near the Reply link of any comment to open the overflow menu and click Report (see slide 1). You can also use the Report link in the Tools menu (see slide 2).

    3. Proceed to file a report, fill the form and submit. An email will be sent to the Trust and Safety Product team, who will be the only ones to see your report. Please note this is a test and so do not use it to report real incidents.

    4. As you test, ponder these questions:

    • What do you think about this reporting process? Especially what you like/don’t like about it?
    • If you are familiar with extensions, how would you feel about having this on your wiki as an extension?
    • Which issues have we missed at this initial reporting stage?

    5. Following your test, please leave your feedback on the talk page.

    If you can't find the overflow menu or Report links, or the form fails to submit, please ensure that:

    • You have logged in
    • Your Beta account email address is confirmed
    • Your account has been created for over 3 hours and you have at least 1 edit.
    • You have enabled DiscussionTools because the MTP is integrated with DiscussionTools

    If DiscussionTools doesn’t load, a report can be filed from the Tools menu.

    If you can't file a second report, please note that there is a limit of 1 report per day for non-confirmed users and 5 reports per day for autoconfirmed users. These requirements before testing help to reduce the possibility of malicious users abusing the system.

    请参阅页面てき研究けんきゅう部分ぶぶん了解りょうかいさら资讯。

    事件じけんどおり报专あんてきよん更新こうしん – 2023 ねん 7 がつ 27 にち

    Hello everyone! For the past couple of months the Trust and Safety Product team has been working on finalising Phase 1 of the Incident Reporting System project.

    The purpose of this phase was to define possible product direction and scope of the project with your feedback. We now have a better understanding of what to do next. Read more.

    项目范围&MVP – 2022ねん11月8にち

    Our main goal for the past couple of months was to understand the problem space and user expectations around this project. The way we would like to approach this is to build something small, a minimum viable product (MVP), that will help us figure out whether the basic experience we are looking at actually works. Read more.

    りゅうほど

    项目阶段

    Figuring out how to manage incident reporting in the Wikimedia space is not an easy task. There are a lot of risks and a lot of unknowns.

    As this is a complex project it needs to be split into multiple iterations and project phases. For each of these phases we will hold one or several cycles of discussions in order to ensure that we are on the right track and that we incorporate community feedback early, before jumping into large chunks of work.

    だいいち阶段

    Preliminary research: collect feedback, reading through existing documentation.

    Conduct interviews in order to better understand the problem space and identify critical questions we need to answer.

    Define and discuss possible product direction and scope of project. Identify possible pilot wikis.

    At the end of this phase we should have a solid understanding of what we are trying to do.

    だい阶段

    Create prototypes to illustrate the ideas that came up in Phase 1.

    Create a list of possible options for more in-depth consultation and review.

    だいさん阶段

    Identify and prioritize the best possible ideas.

    Transition to software development and break down work in Phabricator tickets.

    Continue cycle for next iterations

    研究けんきゅう

    2023ねん9がつ21にち更新こうしんぶんとおる事件じけん报告研究けんきゅう成果せいか

    Research Findings Report on Incident Reporting 2023

    The Incident Reporting System project has completed research about harassment on selected pilot wikis.

    The research, which started in early 2023, studied the Indonesian and Korean Wikipedias to understand harassment, how harassment is reported and how responders to reports go about their work.

    The findings of the studies have been published.

    In summary, we received valuable insights on the improvements needed for both onwiki and offwiki incident reporting. We also learned more about the communities' needs, which can be used as valuable input for the Incident Reporting tool.

    We are keen to share these findings with you; the report has more comprehensive information.

    Please leave any feedback and questions on the talkpage.

    Pre-project research

    The following document is a completed review of research from 2015–2022 the Wikimedia Foundation has done on online harassment on Wikimedia projects. In this review we’ve identified major themes, insights, and areas of concern and provided direct links to the literature.

    Previous work

    The Trust and Safety Tools team has been studying previous research and community consultations to inform our work. We revisited the Community health initiative User reporting system proposal and the User reporting system consultation of 2019. We have also been trying to map out some of the conflict resolution flows across wikis to understand how communities are currently managing conflicts. Below is a map of the Italian Wiki conflict resolution flow. It has notes on opportunities for automation.

    On Italian Wikipedia, there's a 3-step policy in place for conflict resolution. This map visualizes this process and tries to identify opportunities for automation for both editors and admins.

    つね见问题

    Questions and answers from Phase 1 of the project

    Q: Is there data available about how many incidents are reported per year?

    A: Right now there is not a lot of clear data we can use. There are a couple of reasons for this. First, issues are reported in various ways and those differ from community to community. Capturing that data completely and cleanly is highly complicated and would be very time consuming. Second, the interpretation of issues also differs. Some things that are interpreted as harassment are just wiki business (e.g. deleting a promotional article). Review of harassment may also need cultural or community context. We cannot automate and visualize data or count it objectively. The incident reporting system is an opportunity to solve some of these data needs.

    Q: How is harassment being defined?

    A: Please see the definitions in the Universal Code of Conduct.

    Q: How many staff and volunteers will be needed to support the IRS?

    A: Currently the magnitude of the problem is not known. So the amount of people needed to support this is not known. Experimenting with the minimum viable product will provide some insight into the number of people needed to support the IRS.

    Q: What is the purpose of the MVP (minimal viable product)?

    A: The MVP is an experiment and opportunity to learn. This first experimental work will answer the questions that we have right now. Then results will guide the future plans.

    Q: What questions are you trying to answer with the minimum viable product?

    A: Here are the questions we need to answer:

    • What kind of reports will people file?
    • How many people will file reports?
    • How many people would we need in order to process them?
    • How big is this problem?
    • Can we get a clearer picture of the magnitude of harassment issues? Can we get some data around the number of reports? Is harassment underreported or overreported?
    • Are people currently not reporting harassment because it doesn’t happen or because they don’t know how?
    • Will this be a lot to handle with our current setup, or not?
    • How many are valid complaints compared to people who don't understand the wiki process? Can we distinguish/filter valid complaints, and filter invalid reports to save volunteer or staff time?
    • Will we receive lots of reports filed by people who are upset that their edits were reverted or their page was deleted? What will we do with them?

    Q: How does the Wikimedia movement compare to how other big platforms like Facebook/Reddit handle harassment?

    A: While we do not have any identical online affinity groups, the Wikimedia movement is most often connected with Facebook and Reddit in regard to how we handle harassment. What is important to consider is nobody has resolved harassment. Other platforms struggle with content moderation, and often they have paid staff who try to deal with it. Two huge differences between us and Reddit and Facebook are the globally collaborative nature of our projects and how communities work to resolve harassment at the community-level.

    Q: Is WMF trying to change existing community processes?

    A: Our plan for the IRS is not to change any community process. The goal is to connect to existing processes. The ultimate goals are to:

    • Make it easier for people who experience harassment to get help.
    • Eliminate situations in which people do not report because they don’t know how to report harassment.
    • Ensure harassment reports reach the right entities that handle them per local community processes.
    • Ensure responders receive good reports and redirect unfounded complaints and issues to be handled elsewhere.