The company updated its support page a week ago to mention the use of the AI model as part of its harassment filter. The filter is powered by the LLM and is trained on content removed by Reddit’s internal tools and enforcement teams. It remains to be seen how effective this new tool will be in flagging objectionable content on the platform.
Key takeaways:
- Reddit is planning to use an AI model to detect harassment on its platform.
- The AI model is a Large Language Model (LLM) trained on content that has previously violated Reddit's terms.
- The model is also trained on moderator actions and content removed by Reddit's internal tools and enforcement teams.
- It's still uncertain how effective this new tool will be in flagging objectionable content.