What Happens When I Report Something to Facebook?

In the age of social media, it has become increasingly common for users to encounter content that violates community standards or policies. In such cases, Facebook provides its users with a reporting feature, which allows them to report the content to the platform for review. However, many users wonder what happens when they report something on Facebook, and whether the person they report gets notified. In this article, we will explore what happens when you report something on Facebook, the consequences of such reports, and whether the person reported gets notified.

The Reporting Feature on Facebook

Facebook’s reporting feature allows users to report content that violates community standards, including hate speech, violence, harassment, and fake news, among others. The reporting feature is an essential tool that helps Facebook maintain a safe and respectful environment for all users.

How to Report Content on Facebook

Reporting content on Facebook is a simple process. To report content on Facebook, follow these steps:

  • Click on the three dots (…) in the top right corner of the post.
  • Select “Find support or report post.”
  • Select the reason why you’re reporting the post.
  • Follow the on-screen instructions to complete the report.

Types of Content that Can Be Reported

Facebook has community standards that dictate what types of content are not allowed on the platform. Users can report any content that they believe violates these standards, including:

  • Hate speech
  • Violence and incitement
  • Harassment and bullying
  • Nudity and sexual activity
  • Spam and false information

What Happens When You Report Something on Facebook?

Review Process

When a user reports content on Facebook, the platform’s content moderation team reviews the report and the content in question. The review process involves a team of human moderators who check whether the content violates Facebook’s community standards.

Possible Outcomes of Reports

Once the moderation team reviews the report, they can take various actions based on their findings. These actions can include:

  • Removing the content
  • Warning the person who posted the content
  • Disabling the person’s account
  • Reporting the content to law enforcement agencies if necessary

Does the Person Reported Get Notified?

One of the most commonly asked questions about Facebook’s reporting feature is whether the person who posted the content gets notified when it is reported. The answer is no. When a user reports content on Facebook, the platform keeps the identity of the person who made the report anonymous. The person who posted the content is not notified of who reported it or that it was reported.


To conclude, reporting something on Facebook can have significant consequences, both for the person being reported and the person making the report. It is important to use this feature responsibly and only report content that violates Facebook’s Community Standards. By doing so, we can help create a safer and more respectful online community for everyone. It’s also important to remember that reporting someone on Facebook is not the only solution to conflict or disagreement. Communication, empathy, and understanding can go a long way in resolving issues and building stronger relationships. Let’s strive to use social media as a tool for connection and positive change, and always remember to treat others with kindness and respect both online and offline.

FAQs about Reporting on Facebook

How do you know if someone reported you on Facebook?

You won’t know who reported you on Facebook, and Facebook won’t notify you of the person who reported you. However, if your post or account has been removed or disabled, it could be a sign that someone reported you. Facebook may also send you a notification or email explaining why your content was removed or why your account was disabled.

How long does it take for Facebook to review a report?

The amount of time it takes for Facebook to review a report can vary depending on the nature of the report and the volume of reports being processed. Facebook aims to review reports as quickly as possible, and in some cases, they can take action within minutes. However, in more complex cases, it may take several days for Facebook to complete their investigation and take appropriate action. If you have reported content that violates Facebook’s Community Standards, you can rest assured that Facebook’s content moderation team will take appropriate action based on their findings, even if it takes some time.

Can I report content anonymously on Facebook?

Yes, it’s obvious that you can report content anonymously as your identity is not revealed to the person or people involved in the report. This means that the person being reported will not know that it was you who made the report. Facebook takes the privacy of their users seriously, and they ensure that your identity is kept confidential to protect your safety and privacy. Therefore, it is clear that reporting content anonymously on Facebook is a secure and private process.

Will Facebook tell me what action was taken after I make a report?

Facebook may send you a notification or email after you make a report, explaining whether or not the reported content was found to violate their Community Standards and what action was taken as a result. However, Facebook does not typically provide detailed information about the specific action taken against the reported content or the person responsible for it. 

Can I appeal a Facebook content moderation decision?

Yes, you can appeal a Facebook content moderation decision. If you believe that Facebook has made a mistake or that your content was wrongfully removed, you can appeal the decision by submitting a request for review. To do this, go to the Help Center on Facebook and follow the steps to appeal the decision.

Facebook will then review your appeal and determine whether or not to reinstate the content or account. It’s important to note that not all appeals are successful, and Facebook’s content moderation team makes their decisions based on their Community Standards and other policies. However, if you believe that your content was removed in error or that your account was disabled unfairly, it is worth submitting an appeal for review.

How many reports does it take to get banned from Facebook?

The number of reports it takes to get banned from Facebook depends on the severity and frequency of the violations. Facebook’s policies prohibit various forms of harmful or offensive behaviour, such as hate speech, harassment, and violence. If a user repeatedly violates these policies or engages in behaviour that is deemed to be severe or egregious, Facebook may take action, including issuing warnings, disabling accounts, or banning users from the platform.

However, Facebook does not have a set number of reports that will automatically result in a ban. Instead, Facebook evaluates reports on a case-by-case basis and takes appropriate action based on their findings. If a user violates Facebook’s policies, they may receive consequences, including warnings, temporary suspensions, or permanent bans, depending on the nature and severity of their violations.

Worth checking: Does Facebook Notify When You Screenshot?

Photo of author

Bibek Sapkota

I'm Bibek | Tech Enthusiast & Lifelong Learner. | Playing on the Web for the Past Few Years as a Full-Time Tech Blogger. At TechnoBite, I curate insightful content ranging from how-to guides to in-depth tech reviews. Passionate about continuous learning and sharing knowledge, my mission is to empower my audience with valuable information to navigate the ever-evolving world of technology.

Leave a Comment