Instagram is rolling out a new policy where, in addition to removing accounts with a certain percentage of content violating guidelines, a certain number of violations within a given window of time will also result in the removal of the account.
Previously, under the policy, the platform would disable accounts that have a certain percentage of violating content. The platform is also introducing a new notification process to help you understand if your account is at risk of being disabled.
If you post something on Instagram that goes against the Community Guidelines, you will be notified by a warning and the reason for removal along with the date and time it was posted.
Also Read: How to optimize Video Ad Campaigns on Facebook & Instagram
For instance, "Post removed for hate speech or symbols. Posted April 1 at 1:03 PM".
You will also see an illustration of the guideline you have violated. This notification will also offer you the opportunity to appeal the content deleted.
Appeals will be available for content deleted for violations of nudity & pornography, bullying & harassment, hate speech, drug sales, and counter-terrorism policies, but Instagram will be expanding appeals in the coming months.
If the content is found to be removed in error, it will be restored and the violation will be removed from the account’s record.
Instagram also mentioned that the option to appeal will also be able available within Instagram within the coming months.