The new feature began rolling out to the app’s billions of users on Monday.
The AI operates from a list of words and phrases that have been reported as offensive in the past.
If the AI detects offensive language on a comment, it will send a prompt to the writer and give them a chance to edit or delete their comment before it is posted.
The app hopes to encourage users “to pause and reconsider their words before posting."
In order to create the new system, Instagram partnered with suicide prevention programs.
Best-selling author and Smith College educator Rachel Simmons told “Good Morning America” that it was a great first step.
“We want to see social media platforms like Instagram stopping bullies before they start," she said.
The new tool is the latest arrow in Instagram’s anti-bullying quiver. In October, the app launched Restrict, a feature that allows users to easily shadow-ban other users who may post bullying or offensive comments.
©2019 New York Daily News Distributed by Tribune Content Agency, LLC.