This workflow identifies toxic language (such as profanity, insults, or threats) in messages sent via Telegram.
Telegram Trigger: Initiates the workflow when a new message is posted in a Telegram chat.
Google Perspective: Analyzes the message text and provides a probability score between 0 and 1 indicating the likelihood of toxic content.
IF: Filters messages with a toxic probability score above 0.7.
Telegram: Posts a warning message in the chat saying "I don't tolerate toxic language" if the probability score is above 0.7.
NoOp: No action is taken if the probability score is below 0.7.
Your connected stack awaits to automate AI workflows with 24-7 uptime performance and engagement
"I highly recommend Rantir, they are a great dev team with quick turn around on all projects and requests. We recently worked with them on updating our website and any changes, updates or modifications I needed were always taken care of quickly!"
"The team at Rantir has lived up to every definition of the word "partner". They're adaptive, fast, and flexible (all the things you'd hope for). We're so thrilled with what we've accomplished so far and look forward to working alongside them in the future."
"Working with the Rantir team was a pleasure. They guided us through the whole process from design to implementation, creating a great site on a tight deadline. They were responsive and adaptable throughout, and we'd be happy to work with them again in the future."
"Working with the Rantir team early on made combined design and development with early conversations to implement AI within Onder. We were happy to work together to help bring no-code, with code and AI."
Rantir University for learning how to build powerful AI Agents & Software you own.