The surge of low-quality data online has led to potentially harmful knowledge instilled in Large Language Models (LLMs). This problem elevates risks when LLMs are deployed in chatbots that might expose users to harmful advice or aggressive interactions. Existing toxicity evaluation datasets focus mainly on English, limiting their capability to detect multilingual toxicity which compromises…