Is Multilingual LLM Watermarking Truly Multilingual? Scaling Robustness to 100+ Languages via Back-Translation
📰 ArXiv cs.AI
Existing multilingual LLM watermarking methods are not truly multilingual and fail to remain robust under translation attacks in medium- and low-resource languages
Action Steps
- Identify the limitations of current multilingual watermarking methods
- Evaluate the robustness of these methods in medium- and low-resource languages
- Use back-translation to scale robustness to 100+ languages
- Develop new methods that can effectively watermark LLM outputs across languages
Who Needs to Know This
ML researchers and engineers working on LLMs and watermarking techniques can benefit from this research to improve the robustness of their models across multiple languages
Key Insight
💡 Current multilingual watermarking methods are not truly multilingual and require improvement to scale to 100+ languages
Share This
🚨 Multilingual LLM watermarking methods are not as robust as claimed, especially in medium- and low-resource languages 🚨
DeepCamp AI