Published
Dec 18, 2024
Updated
Dec 18, 2024

How Google Uses AI to Zap Bad Ads

Zero-Shot Image Moderation in Google Ads with LLM-Assisted Textual Descriptions and Cross-modal Co-embeddings
By
Enming Luo|Wei Qiao|Katie Warren|Jingxiang Li|Eric Xiao|Krishna Viswanathan|Yuan Wang|Yintao Liu|Jimin Li|Ariel Fuxman

Summary

Ever wonder how Google keeps inappropriate ads off your screen? It's a monumental task, given the sheer volume of ads they process daily. A new research paper reveals Google's secret weapon: a cutting-edge AI system that can spot bad ads *before* they even go live. Instead of relying on traditional methods of training AI with countless labeled examples of bad ads (a time-consuming and resource-intensive process), Google's researchers have developed a clever shortcut. They use large language models (LLMs), the same technology that powers ChatGPT, to generate detailed text descriptions of what constitutes a policy violation. These descriptions are then turned into “co-embeddings,” mathematical representations that capture the semantic relationship between images and text. When a new ad image comes in, the system checks its similarity to these policy-violation co-embeddings. If it matches closely with descriptions of banned content (like tobacco products or misleading imagery), it’s flagged for further review or outright blocked. This allows for “zero-shot” moderation – meaning the system can identify violations without ever having specifically seen them before. The system is also surprisingly agile. Updating policies is as simple as tweaking the text descriptions, no lengthy retraining required. Early tests show this new approach catches over 100% more bad ads than previous methods, maintaining a high level of accuracy. This clever use of LLMs and co-embeddings is a game-changer for online safety. It allows Google to stay one step ahead of bad actors while maintaining a vibrant and trustworthy ad ecosystem. While the research focuses on ads, this technology holds exciting potential for other areas of content moderation, promising a safer and cleaner online experience for all.
🍰 Interesting in building your own agents?
PromptLayer provides the tools to manage and monitor prompts with your whole team. Get started for free.

Question & Answers

How does Google's co-embedding system technically work to identify policy-violating ads?
Google's co-embedding system converts both policy descriptions and ad images into mathematical representations that exist in the same semantic space. The process works in three main steps: First, large language models generate detailed text descriptions of policy violations. Second, these descriptions are transformed into co-embeddings - mathematical vectors that capture semantic meaning. Finally, when new ad images arrive, the system computes their embeddings and measures similarity to the policy violation embeddings. If the similarity score exceeds a threshold, the ad is flagged. For example, if an ad image's embedding closely matches the embedding for 'tobacco products,' it would be automatically flagged for review or blocked.
What are the main benefits of AI-powered content moderation for businesses?
AI-powered content moderation offers three key advantages for businesses. First, it provides scalable, real-time protection against inappropriate content without requiring massive human moderation teams. Second, it's highly adaptable - policies can be updated quickly without extensive system retraining. Third, it's proactive rather than reactive, catching potential violations before they reach users. This technology can benefit various industries, from social media platforms to e-commerce sites, helping maintain brand safety while reducing operational costs. For instance, an online marketplace could automatically screen product listings for counterfeit items or misleading claims.
How is AI changing the future of online advertising?
AI is revolutionizing online advertising by making it safer, more efficient, and more trustworthy. It enables automatic screening of millions of ads in real-time, ensuring compliance with policies and protecting users from harmful or misleading content. The technology also allows for quick adaptation to new types of policy violations or emerging threats. This creates a more reliable advertising ecosystem where businesses can confidently invest in digital marketing, knowing their ads will appear alongside appropriate content. For users, this means a better online experience with fewer unwanted or potentially harmful advertisements.

PromptLayer Features

  1. Prompt Management
  2. Similar to how Google manages policy violation descriptions, PromptLayer can version and manage text prompts that define content guidelines
Implementation Details
Store policy descriptions as versioned prompts, use template system for consistent formatting, implement access controls for policy updates
Key Benefits
• Centralized policy management • Version history tracking • Controlled collaborative editing
Potential Improvements
• Add semantic search for similar policies • Implement policy conflict detection • Create policy template suggestions
Business Value
Efficiency Gains
Reduce policy update time by 70% through centralized management
Cost Savings
Minimize errors and rework through version control
Quality Improvement
Ensure consistency across all content policies
  1. Testing & Evaluation
  2. Like Google's system testing against policy violations, PromptLayer can facilitate batch testing of content against moderation rules
Implementation Details
Create test suites for policy compliance, implement A/B testing for policy effectiveness, establish evaluation metrics
Key Benefits
• Automated compliance testing • Performance benchmarking • Policy effectiveness measurement
Potential Improvements
• Add real-time policy testing • Implement automated regression testing • Create policy impact analysis tools
Business Value
Efficiency Gains
Reduce policy testing time by 80% through automation
Cost Savings
Prevent costly policy violations through proactive testing
Quality Improvement
Maintain 99.9% accuracy in content moderation

The first platform built for prompt engineering