Robots & Indexing Troubleshooting
Common issues and solutions for robots.txt, indexing, and content protection
Common Issues Quick Reference
Robots.txt Issues
Robots.txt Issues
Physical File Conflict
Problem:
Warning message: "A physical robots.txt file exists in your site root"
Solution:
- Access your site via FTP or file manager
- Navigate to your WordPress root directory (where wp-config.php is located)
- Delete or rename the robots.txt file
- Clear any caching plugins
- Verify at yoursite.com/robots.txt
# Via SSH/Terminal
cd /path/to/wordpress/root
mv robots.txt robots.txt.backup
# Or remove it
rm robots.txtRules Not Taking Effect
Common Causes:
- Caching at server or CDN level
- Virtual editor not enabled
- Syntax errors in rules
- Conflicting plugins
Solutions:
- Clear all caches (WordPress, CDN, browser)
- Verify "Enable Virtual Robots.txt Editor" is ON
- Check rules syntax:
# Correct syntax
User-agent: *
Disallow: /private/
# Incorrect syntax
User-agent * # Missing colon
Disallow /private/ # Missing colon- Deactivate other SEO plugins temporarily
- Check .htaccess for conflicting rules
Sitemap Not Appearing in Robots.txt
Checklist:
- Ensure XML Sitemaps are enabled in ProRank SEO → Technical SEO → Sitemaps
- Verify sitemap exists at yoursite.com/sitemap_index.xml
- Check "Auto-add sitemap to robots.txt" is enabled
- Save robots.txt settings again to regenerate
Manual Addition:
If auto-add fails, manually add to your robots.txt rules:
# Add at the bottom of your rules
Sitemap: https://yoursite.com/sitemap_index.xmlAI Bots Still Crawling Despite Blocking
Verification Steps:
- Check robots.txt shows AI bot rules:
User-agent: GPTBot
Disallow: /
User-agent: ChatGPT-User
Disallow: /- Verify meta tags are present in page source:
<meta name="robots" content="noai, noimageai" />- Check server logs for bot user agents
- Note: Some bots may not respect robots.txt immediately
Additional Measures:
- Enable Content Safeguard for noai directives
- Consider IP blocking at server level for persistent bots
- Use .htaccess rules as additional layer
Indexing API Issues
IndexNow Not Working
Common Issues:
- Missing API Key:Generate a new key in settings and save
- Key file not accessible:Verify key file exists at yoursite.com/[your-key].txt
- Invalid URL format:Ensure URLs use https:// and are publicly accessible
- Rate limiting:IndexNow has daily limits - check submission log
Debugging Steps:
# Test IndexNow manually
curl -X POST "https://api.indexnow.org/indexnow" \
-H "Content-Type: application/json" \
-d '{
"host": "yoursite.com",
"key": "your-api-key",
"urlList": ["https://yoursite.com/test-page"]
}'Response Codes:
200- Success202- Accepted, processing400- Invalid format403- Invalid key429- Rate limited
Google Indexing API Errors
Service Account Issues:
- Verify JSON key file is valid:
{
"type": "service_account",
"project_id": "your-project",
"private_key_id": "...",
"private_key": "-----BEGIN PRIVATE KEY-----\n...\n-----END PRIVATE KEY-----\n",
"client_email": "indexing@your-project.iam.gserviceaccount.com"
}- Add service account email to Search Console as owner
- Enable Indexing API in Google Cloud Console
- Check API quotas (200 URLs/day limit)
Common Error Messages:
Permission denied→ Service account not added to Search ConsoleInvalid grant→ JSON key file corrupted or expiredQuota exceeded→ Daily limit reached (200 URLs)API not enabled→ Enable API in Google Cloud ConsoleMeta Tags & Noindex Issues
Pages Still Appearing in Search After Noindex
Understanding the Delay:
Search engines need to recrawl pages to discover noindex tags. This can take days or weeks.
Speed Up Deindexing:
- Use Google Search Console URL Removal tool for urgent cases
- Submit URL to IndexNow for faster recrawling
- Update sitemap to trigger recrawl
- Ensure page returns 200 status (not 404) with noindex tag
Verification:
# Check if noindex is present
curl -I https://yoursite.com/page | grep -i "x-robots-tag"
# Or check HTML source for:
# <meta name="robots" content="noindex" />Content Safeguard Issues
AI Protection Not Working
Verification Steps:
- Check page source for noai/noimageai tags:
<!-- Should see this in <head> -->
<meta name="robots" content="index, follow, noai, noimageai" />- For Pro+ users, check HTTP headers:
curl -I https://yoursite.com/page
# Should include:
# X-Robots-Tag: noai, noimageaiIf Tags Missing:
- Verify Content Safeguard is enabled
- Check license tier (Core+ or Pro+ required)
- Clear all caches
- Check for theme/plugin conflicts
Performance Issues
Slow Loading or Timeouts
Large Robots.txt File:
If blocking many AI bots, robots.txt can become large. Optimize by:
- Using wildcard patterns where possible
- Combining similar bot rules
- Removing outdated bot entries
API Rate Limiting:
Prevent API exhaustion:
- Batch IndexNow submissions (max 10,000 URLs per request)
- Implement submission queue for Google API (200/day limit)
- Use cron jobs for scheduled submissions
Diagnostic Tools
Useful Testing Tools
ProRank SEO Built-in
- • Robots.txt preview
- • IndexNow submission log
- • Meta tag inspector
- • Conflict detector
External Tools
- • Google Search Console robots.txt tester
- • Bing Webmaster Tools
- • Technical SEO crawlers (Screaming Frog)
- • Browser DevTools Network tab
Quick Diagnostic Commands:
# Check robots.txt
curl https://yoursite.com/robots.txt
# Check meta tags
curl -s https://yoursite.com | grep -i "meta.*robots"
# Check HTTP headers
curl -I https://yoursite.com | grep -i "x-robots"
# Test IndexNow key file
curl https://yoursite.com/[your-key].txtEmergency Fixes
Critical Issues Quick Fixes
Site Blocked from Search Engines:
- Immediately disable virtual robots.txt editor
- Remove any physical robots.txt file
- Clear all caches
- Submit sitemap to Search Console
- Use URL Inspection tool to request indexing
Complete Feature Reset:
If all else fails, reset the module:
-- Backup first, then run in database:
DELETE FROM wp_options WHERE option_name LIKE 'prorank_robots%';
DELETE FROM wp_options WHERE option_name LIKE 'prorank_indexing%';
-- Then reactivate and reconfigure in adminAlways backup your database before running SQL commands. After reset, you'll need to reconfigure all Robots & Indexing settings.
Getting Help
Before Contacting Support
- ✓ Check WordPress and plugin versions
- ✓ Test with default theme
- ✓ Disable other SEO plugins
- ✓ Clear all caches
- ✓ Check error logs
- ✓ Document steps to reproduce
Information to Provide
- • ProRank SEO version and license
- • WordPress version
- • Active theme and plugins
- • Server environment (PHP version)
- • Error messages or logs
- • Screenshots of settings