1. robots.txt
Analysis of Current Status
- The site does have a robots.txt at the root.
- Currently, it allows all bots (
Allow: /), with no significant blocking, and it includes a sitemap. - No crawl delay or restrictions on sensitive/private areas or for duplicate content URLs.
- No mention of AI-specific bots like GPTBot, Gemini, ClaudeBot, etc.
Optimization Recommendations
- Location: Always at the root of the domain (
https://www.deltastream.io/robots.txt). - Structure:
- Allow all by default, unless there are private or sensitive folders to block.
- Explicitly allow modern AI user-agents (search, LLM, AI assistant bots).
- Keep the structure simple unless advanced segmentation or restrictions are needed.
- Include the
Sitemapdirective. - If no private/admin/test paths exist, avoid unneeded
Disallowentries.
- If there are private/admin folders or unwanted bots, specify them explicitly.
- Test regularly using Google Search Console’s robots.txt Tester.
Optimized robots.txt Example—No Private/Admin Paths
User-agent: *
Allow: /
# Explicit allow for major AI bots:
User-agent: Googlebot
Allow: /
User-agent: Gemini
Allow: /
User-agent: ClaudeBot
Allow: /
User-agent: anthropic-ai
Allow: /
User-agent: Claude-Web
Allow: /
User-agent: OAI-SearchBot/1.0; +https://openai.com/searchbot
Allow: /
User-agent: Mozilla/5.0 AppleWebKit/537.36 (KHTML, like Gecko); compatible; GPTBot/1.1; +https://openai.com/gptbot
Allow: /
Sitemap: https://www.deltastream.io/sitemap_index.xml
Optimized robots.txt Example—Blocking Private/Admin Folders
User-agent: *
Disallow: /admin/
Disallow: /private/
Allow: /
# AI and search bots
User-agent: Googlebot
Allow: /
User-agent: Gemini
Allow: /
User-agent: ClaudeBot
Allow: /
User-agent: anthropic-ai
Allow: /
User-agent: Claude-Web
Allow: /
User-agent: OAI-SearchBot/1.0; +https://openai.com/searchbot
Allow: /
User-agent: Mozilla/5.0 AppleWebKit/537.36 (KHTML, like Gecko); compatible; GPTBot/1.1; +https://openai.com/gptbot
Allow: /
Sitemap: https://www.deltastream.io/sitemap_index.xml
Keep this at the root of your domain (/robots.txt).
2. llms.txt
Analysis of Current Status
- No
llms.txtexists (404 Not Found) at https://www.deltastream.io/llms.txt. - This means AI bots seeking to ingest structured info about LLM-related services, capabilities, and contact/support/metadata receive nothing.
Optimization Recommendations
- Location: Place
llms.txtat the domain root (https://www.deltastream.io/llms.txt). - Content:
- Company name/description
- Core LLM integrations/support
- Use cases
- Documentation links
- Contact/support info
- Metadata: title, domain, language, keywords, etc.
- Formatting:
- Use clear headings, bullet lists, and concise descriptions for easy AI ingestion.
- Return HTTP status 200 OK.
Optimized llms.txt Example
# Deltastream—Large Language Model Services Manifest
Deltastream provides real-time data streaming and analytics, enhanced with AI and LLM integrations.
### Metadata
title: Deltastream | Real-Time Data Streaming Powered by AI
description: Deltastream offers real-time data analytics with deep integration of leading Large Language Models for advanced enrichment, Q&A, and summarization.
domain: www.deltastream.io
language: en
keywords: Deltastream, LLM, Streaming, Data Enrichment, Real-Time Analytics, AI, GPT-4, Gemini, Anthropic, Mistral, Cohere, Data Platform
### LLM Integrations Supported
- OpenAI GPT-4
- Google Gemini
- Anthropic Claude
- Mistral AI
- Cohere Command
- Databricks Dolly
### Use Cases
- Stream data enrichment
- Real-time analytics with LLM insights
- Automated summarization and Q&A
- Sentiment analysis
### Documentation
- API Docs: https://www.deltastream.io/docs/llms
- Quickstart: https://www.deltastream.io/quickstart
- Support: [email protected]
### Contact
- Join our Slack: https://join.slack.com/deltastream
- Email: [email protected]
### Accessibility & SEO
alt_text_present: true
structured_data: true
mobile_friendly: true
robots_txt: /robots.txt
Summary Table
| File | Location | Status | Optimization Actions |
|---|---|---|---|
| robots.txt | /robots.txt at root | Present | Add explicit AI bot entries if needed, keep simple, add sitemap |
| llms.txt | /llms.txt at root | MISSING | Create file with structured AI-friendly content and metadata |
Final Checklist
- Both files belong at the domain root.
- robots.txt: Allow all, add specific AI bots, restrict only as needed, include sitemap.
- llms.txt: MUST be created with high-quality, structured, markdown or plaintext content—following the template above.
- Test both files for accessibility (
HTTP 200) and structure.
If you have specific private URLs/folders or need AI-focused indexing adjustments, provide them for tailored rules.