
內容簡介
總結:這個 WordPress 外掛 Crawler Record 記錄特定使用者代理(Google、Bing、OpenAI、Anthropic、Perplexity、DuckDuckGo,以及你自己)上次訪問你內容的時間,提供方便的管理和隱私保護功能。
1. 這個外掛能夠記錄哪些特定的使用者代理訪問了你的內容?
- Google、Bing、OpenAI(GPTBot/OAI-SearchBot/ChatGPT-User)、Anthropic(ClaudeBot/Claude-User/Claude-SearchBot)、Perplexity、DuckDuckGo,以及可自訂的選項。
2. 外掛如何評估 Allow/Disallow 規則以及處理 robots.txt?
- 外掛在本地渲染 robots.txt(實體文件或通過 do_robots 提供的 WordPress 虛擬輸出),並評估給定路徑的 Allow/Disallow 規則。如果啟用「Settings → Reading → Disourage search engines」(blog_public = 0),所有代理都會被列為被阻擋。
3. 這個外掛如何提供使用者介面?
- 提供清晰的界面,包含分組節段、微秒時間戳記,以及一個小的診斷開關,展示匹配的機器人組和規則。
4. 外掛有哪些已知限制?
- robots.txt 的通配符(*)和行尾標記($)不被解釋;匹配僅為基於前綴。未來版本可能會添加完整的規格支持。
5. 這個外掛如何確保使用者隱私?
- 外掛僅保存機器人訪問時間戳記和最後瀏覽的 URL,沒有收集或存儲有關網站訪客的個人數據。沒有數據傳送給第三方。
6.外掛的授權是什麼?
- GPLv2 或之後版本。請參見 LICENSE 文件。
外掛標籤
開發者團隊
原文外掛簡介
Crawler Record logs the last time specific user agents (for search and AI chat/LLMs) accessed your content including:
Google
Bing
ChatGPT (OpenAI)
Claude (Anthropic)
Perplexity
DuckDuckGo
Meta
Apple
You can view this information:
Admin page: a grouped list of all crawlers, with last seen time, last page, and robots status for the crawler.
Admin Bar (front-end): quick-glance status for the current singular or URL context (no dropdowns in wp-admin).
Per Page/Post: From the edit screen on any page or post, you can see that last time each crawler visited that page.
Robots-aware: The plugin checks your robots.txt and evaluates Allow/Disallow rules for a given path. If Settings → Reading → “Discourage search engines” is enabled, all agents are shown as blocked with a prominent warning.
Performance-friendly by design: Write-throttling (default 10 minutes) and an auxiliary “last post ID per agent” record avoids heavy admin queries on large sites.
Privacy-friendly: Saves only bot visit timestamps and last URLs crawled — no personal data.
Learn how to use this plugin.
Highlights
Supports common user-agent variants for Google, Bing, OpenAI (ChatGPT), Anthropic (Claude), Perplexity, Meta, Apple, and DuckDuckGo.
Robots status computed from local robots.txt (physical or virtual) without outbound requests.
Clear UI with grouped sections, microsecond timestamps, and a small diagnostics toggle showing the matched robots group and rule.
Known limitations
robots.txt wildcards (*) and end-of-line marker ($) are not interpreted; matching is prefix-based only. Future versions may add full spec support.
Privacy
This plugin stores:
– Timestamps of crawler visits (float, with microseconds)
– Last URL seen per crawler (per-URL records)
– Last post ID per crawler (for admin performance)
It does not collect or store personal data about site visitors. No data is transmitted to third parties.
License
GPLv2 or later. See LICENSE file.
