安全扫描
OpenClaw
安全
medium confidenceThe skill's code, declared requirements, and runtime instructions align with a straightforward web-scraping tool and do not request unrelated credentials or install artifacts, though part of the source shown was truncated so the full file should be reviewed before deployment.
评估建议
This skill appears to be a coherent, self-contained web scraper that doesn't request secrets or install external code. Before installing: (1) review the full src/index.ts (the provided snippet was truncated) to confirm there are no hidden network callbacks or logging endpoints; (2) ensure use complies with target sites' robots.txt, terms of service, and legal/privacy rules; (3) enforce rate limits and avoid scraping protected or paywalled content; (4) if you run in a sensitive environment, sandb...详细分析 ▾
✓ 用途与能力
Name/description (fetching, extracting text/links/metadata) match the actual tools and code: scrape_url, extract_links, scrape_multiple. No unrelated env vars, binaries, or services are requested.
✓ 指令范围
SKILL.md describes limited scraping actions (single page, link extraction, multi-page up to 10). Instructions recommend rate-limiting and content-size limits and do not instruct access to unrelated files, credentials, or external endpoints beyond the target pages.
✓ 安装机制
No install spec; skill is instruction-plus-code and relies on built-in Node fetch. No downloads, package registry installs, or archive extraction are present in the provided metadata.
✓ 凭证需求
Skill requires no environment variables, credentials, or config paths. The code uses only network fetch and in-memory parsing; requested access is proportional to web-scraping functionality.
✓ 持久化与权限
always is false and disable-model-invocation is false (normal). The skill does not request persistent system-wide privileges or modify other skills. Autonomous invocation is allowed by platform default but not combined with other red flags.
安全有层次,运行前请审查代码。
运行时依赖
无特殊依赖
版本
latestv0.2.112026/2/14
- Added SKILL.md documentation detailing available web scraping tools and usage guidelines. - Described supported actions: single-page scraping, link extraction, and multi-page scraping. - Clarified content formats (text, markdown, html) and their recommended uses. - Provided best practices for scraping frequency, file size limits, and saving results.
● 可疑
安装命令 点击复制
官方npx clawhub@latest install xpr-web-scraping
镜像加速npx clawhub@latest install xpr-web-scraping --registry https://cn.clawhub-mirror.com
技能文档
Web Scraping
You have web scraping tools for fetching and extracting data from web pages:
Single page:
scrape_url— fetch a URL and get cleaned text content + metadata (title, description, link count)
Link discovery:
extract_links— fetch a page and extract all links with text and type (internal/external)
pattern parameter to filter by regex (e.g. "\\.pdf$" for PDF links)
- Links are deduplicated and resolved to absolute URLsMulti-page research:
scrape_multiple— fetch up to 10 URLs in parallel for comparison/research
Best practices:
- Prefer "text" format for content extraction, "markdown" for preserving structure
- Don't scrape the same domain more than 5 times per minute
- Combine with
store_deliverableto save scraped content as job evidence - For very large pages, the content is limited to 5MB
数据来源:ClawHub ↗ · 中文优化:龙虾技能库
OpenClaw 技能定制 / 插件定制 / 私有工作流定制
免费技能或插件可能存在安全风险,如需更匹配、更安全的方案,建议联系付费定制