跳至主要内容
小龙虾小龙虾AI
🤖

Ping Model

Measure and display AI model response latency. Use when the user types /ping or /ping followed by a model name to test round-trip time. Captures precise timi...

下载38
星标0
版本1.0.0
开发工具
安全通过
💬Prompt

技能说明


name: ping-model description: Measure and display AI model response latency. Use when the user types /ping or /ping followed by a model name to test round-trip time. Captures precise timing between command receipt and response generation, with smart duration formatting (ms, seconds, or minutes). Supports cross-model testing by temporarily switching models and measuring latency. metadata: {"clawdbot":{"emoji":"🧪","requires":{"bins":["node"]}}}

Ping Model

Measure AI model response latency with consistent formatting.

Quick Start

Simple Ping (current model)

bash command:"node {baseDir}/ping-model.js"

Ping Specific Model

bash command:"node {baseDir}/ping-model.js --model minimax"

Compare Multiple Models

bash command:"node {baseDir}/ping-model.js --compare kimi,minimax,deepseek"

Command Reference

CommandDescription
/pingPing current active model
/ping kimiSwitch to kimi, ping, return
/ping minimaxSwitch to minimax, ping, return
/ping deepseekSwitch to deepseek, ping, return
/ping allCompare all available models

Output Format

Required format - ALWAYS use this exact structure:

🧪 PING {model-name}

📤 Sent:     {HH:MM:SS.mmm}
📥 Received: {HH:MM:SS.mmm}
⏱️  Latency:  {formatted-duration}

🎯 Pong!

Latency Formatting Rules

  • < 1 second: Display as XXXms (e.g., 847ms)
  • ≥ 1 second, < 60 seconds: Display as X.XXs (e.g., 1.23s)
  • ≥ 60 seconds: Display as X.XXmin (e.g., 2.5min)

Examples

Fast response (< 1s):

🧪 PING kimi

📤 Sent:     09:34:15.123
📥 Received: 09:34:15.247
⏱️  Latency:  124ms

🎯 Pong!

Medium response (1-60s):

🧪 PING minimax

📤 Sent:     09:34:15.123
📥 Received: 09:34:16.456
⏱️  Latency:  1.33s

🎯 Pong!

Slow response (> 60s):

🧪 PING gemini

📤 Sent:     09:34:15.123
📥 Received: 09:35:25.456
⏱️  Latency:  1.17min

🎯 Pong!

Cross-Model Testing

When testing a non-active model:

  1. Save current model context
  2. Switch to target model
  3. Execute ping
  4. Measure latency
  5. Restore original model
  6. Display result

Critical: Always return to the original model after testing.

Comparison Mode

bash command:"node {baseDir}/ping-model.js --compare kimi,minimax,deepseek,gpt"

Output format:

══════════════════════════════════════════════════
🧪 MODEL COMPARISON
══════════════════════════════════════════════════

🥇 kimi      124ms
🥈 minimax   1.33s
🥉 deepseek  2.45s
4️⃣  openai    5.67s

🏆 Fastest: kimi (124ms)

Implementation

The ping latency is measured as the time between:

  • T1: Message received by the agent
  • T2: Response ready to send

This captures the model's internal processing time, not network latency.

如何使用「Ping Model」?

  1. 打开小龙虾AI(Web 或 iOS App)
  2. 点击上方「立即使用」按钮,或在对话框中输入任务描述
  3. 小龙虾AI 会自动匹配并调用「Ping Model」技能完成任务
  4. 结果即时呈现,支持继续对话优化

相关技能