-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathrun_all.sh
More file actions
executable file
·54 lines (44 loc) · 1.13 KB
/
run_all.sh
File metadata and controls
executable file
·54 lines (44 loc) · 1.13 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
#!/usr/bin/env bash
set -euo pipefail
MODEL="qwen3:8b"
OLLAMA_URL="http://localhost:11434"
echo "=== LLM Agent Benchmark ==="
echo ""
# 1. Check Ollama is running
echo -n "Checking Ollama... "
if ! curl -sf "${OLLAMA_URL}/api/tags" > /dev/null; then
echo "NOT running"
echo "Start Ollama first: ollama serve"
exit 1
fi
echo "OK"
# 2. Check / pull model
echo -n "Checking model ${MODEL}... "
if ollama list 2>/dev/null | grep -q "^${MODEL}"; then
echo "already present"
else
echo "not found - pulling..."
ollama pull "${MODEL}"
fi
# 3. Pre-install all dependencies (warms up uv venv cache)
echo ""
echo "--- Pre-installing dependencies ---"
for script in tool_calling_test.py deepagents_test.py adk_test.py pydanticai_test.py; do
echo -n " ${script} ... "
BENCH_WARMUP=1 uv run "${script}"
done
# 4. Run benchmarks
echo ""
echo "--- Raw Ollama ---"
uv run tool_calling_test.py
echo ""
echo "--- Deep Agents ---"
uv run deepagents_test.py
echo ""
echo "--- Google ADK ---"
uv run adk_test.py
echo ""
echo "--- Pydantic AI ---"
uv run pydanticai_test.py
echo ""
echo "=== Done! Results saved to bench_results.log ==="