forked from Ishabdullah/Codey-v2
-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathrequirements.txt
More file actions
83 lines (75 loc) · 3.62 KB
/
requirements.txt
File metadata and controls
83 lines (75 loc) · 3.62 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
# ── Core (required for agent, GUI, and daemon) ─────────────────────────────
# install.sh installs these automatically.
# On Termux, run first: pkg install python-pyarrow python-pandas
rich>=14.0.0
numpy>=1.24.0
watchdog>=3.0.0
aiohttp>=3.9.0 # GUI server + HTTP inference calls
requests>=2.31.0
httpx>=0.27.0
pyyaml>=6.0
filelock>=3.13.0
tqdm>=4.65.0
hnswlib>=0.7.0 # vector search (falls back to numpy if missing)
pytest>=9.0.0 # optional: run tests with pytest tests/
# LLM inference — we use the llama-server binary, not the Python binding
# llama-cpp-python>=0.2.50 # uncomment only if you need the Python API directly
# Phase 4: Hierarchical Memory
# sentence-transformers>=2.2.0 # Requires torch - not available on Termux/Android
# We use llama.cpp embeddings instead (see embed_server.py)
numpy>=1.24.0
# Phase 5: Background Execution
watchdog>=3.0.0
# Phase 6: Observability (optional — sysmon.py falls back to /proc/ on Android)
# psutil>=5.9.0 # Skip on Termux/Android — wheel fails to build, not needed
# ──────────────────────────────────────────────────────────────────────────────
# Training Data Pipeline (pipeline/)
# ──────────────────────────────────────────────────────────────────────────────
# The pipeline builds fine-tuning datasets from HuggingFace + synthetic data.
# See docs/pipeline.md for the full guide.
#
# IMPORTANT — Termux/Android install order:
#
# Step 1 — Install compiled packages via pkg (pip cannot build these on ARM):
# pkg install python-pyarrow python-pandas
#
# Step 2 — Install pure-Python packages via pip:
# pip install datasets huggingface-hub "fsspec==2026.2.0" \
# httpcore httpx typer tqdm hnswlib \
# aiohttp multiprocess dill xxhash pyyaml \
# filelock requests
#
# On desktop Linux, you can install everything via pip (no pkg step needed).
#
# Notes:
# - fsspec must be pinned to 2026.2.0; datasets 4.8.4 is incompatible with
# newer versions (ImportError at runtime if mismatched).
# - hf-xet build failures are harmless — it is an optional Rust extension
# used only for HuggingFace uploads; ignore any pip build error for it.
# - hnswlib provides fast ANN vector search and builds from source via pip
# (clang must be installed first on Termux: pkg install clang cmake).
# If hnswlib fails to install, the pipeline falls back to numpy brute-force
# cosine search automatically — no action required.
# - sentence-transformers is an optional fallback embedding backend (384-dim
# MiniLM). It requires PyTorch, which is not available on Termux/Android.
# The pipeline defaults to the nomic-embed-text llama-server on port 8082.
# Pipeline — pip-installable dependencies (see install order note above)
datasets>=2.14.0
huggingface-hub>=0.23.0
fsspec==2026.2.0
httpcore>=1.0.0
httpx>=0.27.0
typer>=0.9.0
tqdm>=4.65.0
hnswlib>=0.7.0
aiohttp>=3.9.0
multiprocess>=0.70.0
dill>=0.3.7
xxhash>=3.4.0
pyyaml>=6.0
filelock>=3.13.0
requests>=2.31.0
# Pipeline — must be installed via pkg on Termux (pre-built ARM packages):
# pkg install python-pyarrow python-pandas
# pyarrow>=14.0.0 # DO NOT pip install on Termux — build fails on aarch64
# pandas>=2.0.0 # DO NOT pip install on Termux — build fails on aarch64