NEXO 7.20.25 — local semantic models for Local Memory
Published 2026-05-17. Patch release over v7.20.24.
Why this release exists
Local Memory was warming the right local models, but Local Context search was still persisting and querying deterministic hash vectors. v7.20.25 connects the real local BGE embedding model to chunk indexing and query retrieval while keeping the hash path as a safe offline fallback.
What changed
- Real local embeddings. Chunks and queries now use the pinned local FastEmbed/BGE profile when installed, storing model id, revision and dimension with each vector.
- Automatic vector refresh. Existing hash embeddings are re-queued for lightweight refresh when the BGE profile is available, without re-reading the source files.
- Document-first indexing. PDF, Word/Office/OpenDocument/iWork-style files are discovered and extracted before lower-value text, email, code and unknown files.
- Local reranking. If the cross-encoder reranker is present locally, Local Context reranks the top candidate chunks without network downloads.
- Optional Qwen status. The Desktop-owned local-presence Qwen GGUF is reported as optional when absent from standalone Brain installs.
Validation
python3 -m pytest tests/test_local_context.py tests/test_local_models.py tests/test_model_warmup.py tests/test_local_context_cli.py tests/test_nexo_brain_onboarding_cli.py -q
# 88 passed
Full changelog entry → · Previous Local Memory profile write note