Files
LEANN/packages
ww2283 d226f72bc0 feat: implement true batch processing for Ollama embeddings
Migrate from deprecated /api/embeddings to modern /api/embed endpoint
which supports batch inputs. This reduces HTTP overhead by sending
32 texts per request instead of making individual API calls.

Changes:
- Update endpoint from /api/embeddings to /api/embed
- Change parameter from 'prompt' (single) to 'input' (array)
- Update response parsing for batch embeddings array
- Increase timeout to 60s for batch processing
- Improve error handling for batch requests

Performance:
- Reduces API calls by 32x (batch size)
- Eliminates HTTP connection overhead per text
- Note: Ollama still processes batch items sequentially internally

Related: #151
2025-10-25 10:58:15 -04:00
..
2025-09-19 17:03:55 -07:00
2025-09-16 20:45:22 +00:00
2025-09-24 11:19:04 -07:00
2025-08-14 14:39:45 -07:00
2025-07-27 02:22:54 -07:00