fix: mem compare
This commit is contained in:
@@ -10,6 +10,7 @@ import time
|
|||||||
import psutil
|
import psutil
|
||||||
import gc
|
import gc
|
||||||
import subprocess
|
import subprocess
|
||||||
|
from pathlib import Path
|
||||||
|
|
||||||
# Setup logging
|
# Setup logging
|
||||||
logging.basicConfig(stream=sys.stdout, level=logging.INFO)
|
logging.basicConfig(stream=sys.stdout, level=logging.INFO)
|
||||||
@@ -95,124 +96,39 @@ def test_faiss_hnsw():
|
|||||||
|
|
||||||
|
|
||||||
def test_leann_hnsw():
|
def test_leann_hnsw():
|
||||||
"""Test LEANN HNSW Backend"""
|
"""Test LEANN HNSW Search Memory (load existing index)"""
|
||||||
print("\n" + "=" * 50)
|
print("\n" + "=" * 50)
|
||||||
print("TESTING LEANN HNSW BACKEND")
|
print("TESTING LEANN HNSW SEARCH MEMORY")
|
||||||
print("=" * 50)
|
print("=" * 50)
|
||||||
|
|
||||||
tracker = MemoryTracker("LEANN HNSW")
|
tracker = MemoryTracker("LEANN HNSW Search")
|
||||||
|
|
||||||
# Import and setup
|
# Import and setup
|
||||||
tracker.checkpoint("Initial")
|
tracker.checkpoint("Initial")
|
||||||
|
|
||||||
from llama_index.core import SimpleDirectoryReader
|
from leann.api import LeannSearcher
|
||||||
from llama_index.core.node_parser import SentenceSplitter
|
|
||||||
from leann.api import LeannBuilder, LeannChat
|
|
||||||
from pathlib import Path
|
|
||||||
|
|
||||||
tracker.checkpoint("After imports")
|
tracker.checkpoint("After imports")
|
||||||
|
|
||||||
# Load and parse documents
|
# Find existing LEANN index
|
||||||
documents = SimpleDirectoryReader(
|
index_paths = [
|
||||||
"examples/data",
|
"./test_leann_comparison/comparison.leann",
|
||||||
recursive=True,
|
]
|
||||||
encoding="utf-8",
|
index_path = None
|
||||||
required_exts=[".pdf", ".txt", ".md"],
|
for path in index_paths:
|
||||||
).load_data()
|
if os.path.exists(path + ".meta.json"):
|
||||||
|
index_path = path
|
||||||
|
break
|
||||||
|
|
||||||
tracker.checkpoint("After document loading")
|
if not index_path:
|
||||||
|
print("❌ LEANN index not found. Please build it first")
|
||||||
|
return {"peak_memory": float("inf"), "error": "Index not found"}
|
||||||
|
|
||||||
# Parse into chunks
|
# Load searcher
|
||||||
node_parser = SentenceSplitter(
|
searcher = LeannSearcher(index_path)
|
||||||
chunk_size=256, chunk_overlap=20, separator=" ", paragraph_separator="\n\n"
|
tracker.checkpoint("After searcher loading")
|
||||||
)
|
|
||||||
|
|
||||||
all_texts = []
|
print("Running search queries...")
|
||||||
for doc in documents:
|
|
||||||
nodes = node_parser.get_nodes_from_documents([doc])
|
|
||||||
for node in nodes:
|
|
||||||
all_texts.append(node.get_content())
|
|
||||||
|
|
||||||
tracker.checkpoint("After text chunking")
|
|
||||||
|
|
||||||
# Build LEANN index
|
|
||||||
INDEX_DIR = Path("./test_leann_comparison")
|
|
||||||
INDEX_PATH = str(INDEX_DIR / "comparison.leann")
|
|
||||||
|
|
||||||
# Clean up previous index
|
|
||||||
import shutil
|
|
||||||
|
|
||||||
if INDEX_DIR.exists():
|
|
||||||
shutil.rmtree(INDEX_DIR)
|
|
||||||
|
|
||||||
builder = LeannBuilder(
|
|
||||||
backend_name="hnsw",
|
|
||||||
embedding_model="facebook/contriever",
|
|
||||||
graph_degree=32,
|
|
||||||
complexity=64,
|
|
||||||
is_compact=True,
|
|
||||||
is_recompute=True,
|
|
||||||
num_threads=1,
|
|
||||||
)
|
|
||||||
|
|
||||||
tracker.checkpoint("After builder setup")
|
|
||||||
|
|
||||||
print("Building LEANN HNSW index...")
|
|
||||||
|
|
||||||
for chunk_text in all_texts:
|
|
||||||
builder.add_text(chunk_text)
|
|
||||||
|
|
||||||
builder.build_index(INDEX_PATH)
|
|
||||||
|
|
||||||
tracker.checkpoint("After index building")
|
|
||||||
|
|
||||||
# Ensure any existing embedding server is updated with new data
|
|
||||||
print("Ensuring embedding server data synchronization...")
|
|
||||||
meta_file = INDEX_DIR / "comparison.leann.meta.json"
|
|
||||||
if meta_file.exists():
|
|
||||||
from leann.embedding_server_manager import (
|
|
||||||
_check_port,
|
|
||||||
_update_server_meta_path,
|
|
||||||
_check_server_meta_path,
|
|
||||||
)
|
|
||||||
|
|
||||||
port = 5557 # Default port for HNSW backend
|
|
||||||
if _check_port(port):
|
|
||||||
print(f"Updating server meta path to: {meta_file}")
|
|
||||||
success = _update_server_meta_path(port, str(meta_file))
|
|
||||||
if success:
|
|
||||||
print("✅ Server meta path updated successfully")
|
|
||||||
else:
|
|
||||||
print("❌ Failed to update server meta path")
|
|
||||||
|
|
||||||
# Verify the update and ensure server is ready
|
|
||||||
time.sleep(2) # Give server time to reload data
|
|
||||||
max_retries = 5
|
|
||||||
for retry in range(max_retries):
|
|
||||||
if _check_server_meta_path(port, str(meta_file)):
|
|
||||||
print("✅ Server meta path verification successful")
|
|
||||||
break
|
|
||||||
else:
|
|
||||||
print(
|
|
||||||
f"⏳ Server meta path verification failed (attempt {retry + 1}/{max_retries})"
|
|
||||||
)
|
|
||||||
if retry < max_retries - 1:
|
|
||||||
time.sleep(1)
|
|
||||||
else:
|
|
||||||
print(
|
|
||||||
"❌ Server meta path verification failed after all retries - may cause query issues"
|
|
||||||
)
|
|
||||||
else:
|
|
||||||
print("No existing server found on port 5557")
|
|
||||||
|
|
||||||
# Test queries
|
|
||||||
chat = LeannChat(
|
|
||||||
index_path=INDEX_PATH, llm_config={"type": "simulated", "model": "test"}
|
|
||||||
)
|
|
||||||
|
|
||||||
tracker.checkpoint("After chat setup")
|
|
||||||
|
|
||||||
print("Running queries...")
|
|
||||||
queries = [
|
queries = [
|
||||||
"什么是盘古大模型以及盘古开发过程中遇到了什么阴暗面,任务令一般在什么城市颁发",
|
"什么是盘古大模型以及盘古开发过程中遇到了什么阴暗面,任务令一般在什么城市颁发",
|
||||||
"What is LEANN and how does it work?",
|
"What is LEANN and how does it work?",
|
||||||
@@ -221,18 +137,20 @@ def test_leann_hnsw():
|
|||||||
|
|
||||||
for i, query in enumerate(queries):
|
for i, query in enumerate(queries):
|
||||||
start_time = time.time()
|
start_time = time.time()
|
||||||
_ = chat.ask(query, top_k=20, recompute_beighbor_embeddings=True, complexity=32)
|
# Use same parameters as Faiss: top_k=20, ef=120 (complexity parameter)
|
||||||
|
_ = searcher.search(query, top_k=20, ef=120)
|
||||||
query_time = time.time() - start_time
|
query_time = time.time() - start_time
|
||||||
print(f"Query {i + 1} time: {query_time:.3f}s")
|
print(f"Query {i + 1} time: {query_time:.3f}s")
|
||||||
tracker.checkpoint(f"After query {i + 1}")
|
tracker.checkpoint(f"After query {i + 1}")
|
||||||
|
|
||||||
peak_memory = tracker.summary()
|
peak_memory = tracker.summary()
|
||||||
|
|
||||||
# Get storage size before cleanup - only index files (exclude text data)
|
# Get storage size before cleanup
|
||||||
storage_size = 0
|
storage_size = 0
|
||||||
|
INDEX_DIR = Path(index_path).parent
|
||||||
if INDEX_DIR.exists():
|
if INDEX_DIR.exists():
|
||||||
total_size = 0
|
total_size = 0
|
||||||
for dirpath, dirnames, filenames in os.walk(str(INDEX_DIR)):
|
for dirpath, _, filenames in os.walk(str(INDEX_DIR)):
|
||||||
for filename in filenames:
|
for filename in filenames:
|
||||||
# Only count actual index files, skip text data and backups
|
# Only count actual index files, skip text data and backups
|
||||||
if filename.endswith((".old", ".tmp", ".bak", ".jsonl", ".json")):
|
if filename.endswith((".old", ".tmp", ".bak", ".jsonl", ".json")):
|
||||||
@@ -243,20 +161,19 @@ def test_leann_hnsw():
|
|||||||
total_size += os.path.getsize(filepath)
|
total_size += os.path.getsize(filepath)
|
||||||
storage_size = total_size / (1024 * 1024) # Convert to MB
|
storage_size = total_size / (1024 * 1024) # Convert to MB
|
||||||
|
|
||||||
# Clean up (but keep directory for storage size comparison)
|
# Clean up
|
||||||
del chat, builder
|
del searcher
|
||||||
gc.collect()
|
gc.collect()
|
||||||
|
|
||||||
return {
|
return {
|
||||||
"peak_memory": peak_memory,
|
"peak_memory": peak_memory,
|
||||||
"storage_size": storage_size,
|
"storage_size": storage_size,
|
||||||
"tracker": tracker,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
"""Run comparison tests"""
|
"""Run comparison tests"""
|
||||||
print("Memory Usage Comparison: Faiss HNSW vs LEANN HNSW")
|
print("Storage + Search Memory Comparison: Faiss HNSW vs LEANN HNSW")
|
||||||
print("=" * 60)
|
print("=" * 60)
|
||||||
|
|
||||||
# Test Faiss HNSW
|
# Test Faiss HNSW
|
||||||
@@ -271,7 +188,7 @@ def main():
|
|||||||
|
|
||||||
# Final comparison
|
# Final comparison
|
||||||
print("\n" + "=" * 60)
|
print("\n" + "=" * 60)
|
||||||
print("FINAL COMPARISON")
|
print("STORAGE + SEARCH MEMORY COMPARISON")
|
||||||
print("=" * 60)
|
print("=" * 60)
|
||||||
|
|
||||||
# Get storage sizes
|
# Get storage sizes
|
||||||
@@ -281,33 +198,37 @@ def main():
|
|||||||
# Get Faiss storage size using Python
|
# Get Faiss storage size using Python
|
||||||
if os.path.exists("./storage_faiss"):
|
if os.path.exists("./storage_faiss"):
|
||||||
total_size = 0
|
total_size = 0
|
||||||
for dirpath, dirnames, filenames in os.walk("./storage_faiss"):
|
for dirpath, _, filenames in os.walk("./storage_faiss"):
|
||||||
for filename in filenames:
|
for filename in filenames:
|
||||||
filepath = os.path.join(dirpath, filename)
|
filepath = os.path.join(dirpath, filename)
|
||||||
total_size += os.path.getsize(filepath)
|
total_size += os.path.getsize(filepath)
|
||||||
faiss_storage_size = total_size / (1024 * 1024) # Convert to MB
|
faiss_storage_size = total_size / (1024 * 1024) # Convert to MB
|
||||||
|
|
||||||
# LEANN storage size is already captured in leann_results
|
print("Faiss HNSW:")
|
||||||
|
|
||||||
print(f"Faiss HNSW:")
|
|
||||||
if "error" in faiss_results:
|
if "error" in faiss_results:
|
||||||
print(f" ❌ Failed: {faiss_results['error']}")
|
print(f" ❌ Failed: {faiss_results['error']}")
|
||||||
else:
|
else:
|
||||||
print(f" Peak Memory: {faiss_results['peak_memory']:.1f} MB")
|
print(f" Search Memory: {faiss_results['peak_memory']:.1f} MB")
|
||||||
print(f" Storage Size: {faiss_storage_size:.1f} MB")
|
print(f" Storage Size: {faiss_storage_size:.1f} MB")
|
||||||
|
|
||||||
print(f"\nLEANN HNSW:")
|
print("\nLEANN HNSW:")
|
||||||
print(f" Peak Memory: {leann_results['peak_memory']:.1f} MB")
|
if "error" in leann_results:
|
||||||
print(f" Storage Size: {leann_storage_size:.1f} MB")
|
print(f" ❌ Failed: {leann_results['error']}")
|
||||||
|
else:
|
||||||
|
print(f" Search Memory: {leann_results['peak_memory']:.1f} MB")
|
||||||
|
print(f" Storage Size: {leann_storage_size:.1f} MB")
|
||||||
|
|
||||||
# Calculate improvements only if Faiss test succeeded
|
# Calculate improvements only if both tests succeeded
|
||||||
if "error" not in faiss_results:
|
if "error" not in faiss_results and "error" not in leann_results:
|
||||||
memory_ratio = faiss_results["peak_memory"] / leann_results["peak_memory"]
|
memory_ratio = faiss_results["peak_memory"] / leann_results["peak_memory"]
|
||||||
|
|
||||||
print(f"\nLEANN vs Faiss:")
|
print("\nLEANN vs Faiss Performance:")
|
||||||
print(f" Memory Usage: {memory_ratio:.1f}x less")
|
memory_saving = faiss_results["peak_memory"] - leann_results["peak_memory"]
|
||||||
|
print(
|
||||||
|
f" Search Memory: {memory_ratio:.1f}x less ({memory_saving:.1f} MB saved)"
|
||||||
|
)
|
||||||
|
|
||||||
# Storage comparison - be clear about which is larger
|
# Storage comparison
|
||||||
if leann_storage_size > faiss_storage_size:
|
if leann_storage_size > faiss_storage_size:
|
||||||
storage_ratio = leann_storage_size / faiss_storage_size
|
storage_ratio = leann_storage_size / faiss_storage_size
|
||||||
print(
|
print(
|
||||||
@@ -319,20 +240,16 @@ def main():
|
|||||||
f" Storage Size: {storage_ratio:.1f}x smaller (LEANN uses less storage)"
|
f" Storage Size: {storage_ratio:.1f}x smaller (LEANN uses less storage)"
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
print(f" Storage Size: similar")
|
print(" Storage Size: similar")
|
||||||
|
|
||||||
print(f"\nSavings:")
|
|
||||||
memory_saving = faiss_results["peak_memory"] - leann_results["peak_memory"]
|
|
||||||
storage_diff = faiss_storage_size - leann_storage_size
|
|
||||||
print(f" Memory: {memory_saving:.1f} MB")
|
|
||||||
if storage_diff >= 0:
|
|
||||||
print(f" Storage: {storage_diff:.1f} MB saved")
|
|
||||||
else:
|
|
||||||
print(f" Storage: {abs(storage_diff):.1f} MB additional used")
|
|
||||||
else:
|
else:
|
||||||
print(f"\n✅ LEANN HNSW ran successfully!")
|
if "error" not in leann_results:
|
||||||
print(f"📊 LEANN Memory Usage: {leann_results['peak_memory']:.1f} MB")
|
print("\n✅ LEANN HNSW completed successfully!")
|
||||||
print(f"📊 LEANN Storage Size: {leann_storage_size:.1f} MB")
|
print(f"📊 Search Memory: {leann_results['peak_memory']:.1f} MB")
|
||||||
|
print(f"📊 Storage Size: {leann_storage_size:.1f} MB")
|
||||||
|
if "error" not in faiss_results:
|
||||||
|
print("\n✅ Faiss HNSW completed successfully!")
|
||||||
|
print(f"📊 Search Memory: {faiss_results['peak_memory']:.1f} MB")
|
||||||
|
print(f"📊 Storage Size: {faiss_storage_size:.1f} MB")
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
|
|||||||
@@ -1,14 +1,17 @@
|
|||||||
#!/usr/bin/env python3
|
#!/usr/bin/env python3
|
||||||
"""Test only Faiss HNSW"""
|
"""Test only Faiss HNSW"""
|
||||||
|
|
||||||
import sys
|
import sys
|
||||||
import time
|
import time
|
||||||
import psutil
|
import psutil
|
||||||
import gc
|
import gc
|
||||||
|
|
||||||
|
|
||||||
def get_memory_usage():
|
def get_memory_usage():
|
||||||
process = psutil.Process()
|
process = psutil.Process()
|
||||||
return process.memory_info().rss / 1024 / 1024
|
return process.memory_info().rss / 1024 / 1024
|
||||||
|
|
||||||
|
|
||||||
class MemoryTracker:
|
class MemoryTracker:
|
||||||
def __init__(self, name: str):
|
def __init__(self, name: str):
|
||||||
self.name = name
|
self.name = name
|
||||||
@@ -27,9 +30,21 @@ class MemoryTracker:
|
|||||||
print(f"Peak Memory: {peak_mem:.1f} MB")
|
print(f"Peak Memory: {peak_mem:.1f} MB")
|
||||||
return peak_mem
|
return peak_mem
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
import faiss
|
try:
|
||||||
from llama_index.core import SimpleDirectoryReader, VectorStoreIndex, StorageContext, Settings
|
import faiss
|
||||||
|
except ImportError:
|
||||||
|
print("Faiss is not installed.")
|
||||||
|
print("Please install it with `uv pip install faiss-cpu`")
|
||||||
|
sys.exit(1)
|
||||||
|
|
||||||
|
from llama_index.core import (
|
||||||
|
SimpleDirectoryReader,
|
||||||
|
VectorStoreIndex,
|
||||||
|
StorageContext,
|
||||||
|
Settings,
|
||||||
|
)
|
||||||
from llama_index.vector_stores.faiss import FaissVectorStore
|
from llama_index.vector_stores.faiss import FaissVectorStore
|
||||||
from llama_index.embeddings.huggingface import HuggingFaceEmbedding
|
from llama_index.embeddings.huggingface import HuggingFaceEmbedding
|
||||||
|
|
||||||
@@ -45,7 +60,12 @@ def main():
|
|||||||
faiss_index.hnsw.efConstruction = 64
|
faiss_index.hnsw.efConstruction = 64
|
||||||
tracker.checkpoint("After Faiss index creation")
|
tracker.checkpoint("After Faiss index creation")
|
||||||
|
|
||||||
documents = SimpleDirectoryReader("examples/data", recursive=True, encoding="utf-8", required_exts=[".pdf", ".txt", ".md"]).load_data()
|
documents = SimpleDirectoryReader(
|
||||||
|
"examples/data",
|
||||||
|
recursive=True,
|
||||||
|
encoding="utf-8",
|
||||||
|
required_exts=[".pdf", ".txt", ".md"],
|
||||||
|
).load_data()
|
||||||
tracker.checkpoint("After document loading")
|
tracker.checkpoint("After document loading")
|
||||||
|
|
||||||
print("Building Faiss HNSW index...")
|
print("Building Faiss HNSW index...")
|
||||||
@@ -74,5 +94,6 @@ def main():
|
|||||||
peak_memory = tracker.summary()
|
peak_memory = tracker.summary()
|
||||||
print(f"Peak Memory: {peak_memory:.1f} MB")
|
print(f"Peak Memory: {peak_memory:.1f} MB")
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
main()
|
main()
|
||||||
Reference in New Issue
Block a user