misc / SimpleMem /MCP /reference /utils /embedding.py
NingsenWang's picture
Upload SimpleMem project snapshot
a54fd97 verified
"""
Embedding utilities - Generate vector embeddings using SentenceTransformers
Supports Qwen3 Embedding models through SentenceTransformers interface
"""
from typing import List, Optional, Dict, Any
import numpy as np
import config
import os
class EmbeddingModel:
"""
Embedding model using SentenceTransformers (supports Qwen3 and other models)
"""
def __init__(self, model_name: str = None, use_optimization: bool = True):
self.model_name = model_name or config.EMBEDDING_MODEL
self.use_optimization = use_optimization
print(f"Loading embedding model: {self.model_name}")
# Check if it's a Qwen3 model (through SentenceTransformers)
if self.model_name.startswith("qwen3"):
self._init_qwen3_sentence_transformer()
else:
self._init_standard_sentence_transformer()
def _init_qwen3_sentence_transformer(self):
"""Initialize Qwen3 model using SentenceTransformers"""
try:
from sentence_transformers import SentenceTransformer
# Map model names to actual model paths
qwen3_models = {
"qwen3-0.6b": "Qwen/Qwen3-Embedding-0.6B",
"qwen3-4b": "Qwen/Qwen3-Embedding-4B",
"qwen3-8b": "Qwen/Qwen3-Embedding-8B"
}
model_path = qwen3_models.get(self.model_name, self.model_name)
print(f"Loading Qwen3 model via SentenceTransformers: {model_path}")
# Initialize with optimization settings
if self.use_optimization:
try:
# Try to use flash_attention_2 and left padding for better performance
self.model = SentenceTransformer(
model_path,
model_kwargs={
"attn_implementation": "flash_attention_2",
"device_map": "auto"
},
tokenizer_kwargs={"padding_side": "left"},
trust_remote_code=True
)
print("Qwen3 loaded with flash_attention_2 optimization")
except Exception as e:
print(f"Flash attention failed ({e}), using standard loading...")
self.model = SentenceTransformer(model_path, trust_remote_code=True)
else:
self.model = SentenceTransformer(model_path, trust_remote_code=True)
self.dimension = self.model.get_sentence_embedding_dimension()
self.model_type = "qwen3_sentence_transformer"
# Check if Qwen3 supports query prompts
self.supports_query_prompt = hasattr(self.model, 'prompts') and 'query' in getattr(self.model, 'prompts', {})
print(f"Qwen3 model loaded successfully with dimension: {self.dimension}")
if self.supports_query_prompt:
print("Query prompt support detected")
except Exception as e:
print(f"Failed to load Qwen3 model: {e}")
print("Falling back to default SentenceTransformers model...")
self._fallback_to_sentence_transformer()
def _init_standard_sentence_transformer(self):
"""Initialize standard SentenceTransformer model"""
try:
from sentence_transformers import SentenceTransformer
self.model = SentenceTransformer(self.model_name)
self.dimension = self.model.get_sentence_embedding_dimension()
self.model_type = "sentence_transformer"
self.supports_query_prompt = False
print(f"SentenceTransformer model loaded with dimension: {self.dimension}")
except Exception as e:
print(f"Failed to load SentenceTransformer model: {e}")
raise
def _fallback_to_sentence_transformer(self):
"""Fallback to default SentenceTransformer model"""
fallback_model = "sentence-transformers/all-MiniLM-L6-v2"
print(f"Using fallback model: {fallback_model}")
self.model_name = fallback_model
self._init_standard_sentence_transformer()
def encode(self, texts: List[str], is_query: bool = False) -> np.ndarray:
"""
Encode list of texts to vectors
Args:
- texts: List of texts to encode
- is_query: Whether these are query texts (for Qwen3 prompt optimization)
"""
if isinstance(texts, str):
texts = [texts]
# Use query prompt for Qwen3 models when encoding queries
if self.model_type == "qwen3_sentence_transformer" and self.supports_query_prompt and is_query:
return self._encode_with_query_prompt(texts)
else:
return self._encode_standard(texts)
def encode_single(self, text: str, is_query: bool = False) -> np.ndarray:
"""
Encode single text
Args:
- text: Text to encode
- is_query: Whether this is a query text (for Qwen3 prompt optimization)
"""
return self.encode([text], is_query=is_query)[0]
def encode_query(self, queries: List[str]) -> np.ndarray:
"""
Encode queries with optimal settings for Qwen3
"""
return self.encode(queries, is_query=True)
def encode_documents(self, documents: List[str]) -> np.ndarray:
"""
Encode documents (no query prompt)
"""
return self.encode(documents, is_query=False)
def _encode_with_query_prompt(self, texts: List[str]) -> np.ndarray:
"""Encode texts using Qwen3 query prompt"""
try:
embeddings = self.model.encode(
texts,
prompt_name="query", # Use Qwen3's query prompt
show_progress_bar=False,
normalize_embeddings=True
)
return embeddings
except Exception as e:
print(f"Query prompt encoding failed: {e}, falling back to standard encoding")
return self._encode_standard(texts)
def _encode_standard(self, texts: List[str]) -> np.ndarray:
"""Encode texts using standard method"""
embeddings = self.model.encode(
texts,
show_progress_bar=False,
normalize_embeddings=True
)
return embeddings