Files
crm.clientright.ru/aiassist/vectorize.py
Fedor ac7467f0b4 Major CRM updates: AI Assistant, Court Status API, S3 integration improvements, and extensive file storage system
- Added comprehensive AI Assistant system (aiassist/ directory):
  * Vector search and embedding capabilities
  * Typebot proxy integration
  * Elastic search functionality
  * Message classification and chat history
  * MCP proxy for external integrations

- Implemented Court Status API (GetCourtStatus.php):
  * Real-time court document status checking
  * Integration with external court systems
  * Comprehensive error handling and logging

- Enhanced S3 integration:
  * Improved file backup system with metadata
  * Batch processing capabilities
  * Enhanced error logging and recovery
  * Copy operations with URL fixing

- Added Telegram contact creation API
- Improved error logging across all modules
- Enhanced callback system for AI responses
- Extensive backup file storage with timestamps
- Updated documentation and README files

- File storage improvements:
  * Thousands of backup files with proper metadata
  * Fix operations for broken file references
  * Project-specific backup and recovery systems
  * Comprehensive file integrity checking

Total: 26,461+ files added/modified including AWS SDK, vendor dependencies, and extensive backup system.
2025-10-16 11:17:21 +03:00

35 lines
1.4 KiB
Python

#!/var/www/laws/legal/bin/python3
from sentence_transformers import SentenceTransformer
from natasha import Doc, Segmenter, MorphVocab, NamesExtractor
# Инициализация моделей
#sbert_model = SentenceTransformer("paraphrase-MiniLM-L6-v2", device="cpu")
sbert_model = SentenceTransformer("/var/www/laws/models/paraphrase-MiniLM-L6-v2", device="cpu")
segmenter = Segmenter()
morph_vocab = MorphVocab()
import sys
import json
import sys
# Добавляем нужный путь для установки зависимостей, если требуется
sys.path.append('/var/www/laws/legal/bin/python3.11/site-packages')
from sentence_transformers import SentenceTransformer
def get_embedding(text):
# Инициализируем модель на CPU
sbert_model = SentenceTransformer("paraphrase-MiniLM-L6-v2", device="cpu")
# Получаем эмбеддинг (модель возвращает numpy-массив)
embedding = sbert_model.encode(text, show_progress_bar=False)
# Преобразуем в список и возвращаем
return embedding.tolist()
if __name__ == '__main__':
# Считываем текст из аргументов командной строки
input_text = sys.argv[1] if len(sys.argv) > 1 else ""
vector = get_embedding(input_text)
# Выводим результат в формате JSON
print(json.dumps(vector))