Tagged "rag"
- Velr: Embedded Property-Graph Database for Local LLM Applications
- LM Studio Releases Reworked Plugins with Fully Local Web Research
- Powerful AI Search Engine Built on Single GeForce RTX 5090
- Llama 8B Matches 70B Performance on Multi-Hop QA Using Structured Prompting
- LMCache Dramatically Accelerates LLM Inference on Oracle Data Science Platform
- MiniMax-M2.7: New Compact Model Announced for Local Deployment
- Mamba 3: State Space Model Architecture Optimized for Inference
- Gloss: Open-Source, Local-First RAG Alternative to NotebookLM Built in Rust
- AI Agent Reliability Tracker
- Framework Choice Critical: llama.cpp and vLLM Outperform Ollama for Qwen 3.5 Testing
- RAG vs. Skill vs. MCP vs. RLM: Comparing LLM Enhancement Patterns
- RAG-Enterprise – 100% Local RAG System for Enterprise Documents
- Building a Privacy-Preserving RAG System in the Browser
- Researchers Develop Persistent Memory System for Local LLMs—No RAG Required
- Elastic Introduces Best-in-Class Embedding Models for High Performance Semantic Search
- Elastic Introduces Best-in-Class Embedding Models for High Performance Semantic Search
- Search and Analyze Documents from the DOJ Epstein Files Release with Local LLM
- NVIDIA Releases Dynamo v0.9.0: Infrastructure Overhaul With FlashIndexer and Multi-Modal Support
- Local-First RAG: Vector Search in SQLite with Hamming Distance
- InitRunner: YAML-Based AI Agent Framework with RAG and Memory
- GPU-Accelerated DataFrame Library for Local Inference Workloads
- Microsoft MarkItDown: Document Preprocessing Tool for LLMs
- Building a RAG Pipeline on 2M+ Pages: EpsteinFiles-RAG Project