Tagged "local-llm-deployment"
-
New Era of On-Device AI Driven by High-Speed UFS 5.0 Storage
-
Red Hat Launches AI Enterprise for Hybrid AI Deployments
-
Qwen3.5-27B Identified as Sweet Spot for Mid-Range Local Deployment
-
Mirai Tech Raises $10 Million for On-Device AI Innovation
-
No, Local LLMs Can't Replace ChatGPT or Gemini — I Tried
-
Elastic Introduces Best-in-Class Embedding Models for High Performance Semantic Search
-
Show HN: Dypai – Build Backends from Your IDE Using AI and MCP
-
Show HN: Agora – AI API Pricing Oracle with X402 Micropayments
-
How Do You Know Which SKILL.md Is Good?
-
Nvidia Could Launch Its First Laptops With Its Own Processors
-
nanollama: Open-Source Framework for Training Llama 3 from Scratch with One-Command GGUF Export
-
GPT-OSS 20B Demonstrates Practical Agentic Capabilities Running Fully Locally
-
FORTHought: Self-Hosted AI Stack for Physics Labs Built on OpenWebUI
-
AI-Powered Reverse-Engineering of Rosetta 2 for Linux
-
Security Alert: Fraudulent Shade Software Plagiarized from Heretic Project
-
Ollama 0.17 Released With Improved OpenClaw Onboarding
-
AI PCs Explained: 7 Critical Truths About NPUs and Privacy
-
Search and Analyze Documents from the DOJ Epstein Files Release with Local LLM
-
I Run Local LLMs in One of the World's Priciest Energy Markets, and I Can Barely Tell
-
At India AI Impact Summit, Intel Showcases Its AI PCs and Cost-Efficient Frugal AI
-
I Thought I Needed a GPU to Run AI Until I Learned About These Models
-
Google Is Exploring Ways to Use Its Financial Might to Take on Nvidia
-
Running Local LLMs and VLMs on Arduino UNO Q with yzma
-
LayerScale Launches Inference Engine Faster Than vLLM, SGLang, and TRT-LLM
-
GPT4All Replaces Ollama On Mac After Quick Trial
-
Aegis.rs: Open Source Rust-Based LLM Security Proxy Released
-
Tailscale Releases New Tool to Prevent Sensitive Data Leakage to Cloud AI Services
-
Self-Hosted AI: A Complete Roadmap for Beginners
-
Show HN: Inkog – Pre-flight check for AI agents (governance, loops, injection)
-
Chinese AI Chipmaker Axera Semiconductor Plans $379 Million Hong Kong IPO for Edge Inference Hardware
-
Ask HN: What is the best bang for buck budget AI coding?
-
InitRunner: YAML-Based AI Agent Framework with RAG and Memory
-
175,000 Publicly Exposed Ollama AI Servers Discovered Across 130 Countries
-
Scaling llama.cpp On Neoverse N2: Solving Cross-NUMA Performance Issues
-
Running Your Own AI Assistant for €19/Month: Complete Self-Hosting Guide
-
Scaling llama.cpp On Neoverse N2: Solving Cross-NUMA Performance Issues
-
I Tried a Claude Code Rival That's Local, Open Source, and Completely Free
-
5 Practical Ways to Use Local LLMs with MCP Tools
-
Carmack Proposes Using Long Fiber Lines as L2 Cache for Streaming AI Data
-
Anthropic Releases Claude Opus 4.6 Sabotage Risk Assessment