Tagged "on-device-inference"
-
Velr: Embedded Property-Graph Database for Local LLM Applications
-
Self-Hostable AI Agents and Internal Software Framework Released
-
Qt 6.11 Released with Enhanced Cross-Platform Deployment Capabilities
-
Alibaba Commits to Continuous Open-Sourcing of Qwen and Wan Models
-
Building a Production AI Receptionist: Practical Local LLM Deployment Case Study
-
Careless Whisper – Personal Local Speech to Text
-
A Little Gap That Will Ensure the Future of AI Agents Being Autonomous
-
Running an AI Agent on a 448KB RAM Microcontroller
-
Qualcomm and Samsung's 30-Year AI Alliance Enters a New Phase as On-Device AI Chip Race Heats Up
-
NVIDIA Nemotron 3 Nano 4B Enables On-Device Inference Directly in Web Browsers via WebGPU
-
Meet Sarvam Edge: India's AI Model That Runs on Phones and Laptops With No Internet
-
Multiverse Computing Targets On-Device AI With Compressed Models and New API Portal
-
Dell Pro Max 16 Plus Launches With Enterprise-Grade Discrete NPU for On-Device AI
-
On-Device AI: Tether's QVAC Fabric Enables Local Training
-
Snapdragon 8 Elite Gen 5 Hands the Galaxy S26 the AI Upgrade We've Been Waiting For
-
LucidShark – Local-first, open-source quality and security gate
-
OpenJarvis: Local-First AI Agents That Run Entirely On-Device
-
A New Magnetic Material for the AI Era
-
How I Used Lima for an AI Coding Agent Sandbox
-
The Moment AI Agents Stopped Being a Feature and Started Becoming a System
-
Qwen 3.5 122B Demonstrates Exceptional Reasoning for Local Deployment
-
OmniCoder-9B: Efficient Coding Model for 8GB GPUs
-
Apple's On-Device AI Raises Privacy Alarms Across British Parliament
-
AMD Declares 'AI on the PC Has Crossed an Important Line' – Agent Computers as Next Breakthrough
-
India's Mobile-First AI Strategy Could Accelerate Local Inference Adoption in Emerging Markets
-
Hybrid AI Desktop Layer Combining DOM-Automation and API-Integrations
-
Cicikus v3 Prometheus 4.4B – An Experimental Franken-Merge for Edge Reasoning
-
Local Manga Translator: Production LLM Pipeline with YOLO, OCR, and Inpainting
-
I Fed My Home Assistant Logs Into a Local LLM, and It Found Problems I'd Been Ignoring for Months
-
Best Local LLM Models 2026: Developer Comparison
-
MeepaChat – Slack for AI Agents (iOS, macOS, Web / Cloud, Self-Hosted)
-
Local AI Coding Assistant: Complete VS Code + Ollama + Continue Setup
-
Simple Layer Duplication Technique Achieves Top Open LLM Leaderboard Performance
-
Kali Linux Integrates Local Ollama and MCP for AI-Driven Penetration Testing
-
SK Hynix Develops 1c LPDDR6 DRAM to Boost On-Device AI Performance in Mobile Devices
-
Qwen 3.5 Ultra-Compact Models Enable On-Device AI from Watches to Gaming
-
Google Delivers On-Device AI Features in New Chromebook Plus Model
-
Qwen 3.5 Small Expands On-Device AI to Phones and IoT with Offline Support
-
Engram – Open-Source Persistent Memory for AI Agents
-
Samsung Opens Registration for Vision AI QLED and OLED Television Integration
-
Show HN: Ivy – the first proactive, offline AI tutor
-
Windows 11 Notepad Gets On-Device AI Text Generation Without Subscription
-
Building PyTorch-Native Support for IBM Spyre Accelerator
-
Llama.cpp Merges Automatic Parser Generator to Mainline
-
Turning Your Linux Terminal into a Local AI Assistant
-
IBM Granite 4.0 1B Speech Model Released for Multilingual Speech Recognition
-
Show HN: Asterode – Multi-Model AI App with Memory and Power Features
-
Alibaba Releases Qwen 3.5 AI Model with On-Device AI Support
-
Windows 11 Notepad to Feature On-Device AI Text Generation Without Subscription
-
The Emerging Role of SRAM-Centric Chips in AI Inference
-
OPPO and MediaTek Highlight On-Device AI Innovations at MWC 2026
-
Alibaba Releases Qwen 3.5 AI Model with On-Device AI Support
-
MediaTek Advances Omni Model for Efficient Smartphone Inference
-
Kakao Launches Kanana AI for On-Device Schedule and Recommendation Management
-
Apple Unveils MacBook Pro with M5 Pro and M5 Max Featuring On-Device AI
-
RunAnywhere Launches Production-Grade On-Device AI Platform for Enterprise Scale
-
Qualcomm Snapdragon Wear Elite Brings On-Device AI to Smartwatches
-
OpenWrt 25.12.0 – Stable Release
-
Glyph – A Local-First Markdown Notes App for macOS Built With Rust
-
Apple Unveils MacBook Pro With M5 Pro and M5 Max for On-Device AI
-
Apple M5 Pro and M5 Max: 4× Faster LLM Processing
-
AMD Launches Copilot+ Desktop Chips to Compete in On-Device AI Market
-
VibeWhisper – macOS Voice-to-Text with 100% Local Processing Option
-
Qwen 3.5 Small Models Released: 0.8B to 9B Parameters Optimized for On-Device Inference
-
Apple M4 iPad Air Targets AI Users with Double M1 Speed Performance
-
Alibaba's Qwen 3.5 Small Model Runs Directly on iPhone 17
-
Qualcomm Launches Snapdragon Wear Elite for On-Device AI on Wearables
-
HP ZBook Ultra 14 G1a Workstation Reclaims Local AI Workflows for Professionals
-
Browser Use vs. Claude Computer Use: Comparing Agent Automation Frameworks
-
AMD Expands Ryzen AI 400 Series Portfolio for Consumer and Enterprise AI PC Options
-
ParseHive – AI-Powered Invoice Data Extraction for Windows and Mac
-
DeepSeek V4 Multimodal Model Coming Next Week With Image and Video Generation
-
Apple Intelligence, Galaxy AI, Gemini: Why Your AI-Powered Phone Is Worth Repairing
-
Serve Markdown to LLMs from your Next.js app
-
On-Device AI in Mobile Apps: What Should Run on the Phone vs the Cloud (A 2026 Decision Guide)
-
Meta Reveals AI-Packed Smartwatch In 2026 – Why Wearables Shift Now
-
Galaxy S26 Debuts AI-Powered Scam Detection in Bold Security Push
-
Snapdragon 8 Elite Gen 5 for Galaxy Official: 5 Key Improvements that Push the Boundaries
-
Snapdragon 8 Elite Gen 5 Powers Galaxy S26 Series With Enhanced On-Device AI
-
On-Device AI in Mobile Apps: What Should Run on the Phone vs the Cloud (A 2026 Decision Guide)
-
Show HN: Caret – Tab to Complete at Any App on Your Mac
-
Arduino, Qualcomm Bring On-Device AI and Robotics Learning to Indian School Systems
-
Android Phones Are Getting Smarter Without Internet — Here's Why On-Device AI Is the Next Big Shift
-
Android Phones Are Getting Smarter Without Internet — On-Device AI as the Next Shift
-
Building a Privacy-Preserving RAG System in the Browser
-
Researchers Develop Persistent Memory System for Local LLMs—No RAG Required
-
Ollama for JavaScript Developers: Building AI Apps Without API Keys
-
DeepSeek Paper – DualPath: Breaking the Bandwidth Bottleneck in LLM Inference
-
The Complete Developer's Guide to Running LLMs Locally: From Ollama to Production
-
Apple: Python bindings for access to the on-device Apple Intelligence model
-
Agent System – 7 specialized AI agents that plan, build, verify, and ship code
-
New Era of On-Device AI Driven by High-Speed UFS 5.0 Storage
-
Mirai Announces $10M to Advance On-Device AI Performance for Consumer Devices
-
Show HN: MCP-Enabled File Storage for AI Agents, Auth via Ethereum Wallet
-
Mirai Tech Raises $10 Million for On-Device AI Innovation
-
No, Local LLMs Can't Replace ChatGPT or Gemini — I Tried
-
Kioxia Sampling UFS 5.0 Embedded Flash Memory for Next-Generation Mobile Applications
-
Enhanced Interface Speed Enables High-Performance On-Device AI Features in Smartphones
-
Elastic Introduces Best-in-Class Embedding Models for High Performance Semantic Search
-
Apple Accelerates U.S. Manufacturing with Mac Mini Production
-
Comparing Manual vs. AI Requirements Gathering: 2 Sentences vs. 127-Point Spec
-
South Korea to Launch $687 Million Project to Develop On-Device AI Semiconductors
-
Qwen3's Voice Embeddings Enable Local Voice Cloning and Mathematical Voice Manipulation
-
Custom Portable Workstation Optimized for Local AI Inference Builds
-
Open-Source llama.cpp Finds Long-Term Home at Hugging Face
-
Future of Mobile AI: What On-Device Intelligence Means for App Developers
-
Future of Mobile AI: What On-Device Intelligence Means for App Developers
-
The Complete Stack for Local Autonomous Agents: From GGML to Orchestration
-
AI Is Stress Testing Processor Architectures and RISC-V Fits the Moment
-
How Slow Local LLMs Are on My Framework 13 AMD Strix Point
-
At India AI Impact Summit, Intel Showcases AI PCs and Cost-Efficient Frugal AI
-
Asus ExpertBook B3 G2 with 50 TOPS AI Sets New Enterprise Standard
-
AI PCs Explained: 7 Critical Truths About NPUs and Privacy
-
Vellium v0.3.5: Major Writing Mode Overhaul and Native KoboldCpp Support
-
Taalas Etches AI Models onto Transistors to Rocket Boost Inference
-
I Run Local LLMs in One of the World's Priciest Energy Markets, and I Can Barely Tell
-
[Release] Ouro-2.6B-Thinking: ByteDance's Recurrent Model Now Runnable Locally
-
At India AI Impact Summit, Intel Showcases Its AI PCs and Cost-Efficient Frugal AI
-
Open-Source + AI: ggml Joins Hugging Face, llama.cpp Stays Open—Local AI's Long-Term Home
-
GGML.AI Acquired by Hugging Face
-
Apple Researchers Develop On-Device AI Agent That Interacts With Apps for You
-
SanityBoard Adds 27 New Model Evaluations Including Qwen 3.5 Plus, GLM 5, and Gemini 3.1 Pro
-
PaddleOCR-VL Now Integrated into llama.cpp for Multilingual OCR
-
Why AI Models Fail at Iterative Reasoning and What Could Fix It
-
Show HN: Forked – A Local Time-Travel Debugger for OpenClaw Agents
-
Mihup and Qualcomm Collaborate to Advance Secure On-Device Voice AI for BFSI
-
Local-First RAG: Vector Search in SQLite with Hamming Distance
-
Kitten TTS V0.8 Released: State-of-the-Art Super-Tiny Text-to-Speech Model Under 25MB
-
Clipthesis: Free Local App for Video Tagging and Search Across Drives
-
Why My Country's AI Scene Is Built on Sand
-
Sarvam AI Launches Edge Model to Challenge Major AI Players with Local-First Approach
-
Qualcomm Ventures Positions India as Blueprint for Affordable On-Device AI Infrastructure
-
Can We Leverage AI/LLMs for Self-Learning?
-
Meet Sarvam Edge: India's AI Model That Runs on Phones and Laptops With No Internet
-
Cohere Releases Tiny Aya: Efficient 3.3B Multilingual Model for 70+ Languages
-
ASUS Zenbook 14 Launches in India with AI-Capable Hardware, Starting at Rs 1,15,990
-
Asus ExpertBook B3 G2 Laptop Features Ryzen AI 9 HX 470 CPU in 1.41kg Ultraportable Form Factor
-
Ask HN: What is the best bang for buck budget AI coding?
-
I broke into my own AI system in 10 minutes. I built it
-
Sourdine: Open-Source macOS App for 100% Local AI Transcription
-
MiniMax-M2.5 230B MoE Model Released with GGUF Support for Local Deployment
-
Nanbeige4.1-3B: A Small General Model that Reasons, Aligns, and Acts
-
Arm SME2 Technology Expands CPU Capabilities for On-Device AI