The problem: Most AI requires cloud APIs, which means your data leaves your control. For banks, healthcare, defense, and regulated industries – that's a non-starter.
The solution: EdgeAI-OS runs everything locally. No cloud calls. No API keys. No telemetry. Boot the ISO, use AI. Your data never leaves the machine.
Security features: - 100% offline operation – air-gap friendly, zero network dependencies - No external API calls – all inference runs locally on CPU - Command risk assessment – every command classified as Safe/Moderate/Dangerous - Dangerous pattern blocking – prevents rm -rf /, curl|bash, fork bombs, etc. - Open source & auditable – MIT licensed, inspect every line of code - No data exfiltration – nothing phones home, ever
What's in the ISO: - Local LLMs (TinyLlama 1.1B + SmolLM 135M) – runs on CPU, no GPU needed - ai-sh: natural language shell where 80% of queries resolve instantly via templates - Multi-tier routing: simple queries → fast model, complex → larger model
Example ai-sh session: what time is it? [template] date ← instant, no LLM
files larger than 1gb [template] find . -size +1G ← instant, no LLM
rm -rf / [DANGEROUS] Blocked ← security check
configure nginx as reverse proxy [ai-generated] ... ← uses local LLM (~1-2s)
Target use cases: - Air-gapped enterprise environments (banks, healthcare, government) - Defense & classified networks - Edge devices with no internet connectivity - Privacy-conscious developers - Compliance-heavy industries (HIPAA, GDPR, SOC2)
Built with Rust, based on Debian. 4GB RAM recommended.
GitHub: https://github.com/neuralweaves/edgeai-os ISO Download: https://github.com/neuralweaves/edgeai-os/releases (1.2GB)
Would love feedback, especially from anyone working in secure/regulated environments. What features would make this enterprise-ready for you?