Show HN: Replacing cloud LLM APIs with local, domain-specific models
Most current LLM workflows depend on cloud APIs, which means sending data outside your system. We’ve been working on an alternative: a fully local stack that lets you run and adapt models without relying on external providers. The idea is not just to run models locally, but to make them useful for specific domains (legal, medical, internal knowledge) while keeping them small enough to run on commodity hardware.<p>Current state: 1) local inference engine (GGUF, API-compatible with existing tools) 2) prototype model hub with REST endpoints and model metadata 3) pipeline in progress to adapt general models into domain-specific ones<p>The open question we’re trying to answer is whether this process can be made reproducible, not just one-off fine-tuning. If it works, it could reduce the need for cloud-based AI in many real-world use cases. Repo: <a href="https://github.com/eullm/eullm" rel="nofollow">https://github.com/eullm/eullm</a> Curious to hear
More in Pivot Health
More from Pivot News
OpenAI Shuts Down Sora, Refocuses on Code AGI and Model Spud
OpenAI **discontinues Sora and all video generation products** to redeploy compute resources toward competing with Anthropic in enterprise coding and knowledge work. CEO Sam Altman **narrows his role to focus on capital, supply chains, and data centers**; Fiji Simo's product division becomes 'AGI Deployment' team. New model 'Spud' completes pre-training with expectations to 'accelerate the economy'; **Disney cancels $1B investment partnership** following Sora shutdown.