Local model serving - Using Foundry Local
July 17, 2025 • inferencing, local serving
There are several options available for running Large Language Model (LLM) inference locally. Foundry Local by Microsoft is a new entrant.
4 articles in category "Inferencing"
July 17, 2025 • inferencing, local serving
There are several options available for running Large Language Model (LLM) inference locally. Foundry Local by Microsoft is a new entrant.
July 16, 2025 • inferencing, local serving
Docker Model Runner — a faster, simpler way to run and test AI models locally, right from your existing workflow. Whether you’re experimenting with the latest LLMs or deploying to production, Model Runner brings the performance and control you need, without the friction.
July 15, 2025 • inferencing, local serving
There are several options available for running Large Language Model (LLM) inference locally. LM Studio is one such option. It is more comprehensive and offers some great features.
July 14, 2025 • inferencing, local serving
There are several options available for running Large Language Model (LLM) inference locally. Ollama is one such option and my favorite among all. Ollama offers access to a wide range of models and has recently enabled cloud-hosted models as well. It offers both CLI and GUI (chat interface) to interact with the loaded models.