News
variablechange/nemotron-3-nano
12+ hour, 7+ min ago (127+ words) Special features defined by the model author Controls whether the model will think before replying Controls whether thinking history will be truncated to save context space Custom configuration options included with this model The underlying model files this model uses…...
List your models | LM Studio Docs
2+ mon, 1+ day ago (247+ words) Using MCP via API Idle TTL and Auto-Evict Chat with a model Tools and Function Calling LM Studio REST API Get a list of available models on your system, including both LLMs and embedding models. This endpoint has no request…...
Load a model | LM Studio Docs
2+ mon, 1+ day ago (401+ words) Using MCP via API Idle TTL and Auto-Evict Chat with a model Tools and Function Calling LM Studio REST API Load an LLM or Embedding model into memory with custom configuration for inference Unique identifier for the model to load....
Introducing LM Studio 0.4.0
2+ mon, 1+ day ago (641+ words) Introducing LM Studio 0.4.0 Today we are thrilled to share LM Studio 0.4.0, the next generation of LM Studio. This release introduces parallel requests with continuous batching for high throughput serving, all-new non-GUI deployment option, new stateful REST API, and a refreshed…...
Open Responses with local models via LM Studio
2+ mon, 2+ week ago (399+ words) Open Responses with local models via LM Studio We have partnered with OpenAI to support Open Responses - an open source specification based on the OpenAI Responses API. The Open Responses API in LM Studio brings up several new useful features:…...