Skip to contents

Send a prompt to a locally running LM Studio model. This wrapper automatically routes your request to the appropriate subfunction based on the selected API type.

Usage

lms_chat(
  model,
  input,
  system_prompt = NULL,
  host = "http://localhost:1234",
  api_type = c("openresponses", "openai", "native"),
  logprobs = FALSE,
  simplify = TRUE,
  ...
)

Arguments

model

Character. The name of the loaded model.

input

Character. The user prompt to send to the model.

system_prompt

Character. An optional system prompt to guide model behavior.

host

Character. The base URL of the LM Studio server. Default is "http://localhost:1234".

api_type

Character. The LM Studio API endpoint to use. Options are "openresponses" (default), "openai", or "native".

logprobs

Logical. Whether to return the log probabilities of the generated tokens. Default is FALSE.

simplify

Logical. If TRUE, extracts the core text response. Default is TRUE.

...

Additional arguments passed to the selected API body.