{"service":"RunAI Proxy","version":"1.0.0","description":"Proxy API per LM Studio locale","endpoints":{"health":"GET \/health","chatCompletions":"POST \/v1\/chat\/completions","completions":"POST \/v1\/completions","models":"GET \/v1\/models","rerank":"POST \/v1\/rerank"}}