Documentation
¶
Overview ¶
Package llamaserver wraps the OpenAI-compatible HTTP surface exposed by llama.cpp's built-in server.
Index ¶
Constants ¶
View Source
const DefaultBaseURL = "http://localhost:8080/v1"
Variables ¶
View Source
var ProtocolCapabilities = openai.OpenAIProtocolCapabilities
ProtocolCapabilities mirrors the standard OpenAI-compatible surface.
Functions ¶
Types ¶
type UtilizationProbe ¶
type UtilizationProbe struct {
// contains filtered or unexported fields
}
UtilizationProbe queries llama-server observability endpoints and normalizes them into the shared endpoint utilization shape.
func NewUtilizationProbe ¶
func NewUtilizationProbe(baseURL string, client *http.Client) *UtilizationProbe
NewUtilizationProbe creates a probe for an OpenAI-compatible llama-server base URL.
func (*UtilizationProbe) Probe ¶
func (p *UtilizationProbe) Probe(ctx context.Context) utilization.EndpointUtilization
Probe first tries /metrics on the server root and falls back to /slots when metrics are unavailable.
Click to show internal directories.
Click to hide internal directories.