Documentation
¶
Overview ¶
Package llm provides the interface to use Large Language Models in Spin.
Index ¶
Constants ¶
This section is empty.
Variables ¶
This section is empty.
Functions ¶
This section is empty.
Types ¶
type EmbeddingsResult ¶
type EmbeddingsResult struct {
// Embeddings are the embeddings generated by the request.
Embeddings [][]float32
// Usage is usage related to an embeddings generation request.
Usage *EmbeddingsUsage
}
EmbeddingsResult of generating embeddings.
func GenerateEmbeddings ¶
func GenerateEmbeddings(model string, text []string) (*EmbeddingsResult, error)
GenerateEmbeddings generates the embeddings for the supplied list of text.
type EmbeddingsUsage ¶
type EmbeddingsUsage struct {
// PromptTokenCount is number of tokens in the prompt.
PromptTokenCount int
}
Embeddings is usage related to an embeddings generation request.
type InferencingParams ¶
type InferencingParams struct {
// MaxTokens is the maximum tokens that should be inferred.
// Default: 100
//
// Note: the backing implementation may return less tokens.
MaxTokens int32
// RepeatPenalty is the amount the model should avoid repeating tokens.
// Default: 1.1
RepeatPenalty float32
// RepeatPenaltyLastNTokenCount the number of tokens the model should
// apply the repeat penalty to.
// Default: 64
RepeatPenaltyLastNTokenCount int32
// Temperature is the randomness with which the next token is selected.
// Default: 0.8
Temperature float32
// TopK is the number of possible next tokens the model will choose from.
// Default: 40
TopK int32
// TopP is the probability total of next tokens the model will choose
// from.
// Default: 0.9
TopP float32
}
InferenceParams is the optional request parameters.
type InferencingResult ¶
type InferencingResult struct {
// Text is the text generated by the model.
Text string
// Usage is information about the inferencing request.
Usage *InferencingUsage
}
InferencingResult is the result of an inference.
func Infer ¶
func Infer(model, prompt string, params *InferencingParams) (*InferencingResult, error)
Infer performs inferencing using the provided model and prompt with the given optional parameters.
type InferencingUsage ¶
type InferencingUsage struct {
// PromptTokenCount is the number of tokens in the prompt.
PromptTokenCount int
// GeneratedTokenCount is the number of tokens generated by the
// inferencing operation.
GeneratedTokenCount int
}
InferencingUsage represents information related to the inferencing result.
Click to show internal directories.
Click to hide internal directories.