Documentation
¶
Overview ¶
Create a Llama inference endpoint.
Create an inference endpoint to perform an inference task with the `llama` service.
Index ¶
- Variables
- type NewPutLlama
- type PutLlama
- func (r *PutLlama) ChunkingSettings(chunkingsettings types.InferenceChunkingSettingsVariant) *PutLlama
- func (r PutLlama) Do(providedCtx context.Context) (*Response, error)
- func (r *PutLlama) ErrorTrace(errortrace bool) *PutLlama
- func (r *PutLlama) FilterPath(filterpaths ...string) *PutLlama
- func (r *PutLlama) Header(key, value string) *PutLlama
- func (r *PutLlama) HttpRequest(ctx context.Context) (*http.Request, error)
- func (r *PutLlama) Human(human bool) *PutLlama
- func (r PutLlama) Perform(providedCtx context.Context) (*http.Response, error)
- func (r *PutLlama) Pretty(pretty bool) *PutLlama
- func (r *PutLlama) Raw(raw io.Reader) *PutLlama
- func (r *PutLlama) Request(req *Request) *PutLlama
- func (r *PutLlama) Service(service llamaservicetype.LlamaServiceType) *PutLlama
- func (r *PutLlama) ServiceSettings(servicesettings types.LlamaServiceSettingsVariant) *PutLlama
- func (r *PutLlama) Timeout(duration string) *PutLlama
- type Request
- type Response
Constants ¶
This section is empty.
Variables ¶
var ErrBuildPath = errors.New("cannot build path, check for missing path parameters")
ErrBuildPath is returned in case of missing parameters within the build of the request.
Functions ¶
This section is empty.
Types ¶
type NewPutLlama ¶
NewPutLlama type alias for index.
func NewPutLlamaFunc ¶
func NewPutLlamaFunc(tp elastictransport.Interface) NewPutLlama
NewPutLlamaFunc returns a new instance of PutLlama with the provided transport. Used in the index of the library this allows to retrieve every apis in once place.
type PutLlama ¶
type PutLlama struct {
// contains filtered or unexported fields
}
func New ¶
func New(tp elastictransport.Interface) *PutLlama
Create a Llama inference endpoint.
Create an inference endpoint to perform an inference task with the `llama` service.
https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-inference-put-llama
func (*PutLlama) ChunkingSettings ¶
func (r *PutLlama) ChunkingSettings(chunkingsettings types.InferenceChunkingSettingsVariant) *PutLlama
The chunking configuration object. API name: chunking_settings
func (PutLlama) Do ¶
Do runs the request through the transport, handle the response and returns a putllama.Response
func (*PutLlama) ErrorTrace ¶
ErrorTrace When set to `true` Elasticsearch will include the full stack trace of errors when they occur. API name: error_trace
func (*PutLlama) FilterPath ¶
FilterPath Comma-separated list of filters in dot notation which reduce the response returned by Elasticsearch. API name: filter_path
func (*PutLlama) HttpRequest ¶
HttpRequest returns the http.Request object built from the given parameters.
func (*PutLlama) Human ¶
Human When set to `true` will return statistics in a format suitable for humans. For example `"exists_time": "1h"` for humans and `"exists_time_in_millis": 3600000` for computers. When disabled the human readable values will be omitted. This makes sense for responses being consumed only by machines. API name: human
func (PutLlama) Perform ¶
Perform runs the http.Request through the provided transport and returns an http.Response.
func (*PutLlama) Pretty ¶
Pretty If set to `true` the returned JSON will be "pretty-formatted". Only use this option for debugging only. API name: pretty
func (*PutLlama) Raw ¶
Raw takes a json payload as input which is then passed to the http.Request If specified Raw takes precedence on Request method.
func (*PutLlama) Service ¶
func (r *PutLlama) Service(service llamaservicetype.LlamaServiceType) *PutLlama
The type of service supported for the specified task type. In this case, `llama`. API name: service
func (*PutLlama) ServiceSettings ¶
func (r *PutLlama) ServiceSettings(servicesettings types.LlamaServiceSettingsVariant) *PutLlama
Settings used to install the inference model. These settings are specific to the `llama` service. API name: service_settings
type Request ¶
type Request struct {
// ChunkingSettings The chunking configuration object.
ChunkingSettings *types.InferenceChunkingSettings `json:"chunking_settings,omitempty"`
// Service The type of service supported for the specified task type. In this case,
// `llama`.
Service llamaservicetype.LlamaServiceType `json:"service"`
// ServiceSettings Settings used to install the inference model. These settings are specific to
// the `llama` service.
ServiceSettings types.LlamaServiceSettings `json:"service_settings"`
}
Request holds the request body struct for the package putllama
type Response ¶
type Response struct {
// ChunkingSettings Chunking configuration object
ChunkingSettings *types.InferenceChunkingSettings `json:"chunking_settings,omitempty"`
// InferenceId The inference Id
InferenceId string `json:"inference_id"`
// Service The service type
Service string `json:"service"`
// ServiceSettings Settings specific to the service
ServiceSettings json.RawMessage `json:"service_settings"`
// TaskSettings Task settings specific to the service and task type
TaskSettings json.RawMessage `json:"task_settings,omitempty"`
// TaskType The task type
TaskType tasktypellama.TaskTypeLlama `json:"task_type"`
}
Response holds the response body struct for the package putllama