lib/Search/Elasticsearch/Client/8_0/Direct/Inference.pm (9 lines of code) (raw):
# Licensed to Elasticsearch B.V. under one or more contributor
# license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright
# ownership. Elasticsearch B.V. licenses this file to you under
# the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
package Search::Elasticsearch::Client::8_0::Direct::Inference;
use Moo;
with 'Search::Elasticsearch::Client::8_0::Role::API';
with 'Search::Elasticsearch::Role::Client::Direct';
use namespace::clean;
__PACKAGE__->_install_api('inference');
1;
__END__
# ABSTRACT: The inference APIs enable you to use certain services,
such as ELSER, OpenAI, or Hugging Face, in your cluster.
=head2 DESCRIPTION
The full documentation for Inference is available here:
L<https://www.elastic.co/guide/en/elasticsearch/reference/current/inference-apis.html>
=head1 FOLLOW METHODS
=head2 C<follow()>
my $response = $es->inference->put_model(
'task_type' => 'sparse_embedding',
'model_id' => 'my-elser-model',
'body' => {
'service' => 'elser',
'service_settings' => {
'num_allocations' => 1,
'num_threads' => 1
},
'task_settings' => {}
}
);