typedapi/types/xlmrobertatokenizationconfig.go (100 lines of code) (raw):

// Licensed to Elasticsearch B.V. under one or more contributor // license agreements. See the NOTICE file distributed with // this work for additional information regarding copyright // ownership. Elasticsearch B.V. licenses this file to you under // the Apache License, Version 2.0 (the "License"); you may // not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, // software distributed under the License is distributed on an // "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY // KIND, either express or implied. See the License for the // specific language governing permissions and limitations // under the License. // Code generated from the elasticsearch-specification DO NOT EDIT. // https://github.com/elastic/elasticsearch-specification/tree/cbfcc73d01310bed2a480ec35aaef98138b598e5 package types import ( "bytes" "encoding/json" "errors" "fmt" "io" "strconv" "github.com/elastic/go-elasticsearch/v9/typedapi/types/enums/tokenizationtruncate" ) // XlmRobertaTokenizationConfig type. // // https://github.com/elastic/elasticsearch-specification/blob/cbfcc73d01310bed2a480ec35aaef98138b598e5/specification/ml/_types/inference.ts#L200-L200 type XlmRobertaTokenizationConfig struct { // DoLowerCase Should the tokenizer lower case the text DoLowerCase *bool `json:"do_lower_case,omitempty"` // MaxSequenceLength Maximum input sequence length for the model MaxSequenceLength *int `json:"max_sequence_length,omitempty"` // Span Tokenization spanning options. Special value of -1 indicates no spanning // takes place Span *int `json:"span,omitempty"` // Truncate Should tokenization input be automatically truncated before sending to the // model for inference Truncate *tokenizationtruncate.TokenizationTruncate `json:"truncate,omitempty"` // WithSpecialTokens Is tokenization completed with special tokens WithSpecialTokens *bool `json:"with_special_tokens,omitempty"` } func (s *XlmRobertaTokenizationConfig) UnmarshalJSON(data []byte) error { dec := json.NewDecoder(bytes.NewReader(data)) for { t, err := dec.Token() if err != nil { if errors.Is(err, io.EOF) { break } return err } switch t { case "do_lower_case": var tmp any dec.Decode(&tmp) switch v := tmp.(type) { case string: value, err := strconv.ParseBool(v) if err != nil { return fmt.Errorf("%s | %w", "DoLowerCase", err) } s.DoLowerCase = &value case bool: s.DoLowerCase = &v } case "max_sequence_length": var tmp any dec.Decode(&tmp) switch v := tmp.(type) { case string: value, err := strconv.Atoi(v) if err != nil { return fmt.Errorf("%s | %w", "MaxSequenceLength", err) } s.MaxSequenceLength = &value case float64: f := int(v) s.MaxSequenceLength = &f } case "span": var tmp any dec.Decode(&tmp) switch v := tmp.(type) { case string: value, err := strconv.Atoi(v) if err != nil { return fmt.Errorf("%s | %w", "Span", err) } s.Span = &value case float64: f := int(v) s.Span = &f } case "truncate": if err := dec.Decode(&s.Truncate); err != nil { return fmt.Errorf("%s | %w", "Truncate", err) } case "with_special_tokens": var tmp any dec.Decode(&tmp) switch v := tmp.(type) { case string: value, err := strconv.ParseBool(v) if err != nil { return fmt.Errorf("%s | %w", "WithSpecialTokens", err) } s.WithSpecialTokens = &value case bool: s.WithSpecialTokens = &v } } } return nil } // NewXlmRobertaTokenizationConfig returns a XlmRobertaTokenizationConfig. func NewXlmRobertaTokenizationConfig() *XlmRobertaTokenizationConfig { r := &XlmRobertaTokenizationConfig{} return r } // true type XlmRobertaTokenizationConfigVariant interface { XlmRobertaTokenizationConfigCaster() *XlmRobertaTokenizationConfig } func (s *XlmRobertaTokenizationConfig) XlmRobertaTokenizationConfigCaster() *XlmRobertaTokenizationConfig { return s }