lib/anthropic/resources/messages.rb (55 lines of code) (raw):
# frozen_string_literal: true
module Anthropic
module Resources
class Messages
# @return [Anthropic::Resources::Messages::Batches]
attr_reader :batches
# See {Anthropic::Resources::Messages#stream_raw} for streaming counterpart.
#
# Send a structured list of input messages with text and/or image content, and the
# model will generate the next message in the conversation.
#
# The Messages API can be used for either single queries or stateless multi-turn
# conversations.
#
# Learn more about the Messages API in our [user guide](/en/docs/initial-setup)
#
# @overload create(max_tokens:, messages:, model:, metadata: nil, stop_sequences: nil, system_: nil, temperature: nil, thinking: nil, tool_choice: nil, tools: nil, top_k: nil, top_p: nil, request_options: {})
#
# @param max_tokens [Integer]
# @param messages [Array<Anthropic::Models::MessageParam>]
# @param model [Symbol, String, Anthropic::Models::Model]
# @param metadata [Anthropic::Models::Metadata]
# @param stop_sequences [Array<String>]
# @param system_ [String, Array<Anthropic::Models::TextBlockParam>]
# @param temperature [Float]
# @param thinking [Anthropic::Models::ThinkingConfigEnabled, Anthropic::Models::ThinkingConfigDisabled]
# @param tool_choice [Anthropic::Models::ToolChoiceAuto, Anthropic::Models::ToolChoiceAny, Anthropic::Models::ToolChoiceTool, Anthropic::Models::ToolChoiceNone]
# @param tools [Array<Anthropic::Models::Tool, Anthropic::Models::ToolBash20250124, Anthropic::Models::ToolTextEditor20250124>]
# @param top_k [Integer]
# @param top_p [Float]
# @param request_options [Anthropic::RequestOptions, Hash{Symbol=>Object}, nil]
#
# @return [Anthropic::Models::Message]
#
# @see Anthropic::Models::MessageCreateParams
def create(params)
parsed, options = Anthropic::Models::MessageCreateParams.dump_request(params)
if parsed[:stream]
message = "Please use `#stream_raw` for the streaming use case."
raise ArgumentError.new(message)
end
@client.request(
method: :post,
path: "v1/messages",
body: parsed,
model: Anthropic::Models::Message,
options: options
)
end
def stream
raise NotImplementedError.new("higher level helpers are coming soon!")
end
# See {Anthropic::Resources::Messages#create} for non-streaming counterpart.
#
# Send a structured list of input messages with text and/or image content, and the
# model will generate the next message in the conversation.
#
# The Messages API can be used for either single queries or stateless multi-turn
# conversations.
#
# Learn more about the Messages API in our [user guide](/en/docs/initial-setup)
#
# @overload stream_raw(max_tokens:, messages:, model:, metadata: nil, stop_sequences: nil, system_: nil, temperature: nil, thinking: nil, tool_choice: nil, tools: nil, top_k: nil, top_p: nil, request_options: {})
#
# @param max_tokens [Integer]
# @param messages [Array<Anthropic::Models::MessageParam>]
# @param model [Symbol, String, Anthropic::Models::Model]
# @param metadata [Anthropic::Models::Metadata]
# @param stop_sequences [Array<String>]
# @param system_ [String, Array<Anthropic::Models::TextBlockParam>]
# @param temperature [Float]
# @param thinking [Anthropic::Models::ThinkingConfigEnabled, Anthropic::Models::ThinkingConfigDisabled]
# @param tool_choice [Anthropic::Models::ToolChoiceAuto, Anthropic::Models::ToolChoiceAny, Anthropic::Models::ToolChoiceTool, Anthropic::Models::ToolChoiceNone]
# @param tools [Array<Anthropic::Models::Tool, Anthropic::Models::ToolBash20250124, Anthropic::Models::ToolTextEditor20250124>]
# @param top_k [Integer]
# @param top_p [Float]
# @param request_options [Anthropic::RequestOptions, Hash{Symbol=>Object}, nil]
#
# @return [Anthropic::Internal::Stream<Anthropic::Models::RawMessageStartEvent, Anthropic::Models::RawMessageDeltaEvent, Anthropic::Models::RawMessageStopEvent, Anthropic::Models::RawContentBlockStartEvent, Anthropic::Models::RawContentBlockDeltaEvent, Anthropic::Models::RawContentBlockStopEvent>]
#
# @see Anthropic::Models::MessageCreateParams
def stream_raw(params)
parsed, options = Anthropic::Models::MessageCreateParams.dump_request(params)
unless parsed.fetch(:stream, true)
message = "Please use `#create` for the non-streaming use case."
raise ArgumentError.new(message)
end
parsed.store(:stream, true)
@client.request(
method: :post,
path: "v1/messages",
headers: {"accept" => "text/event-stream"},
body: parsed,
stream: Anthropic::Internal::Stream,
model: Anthropic::Models::RawMessageStreamEvent,
options: options
)
end
# Count the number of tokens in a Message.
#
# The Token Count API can be used to count the number of tokens in a Message,
# including tools, images, and documents, without creating it.
#
# Learn more about token counting in our
# [user guide](/en/docs/build-with-claude/token-counting)
#
# @overload count_tokens(messages:, model:, system_: nil, thinking: nil, tool_choice: nil, tools: nil, request_options: {})
#
# @param messages [Array<Anthropic::Models::MessageParam>]
# @param model [Symbol, String, Anthropic::Models::Model]
# @param system_ [String, Array<Anthropic::Models::TextBlockParam>]
# @param thinking [Anthropic::Models::ThinkingConfigEnabled, Anthropic::Models::ThinkingConfigDisabled]
# @param tool_choice [Anthropic::Models::ToolChoiceAuto, Anthropic::Models::ToolChoiceAny, Anthropic::Models::ToolChoiceTool, Anthropic::Models::ToolChoiceNone]
# @param tools [Array<Anthropic::Models::Tool, Anthropic::Models::ToolBash20250124, Anthropic::Models::ToolTextEditor20250124>]
# @param request_options [Anthropic::RequestOptions, Hash{Symbol=>Object}, nil]
#
# @return [Anthropic::Models::MessageTokensCount]
#
# @see Anthropic::Models::MessageCountTokensParams
def count_tokens(params)
parsed, options = Anthropic::Models::MessageCountTokensParams.dump_request(params)
@client.request(
method: :post,
path: "v1/messages/count_tokens",
body: parsed,
model: Anthropic::Models::MessageTokensCount,
options: options
)
end
# @api private
#
# @param client [Anthropic::Client]
def initialize(client:)
@client = client
@batches = Anthropic::Resources::Messages::Batches.new(client: client)
end
end
end
end