mirror of
https://github.com/tvytlx/ai-agent-deep-dive.git
synced 2026-04-20 11:55:09 +08:00
Add extracted source directory and README navigation
This commit is contained in:
123
extracted-source/node_modules/@anthropic-ai/sdk/resources/messages/messages.mjs
generated
vendored
Normal file
123
extracted-source/node_modules/@anthropic-ai/sdk/resources/messages/messages.mjs
generated
vendored
Normal file
@@ -0,0 +1,123 @@
|
||||
// File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
|
||||
import { APIResource } from "../../core/resource.mjs";
|
||||
import { buildHeaders } from "../../internal/headers.mjs";
|
||||
import { stainlessHelperHeader } from "../../lib/stainless-helper-header.mjs";
|
||||
import { MessageStream } from "../../lib/MessageStream.mjs";
|
||||
import { parseMessage, } from "../../lib/parser.mjs";
|
||||
import * as BatchesAPI from "./batches.mjs";
|
||||
import { Batches, } from "./batches.mjs";
|
||||
import { MODEL_NONSTREAMING_TOKENS } from "../../internal/constants.mjs";
|
||||
export class Messages extends APIResource {
|
||||
constructor() {
|
||||
super(...arguments);
|
||||
this.batches = new BatchesAPI.Batches(this._client);
|
||||
}
|
||||
create(body, options) {
|
||||
if (body.model in DEPRECATED_MODELS) {
|
||||
console.warn(`The model '${body.model}' is deprecated and will reach end-of-life on ${DEPRECATED_MODELS[body.model]}\nPlease migrate to a newer model. Visit https://docs.anthropic.com/en/docs/resources/model-deprecations for more information.`);
|
||||
}
|
||||
if (body.model in MODELS_TO_WARN_WITH_THINKING_ENABLED &&
|
||||
body.thinking &&
|
||||
body.thinking.type === 'enabled') {
|
||||
console.warn(`Using Claude with ${body.model} and 'thinking.type=enabled' is deprecated. Use 'thinking.type=adaptive' instead which results in better model performance in our testing: https://platform.claude.com/docs/en/build-with-claude/adaptive-thinking`);
|
||||
}
|
||||
let timeout = this._client._options.timeout;
|
||||
if (!body.stream && timeout == null) {
|
||||
const maxNonstreamingTokens = MODEL_NONSTREAMING_TOKENS[body.model] ?? undefined;
|
||||
timeout = this._client.calculateNonstreamingTimeout(body.max_tokens, maxNonstreamingTokens);
|
||||
}
|
||||
// Collect helper info from tools and messages
|
||||
const helperHeader = stainlessHelperHeader(body.tools, body.messages);
|
||||
return this._client.post('/v1/messages', {
|
||||
body,
|
||||
timeout: timeout ?? 600000,
|
||||
...options,
|
||||
headers: buildHeaders([helperHeader, options?.headers]),
|
||||
stream: body.stream ?? false,
|
||||
});
|
||||
}
|
||||
/**
|
||||
* Send a structured list of input messages with text and/or image content, along with an expected `output_config.format` and
|
||||
* the response will be automatically parsed and available in the `parsed_output` property of the message.
|
||||
*
|
||||
* @example
|
||||
* ```ts
|
||||
* const message = await client.messages.parse({
|
||||
* model: 'claude-sonnet-4-5-20250929',
|
||||
* max_tokens: 1024,
|
||||
* messages: [{ role: 'user', content: 'What is 2+2?' }],
|
||||
* output_config: {
|
||||
* format: zodOutputFormat(z.object({ answer: z.number() })),
|
||||
* },
|
||||
* });
|
||||
*
|
||||
* console.log(message.parsed_output?.answer); // 4
|
||||
* ```
|
||||
*/
|
||||
parse(params, options) {
|
||||
return this.create(params, options).then((message) => parseMessage(message, params, { logger: this._client.logger ?? console }));
|
||||
}
|
||||
/**
|
||||
* Create a Message stream.
|
||||
*
|
||||
* If `output_config.format` is provided with a parseable format (like `zodOutputFormat()`),
|
||||
* the final message will include a `parsed_output` property with the parsed content.
|
||||
*
|
||||
* @example
|
||||
* ```ts
|
||||
* const stream = client.messages.stream({
|
||||
* model: 'claude-sonnet-4-5-20250929',
|
||||
* max_tokens: 1024,
|
||||
* messages: [{ role: 'user', content: 'What is 2+2?' }],
|
||||
* output_config: {
|
||||
* format: zodOutputFormat(z.object({ answer: z.number() })),
|
||||
* },
|
||||
* });
|
||||
*
|
||||
* const message = await stream.finalMessage();
|
||||
* console.log(message.parsed_output?.answer); // 4
|
||||
* ```
|
||||
*/
|
||||
stream(body, options) {
|
||||
return MessageStream.createMessage(this, body, options, { logger: this._client.logger ?? console });
|
||||
}
|
||||
/**
|
||||
* Count the number of tokens in a Message.
|
||||
*
|
||||
* The Token Count API can be used to count the number of tokens in a Message,
|
||||
* including tools, images, and documents, without creating it.
|
||||
*
|
||||
* Learn more about token counting in our
|
||||
* [user guide](https://docs.claude.com/en/docs/build-with-claude/token-counting)
|
||||
*
|
||||
* @example
|
||||
* ```ts
|
||||
* const messageTokensCount =
|
||||
* await client.messages.countTokens({
|
||||
* messages: [{ content: 'string', role: 'user' }],
|
||||
* model: 'claude-opus-4-6',
|
||||
* });
|
||||
* ```
|
||||
*/
|
||||
countTokens(body, options) {
|
||||
return this._client.post('/v1/messages/count_tokens', { body, ...options });
|
||||
}
|
||||
}
|
||||
const DEPRECATED_MODELS = {
|
||||
'claude-1.3': 'November 6th, 2024',
|
||||
'claude-1.3-100k': 'November 6th, 2024',
|
||||
'claude-instant-1.1': 'November 6th, 2024',
|
||||
'claude-instant-1.1-100k': 'November 6th, 2024',
|
||||
'claude-instant-1.2': 'November 6th, 2024',
|
||||
'claude-3-sonnet-20240229': 'July 21st, 2025',
|
||||
'claude-3-opus-20240229': 'January 5th, 2026',
|
||||
'claude-2.1': 'July 21st, 2025',
|
||||
'claude-2.0': 'July 21st, 2025',
|
||||
'claude-3-7-sonnet-latest': 'February 19th, 2026',
|
||||
'claude-3-7-sonnet-20250219': 'February 19th, 2026',
|
||||
'claude-3-5-haiku-latest': 'February 19th, 2026',
|
||||
'claude-3-5-haiku-20241022': 'February 19th, 2026',
|
||||
};
|
||||
const MODELS_TO_WARN_WITH_THINKING_ENABLED = ['claude-opus-4-6'];
|
||||
Messages.Batches = Batches;
|
||||
//# sourceMappingURL=messages.mjs.map
|
||||
Reference in New Issue
Block a user