Chat
The ChatCompletion
endpoint allows you to have a chat conversation with a model.
chat = anthropic.ChatCompletion.create(
model="anthropic.claude-v2",
messages=[
{"role": "user", "content": "Hello!"},
# { role: "system", content: "Hi how are you doing today." },
],
max_tokens_to_sample=300
)
Usage
import os
from bedrock_anthropic import AnthropicBedrock
anthropic = AnthropicBedrock(
access_key=os.getenv("AWS_ACCESS_KEY"),
secret_key=os.getenv("AWS_SECRET_KEY"),
region=os.getenv("AWS_REGION")
)
chat = anthropic.ChatCompletion.create(
model="anthropic.claude-v2",
max_tokens_to_sample=300,
messages=[
{"role": "user", "content": "Hello!"},
# { role: "system", content: "Hi how are you doing today." },
]
)
print(chat["messages"])
Configuration
model
The model that will complete your prompt. Refer to the models page.
anthropic.ChatCompletion.create(
model="anthropic.claude-v2",
max_tokens_to_sample=300,
messages=[
{"role": "user", "content": "Hello!"},
# { role: "system", content: "Hi how are you doing today." },
]
)
messages
The messages to chat with the model.
anthropic.ChatCompletion.create(
model="anthropic.claude-v2",
max_tokens_to_sample=300,
messages=[
{"role": "user", "content": "Hello!"},
# { role: "system", content: "Hi how are you doing today." },
]
)
max_tokens_to_sample (optional)
The maximum number of tokens to generate before stopping.
- Default:
256
- Range depends on the model, refer to the models page.
anthropic.ChatCompletion.create(
model="anthropic.claude-v2",
max_tokens_to_sample=300,
messages=[
{"role": "user", "content": "Hello!"},
# { role: "system", content: "Hi how are you doing today." },
]
)
stop_sequences (optional)
Sequences that will cause the model to stop generating completion text.
- Default:
[]
anthropic.ChatCompletion.create(
model="anthropic.claude-v2",
max_tokens_to_sample=300,
messages=[
{"role": "user", "content": "Hello!"},
# { role: "system", content: "Hi how are you doing today." },
],
stop_sequences=[
"sequence"
]
)
temperature (optional)
Amount of randomness injected in the response.
- Default:
1
- Range:
0-1
anthropic.ChatCompletion.create(
model="anthropic.claude-v2",
max_tokens_to_sample=300,
messages=[
{"role": "user", "content": "Hello!"},
# { role: "system", content: "Hi how are you doing today." },
],
temperature=0.7
)
top_p (optional)
Use nucleus sampling.
- Default:
1
- Range:
0-1
anthropic.ChatCompletion.create(
model="anthropic.claude-v2",
max_tokens_to_sample=300,
messages=[
{"role": "user", "content": "Hello!"},
# { role: "system", content: "Hi how are you doing today." },
],
top_p=0.7
)
top_k (optional)
Only sample from the top K options for each subsequent token.
- Default:
250
- Range:
0-500
anthropic.ChatCompletion.create(
model="anthropic.claude-v2",
max_tokens_to_sample=300,
messages=[
{"role": "user", "content": "Hello!"},
# { role: "system", content: "Hi how are you doing today." },
],
top_k=250
)