in aidial_adapter_bedrock/llm/model/claude/v1_v2/adapter.py [0:0]
def convert_params(params: ModelParameters) -> Dict[str, Any]:
ret = {}
if params.max_tokens is not None:
ret["max_tokens_to_sample"] = params.max_tokens
else:
# The max tokens parameter is required for Anthropic models.
# Choosing reasonable default.
ret["max_tokens_to_sample"] = DEFAULT_MAX_TOKENS_ANTHROPIC
if params.stop:
ret["stop_sequences"] = params.stop
if params.temperature is not None:
ret["temperature"] = params.temperature
if params.top_p is not None:
ret["top_p"] = params.top_p
return ret