in aidial_adapter_bedrock/llm/model/amazon.py [0:0]
def convert_params(params: ModelParameters) -> Dict[str, Any]:
ret = {}
if params.temperature is not None:
ret["temperature"] = params.temperature
if params.top_p is not None:
ret["topP"] = params.top_p
if params.max_tokens is not None:
ret["maxTokenCount"] = params.max_tokens
else:
# The default for max tokens is 128, which is too small for most use cases.
# Choosing reasonable default.
ret["maxTokenCount"] = DEFAULT_MAX_TOKENS_AMAZON
# NOTE: Amazon Titan (amazon.titan-tg1-large) currently only supports
# stop sequences matching pattern "$\|+".
# if params.stop is not None:
# ret["stopSequences"] = params.stop
return ret