what parameter should i pass to get the same prompts_tokens as returned by openai.
{
"messages": [
{
"role": "system",
"content": ""
},
{
"role": "user",
"content": "你好"
},
{
"role": "assistant",
"content": "你好!有什么我可以帮助你的吗?"
},
{
"role": "user",
"content": "介绍下你自己"
}
]
}
This is my prompt, how to stringfy it and pass to encode?
what parameter should i pass to get the same prompts_tokens as returned by openai.
{ "messages": [ { "role": "system", "content": "" }, { "role": "user", "content": "你好" }, { "role": "assistant", "content": "你好!有什么我可以帮助你的吗?" }, { "role": "user", "content": "介绍下你自己" } ] }This is my prompt, how to stringfy it and pass to encode?