Skip to content

How can I get prompts_tokens? #10

@magelikescoke

Description

@magelikescoke

what parameter should i pass to get the same prompts_tokens as returned by openai.

{
  "messages": [
        {
            "role": "system",
            "content": ""
        },
        {
            "role": "user",
            "content": "你好"
        },
        {
            "role": "assistant",
            "content": "你好!有什么我可以帮助你的吗?"
        },
        {
            "role": "user",
            "content": "介绍下你自己"
        }
    ]
}

This is my prompt, how to stringfy it and pass to encode?

Metadata

Metadata

Assignees

No one assigned

    Labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions