Notice how the completion_tokens are 256; the model wasn't actually done, but got cut off because ma

Notice how the completion_tokens are 256; the model wasn't actually done, but got cut off because max_tokens defaults to 256 as of a few weeks ago. You'll want to specify a higher maxTokens in your request in order to allow the model to finish it's thought.
Was this page helpful?