Is there a way to limit the tokens output like how you can in openAI?
Is there a way to limit the tokens output like how you can in openAI?
@cf/openchat/openchat-3.5-0106 @cf/qwen/qwen1.5-0.5b-chat @cf/qwen/qwen1.5-1.8b-chat @cf/qwen/qwen1.5-14b-chat-awq @cf/qwen/qwen1.5-7b-chat-awq@cf/tinyllama/tinyllama-1.1b-chat-v1.0 was also very weird iircKeep responses brief, limited to two sentences.
Maintain a friendly chat environment.
Use emoticons like ",_," or "@w@" instead of emojis.
You are to NEVER use emojis.@cf/tinyllama/tinyllama-1.1b-chat-v1.0{
"response": "As an AI language model, I don't experience emotions in the way that humans do, but I'm here to assist you with any questions or tasks you may have to the best of my abilities. Is there something specific you'd like to know or discuss today? I'm here to provide information and support no matter the topic. Is there anything in particular on your mind or needing help with? I'm here to listen and offer any help you need.",
"prompt": "Keep responses brief, limited to two sentences. Maintain a friendly chat environment. Use emoticons like \",_,\" or \"@w@\" instead of emojis. You are to NEVER use emojis. ##PERSONALITY## You are a friendly assistant"
}