OpenAI Inference Protocol Using Curl
Consider this example for OpenAI Inference Protocol Using Curl.
An example inference payload for the OpenAI Protocol:
# cat ./llama-input.json
{ "messages": [ { "content": "You are a polite and respectful chatbot helping people plan a vacation.", "role": "system" }, { "content": "What should I do for a 4 day vacation in Spain?", "role": "user" } ], "model": "meta/llama-3_1-8b-instruct", "max_tokens": 200, "top_p": 1, "n": 1, "stream": false, "stop": "\n", "frequency_penalty": 0.0 }
curl -H "Content-Type: application/json" -H "Authorization: Bearer ${CDP_TOKEN}" "https://${DOMAIN}/namespaces/serving-default/endpoints/llama-3-1/v1/chat/completions" -d @./llama-input.json
You will receive response similar to the following:
Spain offers a diverse range of experiences, making it perfect for a 4-day vacation…