mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 02:34:29 +00:00
37 lines
No EOL
858 B
JavaScript
37 lines
No EOL
858 B
JavaScript
const openai = require('openai');
|
|
|
|
// set DEBUG=true in env
|
|
process.env.DEBUG=false;
|
|
async function runOpenAI() {
|
|
const client = new openai.OpenAI({
|
|
apiKey: 'sk-JkKeNi6WpWDngBsghJ6B9g',
|
|
baseURL: 'http://0.0.0.0:8000'
|
|
});
|
|
|
|
try {
|
|
const response = await client.chat.completions.create({
|
|
model: 'sagemaker',
|
|
stream: true,
|
|
max_tokens: 1000,
|
|
messages: [
|
|
{
|
|
role: 'user',
|
|
content: 'write a 20 pg essay about YC ',
|
|
},
|
|
],
|
|
});
|
|
|
|
console.log(response);
|
|
for await (const chunk of response) {
|
|
console.log(chunk);
|
|
console.log(chunk.choices[0].delta.content);
|
|
}
|
|
} catch (error) {
|
|
console.log("got this exception from server");
|
|
console.error(error);
|
|
console.log("done with exception from proxy");
|
|
}
|
|
}
|
|
|
|
// Call the asynchronous function
|
|
runOpenAI(); |