@@ -28,7 +28,7 @@ const client = new Groq({
2828
2929const chatCompletion = await client .chat .completions .create ({
3030 messages: [{ role: ' user' , content: ' Explain the importance of low latency LLMs' }],
31- model: ' llama3-8b-8192 ' ,
31+ model: ' openai/gpt-oss-20b ' ,
3232});
3333
3434console .log (chatCompletion .choices [0 ].message .content );
@@ -51,7 +51,7 @@ const params: Groq.Chat.CompletionCreateParams = {
5151 { role: ' system' , content: ' You are a helpful assistant.' },
5252 { role: ' user' , content: ' Explain the importance of low latency LLMs' },
5353 ],
54- model: ' llama3-8b-8192 ' ,
54+ model: ' openai/gpt-oss-20b ' ,
5555};
5656const chatCompletion: Groq .Chat .ChatCompletion = await client .chat .completions .create (params );
5757```
@@ -117,7 +117,7 @@ const chatCompletion = await client.chat.completions
117117 { role: ' system' , content: ' You are a helpful assistant.' },
118118 { role: ' user' , content: ' Explain the importance of low latency LLMs' },
119119 ],
120- model: ' llama3-8b-8192 ' ,
120+ model: ' openai/gpt-oss-20b ' ,
121121 })
122122 .catch (async (err ) => {
123123 if (err instanceof Groq .APIError ) {
@@ -159,7 +159,7 @@ const client = new Groq({
159159});
160160
161161// Or, configure per-request:
162- await client .chat .completions .create ({ messages: [{ role: ' system' , content: ' You are a helpful assistant.' }, { role: ' user' , content: ' Explain the importance of low latency LLMs' }], model: ' llama3-8b-8192 ' }, {
162+ await client .chat .completions .create ({ messages: [{ role: ' system' , content: ' You are a helpful assistant.' }, { role: ' user' , content: ' Explain the importance of low latency LLMs' }], model: ' openai/gpt-oss-20b ' }, {
163163 maxRetries: 5 ,
164164});
165165```
@@ -176,7 +176,7 @@ const client = new Groq({
176176});
177177
178178// Override per-request:
179- await client .chat .completions .create ({ messages: [{ role: ' system' , content: ' You are a helpful assistant.' }, { role: ' user' , content: ' Explain the importance of low latency LLMs' }], model: ' llama3-8b-8192 ' }, {
179+ await client .chat .completions .create ({ messages: [{ role: ' system' , content: ' You are a helpful assistant.' }, { role: ' user' , content: ' Explain the importance of low latency LLMs' }], model: ' openai/gpt-oss-20b ' }, {
180180 timeout: 5 * 1000 ,
181181});
182182```
@@ -203,7 +203,7 @@ const response = await client.chat.completions
203203 { role: ' system' , content: ' You are a helpful assistant.' },
204204 { role: ' user' , content: ' Explain the importance of low latency LLMs' },
205205 ],
206- model: ' llama3-8b-8192 ' ,
206+ model: ' openai/gpt-oss-20b ' ,
207207 })
208208 .asResponse ();
209209console .log (response .headers .get (' X-My-Header' ));
@@ -215,7 +215,7 @@ const { data: chatCompletion, response: raw } = await client.chat.completions
215215 { role: ' system' , content: ' You are a helpful assistant.' },
216216 { role: ' user' , content: ' Explain the importance of low latency LLMs' },
217217 ],
218- model: ' llama3-8b-8192 ' ,
218+ model: ' openai/gpt-oss-20b ' ,
219219 })
220220 .withResponse ();
221221console .log (raw .headers .get (' X-My-Header' ));
@@ -329,7 +329,7 @@ await client.chat.completions.create(
329329 { role: ' system' , content: ' You are a helpful assistant.' },
330330 { role: ' user' , content: ' Explain the importance of low latency LLMs' },
331331 ],
332- model: ' llama3-8b-8192 ' ,
332+ model: ' openai/gpt-oss-20b ' ,
333333 },
334334 {
335335 httpAgent: new http .Agent ({ keepAlive: false }),
0 commit comments