@@ -2,25 +2,25 @@ import * as utils from './utils.js'
2
2
import 'whatwg-fetch'
3
3
4
4
import type {
5
- Fetch ,
5
+ ChatRequest ,
6
+ ChatResponse ,
6
7
Config ,
7
- GenerateRequest ,
8
- PullRequest ,
9
- PushRequest ,
8
+ CopyRequest ,
9
+ CreateRequest ,
10
+ DeleteRequest ,
10
11
EmbeddingsRequest ,
11
- GenerateResponse ,
12
12
EmbeddingsResponse ,
13
+ ErrorResponse ,
14
+ Fetch ,
15
+ GenerateRequest ,
16
+ GenerateResponse ,
13
17
ListResponse ,
14
18
ProgressResponse ,
15
- ErrorResponse ,
16
- StatusResponse ,
17
- DeleteRequest ,
18
- CopyRequest ,
19
- ShowResponse ,
19
+ PullRequest ,
20
+ PushRequest ,
20
21
ShowRequest ,
21
- ChatRequest ,
22
- ChatResponse ,
23
- CreateRequest ,
22
+ ShowResponse ,
23
+ StatusResponse ,
24
24
} from './interfaces.js'
25
25
26
26
export class Ollama {
@@ -50,6 +50,17 @@ export class Ollama {
50
50
this . abortController = new AbortController ( )
51
51
}
52
52
53
+ /**
54
+ * Processes a request to the Ollama server. If the request is streamable, it will return an
55
+ * AsyncGenerator that yields the response messages. Otherwise, it will return the response
56
+ * object.
57
+ * @param endpoint {string} - The endpoint to send the request to.
58
+ * @param request {object} - The request object to send to the endpoint.
59
+ * @protected {T | AsyncGenerator<T>} - The response object or an AsyncGenerator that yields
60
+ * response messages.
61
+ * @throws {Error } - If the response body is missing or if the response is an error.
62
+ * @returns {Promise<T | AsyncGenerator<T>> } - The response object or an AsyncGenerator that yields the streamed response.
63
+ */
53
64
protected async processStreamableRequest < T extends object > (
54
65
endpoint : string ,
55
66
request : { stream ?: boolean } & Record < string , any > ,
@@ -94,13 +105,17 @@ export class Ollama {
94
105
}
95
106
}
96
107
108
+ /**
109
+ * Encodes an image to base64 if it is a Uint8Array.
110
+ * @param image {Uint8Array | string} - The image to encode.
111
+ * @returns {Promise<string> } - The base64 encoded image.
112
+ */
97
113
async encodeImage ( image : Uint8Array | string ) : Promise < string > {
98
114
if ( typeof image !== 'string' ) {
99
115
// image is Uint8Array convert it to base64
100
116
const uint8Array = new Uint8Array ( image )
101
117
const numberArray = Array . from ( uint8Array )
102
- const base64String = btoa ( String . fromCharCode . apply ( null , numberArray ) )
103
- return base64String
118
+ return btoa ( String . fromCharCode . apply ( null , numberArray ) )
104
119
}
105
120
// the string may be base64 encoded
106
121
return image
@@ -110,7 +125,12 @@ export class Ollama {
110
125
request : GenerateRequest & { stream : true } ,
111
126
) : Promise < AsyncGenerator < GenerateResponse > >
112
127
generate ( request : GenerateRequest & { stream ?: false } ) : Promise < GenerateResponse >
113
-
128
+ /**
129
+ * Generates a response from a text prompt.
130
+ * @param request {GenerateRequest} - The request object.
131
+ * @returns {Promise<GenerateResponse | AsyncGenerator<GenerateResponse>> } - The response object or
132
+ * an AsyncGenerator that yields response messages.
133
+ */
114
134
async generate (
115
135
request : GenerateRequest ,
116
136
) : Promise < GenerateResponse | AsyncGenerator < GenerateResponse > > {
@@ -122,7 +142,14 @@ export class Ollama {
122
142
123
143
chat ( request : ChatRequest & { stream : true } ) : Promise < AsyncGenerator < ChatResponse > >
124
144
chat ( request : ChatRequest & { stream ?: false } ) : Promise < ChatResponse >
125
-
145
+ /**
146
+ * Chats with the model. The request object can contain messages with images that are either
147
+ * Uint8Arrays or base64 encoded strings. The images will be base64 encoded before sending the
148
+ * request.
149
+ * @param request {ChatRequest} - The request object.
150
+ * @returns {Promise<ChatResponse | AsyncGenerator<ChatResponse>> } - The response object or an
151
+ * AsyncGenerator that yields response messages.
152
+ */
126
153
async chat ( request : ChatRequest ) : Promise < ChatResponse | AsyncGenerator < ChatResponse > > {
127
154
if ( request . messages ) {
128
155
for ( const message of request . messages ) {
@@ -140,7 +167,11 @@ export class Ollama {
140
167
request : CreateRequest & { stream : true } ,
141
168
) : Promise < AsyncGenerator < ProgressResponse > >
142
169
create ( request : CreateRequest & { stream ?: false } ) : Promise < ProgressResponse >
143
-
170
+ /**
171
+ * Creates a new model from a stream of data.
172
+ * @param request {CreateRequest} - The request object.
173
+ * @returns {Promise<ProgressResponse | AsyncGenerator<ProgressResponse>> } - The response object or a stream of progress responses.
174
+ */
144
175
async create (
145
176
request : CreateRequest ,
146
177
) : Promise < ProgressResponse | AsyncGenerator < ProgressResponse > > {
@@ -154,7 +185,13 @@ export class Ollama {
154
185
155
186
pull ( request : PullRequest & { stream : true } ) : Promise < AsyncGenerator < ProgressResponse > >
156
187
pull ( request : PullRequest & { stream ?: false } ) : Promise < ProgressResponse >
157
-
188
+ /**
189
+ * Pulls a model from the Ollama registry. The request object can contain a stream flag to indicate if the
190
+ * response should be streamed.
191
+ * @param request {PullRequest} - The request object.
192
+ * @returns {Promise<ProgressResponse | AsyncGenerator<ProgressResponse>> } - The response object or
193
+ * an AsyncGenerator that yields response messages.
194
+ */
158
195
async pull (
159
196
request : PullRequest ,
160
197
) : Promise < ProgressResponse | AsyncGenerator < ProgressResponse > > {
@@ -167,7 +204,13 @@ export class Ollama {
167
204
168
205
push ( request : PushRequest & { stream : true } ) : Promise < AsyncGenerator < ProgressResponse > >
169
206
push ( request : PushRequest & { stream ?: false } ) : Promise < ProgressResponse >
170
-
207
+ /**
208
+ * Pushes a model to the Ollama registry. The request object can contain a stream flag to indicate if the
209
+ * response should be streamed.
210
+ * @param request {PushRequest} - The request object.
211
+ * @returns {Promise<ProgressResponse | AsyncGenerator<ProgressResponse>> } - The response object or
212
+ * an AsyncGenerator that yields response messages.
213
+ */
171
214
async push (
172
215
request : PushRequest ,
173
216
) : Promise < ProgressResponse | AsyncGenerator < ProgressResponse > > {
@@ -178,38 +221,62 @@ export class Ollama {
178
221
} )
179
222
}
180
223
224
+ /**
225
+ * Deletes a model from the server. The request object should contain the name of the model to
226
+ * delete.
227
+ * @param request {DeleteRequest} - The request object.
228
+ * @returns {Promise<StatusResponse> } - The response object.
229
+ */
181
230
async delete ( request : DeleteRequest ) : Promise < StatusResponse > {
182
231
await utils . del ( this . fetch , `${ this . config . host } /api/delete` , {
183
232
name : request . model ,
184
233
} )
185
234
return { status : 'success' }
186
235
}
187
236
237
+ /**
238
+ * Copies a model from one name to another. The request object should contain the name of the
239
+ * model to copy and the new name.
240
+ * @param request {CopyRequest} - The request object.
241
+ * @returns {Promise<StatusResponse> } - The response object.
242
+ */
188
243
async copy ( request : CopyRequest ) : Promise < StatusResponse > {
189
244
await utils . post ( this . fetch , `${ this . config . host } /api/copy` , { ...request } )
190
245
return { status : 'success' }
191
246
}
192
247
248
+ /**
249
+ * Lists the models on the server.
250
+ * @returns {Promise<ListResponse> } - The response object.
251
+ * @throws {Error } - If the response body is missing.
252
+ */
193
253
async list ( ) : Promise < ListResponse > {
194
254
const response = await utils . get ( this . fetch , `${ this . config . host } /api/tags` )
195
- const listResponse = ( await response . json ( ) ) as ListResponse
196
- return listResponse
255
+ return ( await response . json ( ) ) as ListResponse
197
256
}
198
257
258
+ /**
259
+ * Shows the metadata of a model. The request object should contain the name of the model.
260
+ * @param request {ShowRequest} - The request object.
261
+ * @returns {Promise<ShowResponse> } - The response object.
262
+ */
199
263
async show ( request : ShowRequest ) : Promise < ShowResponse > {
200
264
const response = await utils . post ( this . fetch , `${ this . config . host } /api/show` , {
201
265
...request ,
202
266
} )
203
- const showResponse = ( await response . json ( ) ) as ShowResponse
204
- return showResponse
267
+ return ( await response . json ( ) ) as ShowResponse
205
268
}
206
269
270
+ /**
271
+ * Embeds a text prompt into a vector.
272
+ * @param request {EmbeddingsRequest} - The request object.
273
+ * @returns {Promise<EmbeddingsResponse> } - The response object.
274
+ */
207
275
async embeddings ( request : EmbeddingsRequest ) : Promise < EmbeddingsResponse > {
208
276
const response = await utils . post ( this . fetch , `${ this . config . host } /api/embeddings` , {
209
277
...request ,
210
278
} )
211
- const embeddingsResponse = ( await response . json ( ) ) as EmbeddingsResponse
212
- return embeddingsResponse
279
+ return ( await response . json ( ) ) as EmbeddingsResponse
213
280
}
214
281
}
215
282
0 commit comments