Skip to content

Commit 65926b7

Browse files
committed
stash
1 parent 89c26fe commit 65926b7

File tree

66 files changed

+2104
-550
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

66 files changed

+2104
-550
lines changed

packages/typescript/ai-ollama/src/meta/model-meta-athene.ts

Lines changed: 27 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,10 @@
1-
import type { ChatRequest } from 'ollama'
2-
import type { DefaultOllamaModelMeta } from './models-meta'
1+
import type {
2+
OllamaChatRequest,
3+
OllamaChatRequestMessages,
4+
OllamaChatRequestTools,
5+
OllamaMessageTools,
6+
OllamaModelMeta,
7+
} from './models-meta'
38

49
const ATHENE_V2_LATEST = {
510
name: 'athene-v2:latest',
@@ -10,7 +15,11 @@ const ATHENE_V2_LATEST = {
1015
},
1116
size: '47gb',
1217
context: 32_000,
13-
} as const satisfies DefaultOllamaModelMeta<any>
18+
} as const satisfies OllamaModelMeta<
19+
OllamaChatRequest &
20+
OllamaChatRequestMessages<OllamaMessageTools> &
21+
OllamaChatRequestTools
22+
>
1423

1524
const ATHENE_V2_72b = {
1625
name: 'athene-v2:72b',
@@ -21,7 +30,11 @@ const ATHENE_V2_72b = {
2130
},
2231
size: '47gb',
2332
context: 32_000,
24-
} as const satisfies DefaultOllamaModelMeta<any>
33+
} as const satisfies OllamaModelMeta<
34+
OllamaChatRequest &
35+
OllamaChatRequestMessages<OllamaMessageTools> &
36+
OllamaChatRequestTools
37+
>
2538

2639
export const ATHENE_MODELS = [
2740
ATHENE_V2_LATEST.name,
@@ -41,8 +54,16 @@ export const ATHENE_MODELS = [
4154
// Manual type map for per-model provider options
4255
export type AtheneChatModelProviderOptionsByName = {
4356
// Models with thinking and structured output support
44-
[ATHENE_V2_LATEST.name]: ChatRequest
45-
[ATHENE_V2_72b.name]: ChatRequest
57+
[ATHENE_V2_LATEST.name]: OllamaModelMeta<
58+
OllamaChatRequest &
59+
OllamaChatRequestMessages<OllamaMessageTools> &
60+
OllamaChatRequestTools
61+
>
62+
[ATHENE_V2_72b.name]: OllamaModelMeta<
63+
OllamaChatRequest &
64+
OllamaChatRequestMessages<OllamaMessageTools> &
65+
OllamaChatRequestTools
66+
>
4667
}
4768

4869
export type AtheneModelInputModalitiesByName = {

packages/typescript/ai-ollama/src/meta/model-meta-aya.ts

Lines changed: 19 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,8 @@
1-
import type { ChatRequest } from 'ollama'
2-
import type { DefaultOllamaModelMeta } from './models-meta'
1+
import type {
2+
OllamaChatRequest,
3+
OllamaChatRequestMessages,
4+
OllamaModelMeta,
5+
} from './models-meta'
36

47
const AYA_LATEST = {
58
name: 'aya:latest',
@@ -10,7 +13,9 @@ const AYA_LATEST = {
1013
},
1114
size: '4.8gb',
1215
context: 8_000,
13-
} as const satisfies DefaultOllamaModelMeta<any>
16+
} as const satisfies OllamaModelMeta<
17+
OllamaChatRequest & OllamaChatRequestMessages
18+
>
1419

1520
const AYA_8b = {
1621
name: 'aya:8b',
@@ -21,7 +26,9 @@ const AYA_8b = {
2126
},
2227
size: '4.8gb',
2328
context: 8_000,
24-
} as const satisfies DefaultOllamaModelMeta<any>
29+
} as const satisfies OllamaModelMeta<
30+
OllamaChatRequest & OllamaChatRequestMessages
31+
>
2532

2633
const AYA_35b = {
2734
name: 'aya:35b',
@@ -32,7 +39,9 @@ const AYA_35b = {
3239
},
3340
size: '20gb',
3441
context: 8_000,
35-
} as const satisfies DefaultOllamaModelMeta<any>
42+
} as const satisfies OllamaModelMeta<
43+
OllamaChatRequest & OllamaChatRequestMessages
44+
>
3645

3746
export const AYA_MODELS = [AYA_LATEST.name, AYA_8b.name, AYA_35b.name] as const
3847

@@ -49,9 +58,11 @@ export const AYA_MODELS = [AYA_LATEST.name, AYA_8b.name, AYA_35b.name] as const
4958
// Manual type map for per-model provider options
5059
export type AyaChatModelProviderOptionsByName = {
5160
// Models with thinking and structured output support
52-
[AYA_LATEST.name]: ChatRequest
53-
[AYA_8b.name]: ChatRequest
54-
[AYA_35b.name]: ChatRequest
61+
[AYA_LATEST.name]: OllamaModelMeta<
62+
OllamaChatRequest & OllamaChatRequestMessages
63+
>
64+
[AYA_8b.name]: OllamaModelMeta<OllamaChatRequest & OllamaChatRequestMessages>
65+
[AYA_35b.name]: OllamaModelMeta<OllamaChatRequest & OllamaChatRequestMessages>
5566
}
5667

5768
export type AyaModelInputModalitiesByName = {

packages/typescript/ai-ollama/src/meta/model-meta-codegemma.ts

Lines changed: 23 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,8 @@
1-
import type { ChatRequest } from 'ollama'
2-
import type { DefaultOllamaModelMeta } from './models-meta'
1+
import type {
2+
OllamaChatRequest,
3+
OllamaChatRequestMessages,
4+
OllamaModelMeta,
5+
} from './models-meta'
36

47
const CODEGEMMA_LATEST = {
58
name: 'codegemma:latest',
@@ -10,7 +13,9 @@ const CODEGEMMA_LATEST = {
1013
},
1114
size: '5gb',
1215
context: 8_000,
13-
} as const satisfies DefaultOllamaModelMeta<any>
16+
} as const satisfies OllamaModelMeta<
17+
OllamaChatRequest & OllamaChatRequestMessages
18+
>
1419

1520
const CODEGEMMA_8b = {
1621
name: 'codegemma:2b',
@@ -21,7 +26,9 @@ const CODEGEMMA_8b = {
2126
},
2227
size: '1.65gb',
2328
context: 8_000,
24-
} as const satisfies DefaultOllamaModelMeta<any>
29+
} as const satisfies OllamaModelMeta<
30+
OllamaChatRequest & OllamaChatRequestMessages
31+
>
2532

2633
const CODEGEMMA_35b = {
2734
name: 'codegemma:7b',
@@ -32,7 +39,9 @@ const CODEGEMMA_35b = {
3239
},
3340
size: '5gb',
3441
context: 8_000,
35-
} as const satisfies DefaultOllamaModelMeta<any>
42+
} as const satisfies OllamaModelMeta<
43+
OllamaChatRequest & OllamaChatRequestMessages
44+
>
3645

3746
export const CODEGEMMA_MODELS = [
3847
CODEGEMMA_LATEST.name,
@@ -53,9 +62,15 @@ export const CODEGEMMA_MODELS = [
5362
// Manual type map for per-model provider options
5463
export type CodegemmaChatModelProviderOptionsByName = {
5564
// Models with thinking and structured output support
56-
[CODEGEMMA_LATEST.name]: ChatRequest
57-
[CODEGEMMA_8b.name]: ChatRequest
58-
[CODEGEMMA_35b.name]: ChatRequest
65+
[CODEGEMMA_LATEST.name]: OllamaModelMeta<
66+
OllamaChatRequest & OllamaChatRequestMessages
67+
>
68+
[CODEGEMMA_8b.name]: OllamaModelMeta<
69+
OllamaChatRequest & OllamaChatRequestMessages
70+
>
71+
[CODEGEMMA_35b.name]: OllamaModelMeta<
72+
OllamaChatRequest & OllamaChatRequestMessages
73+
>
5974
}
6075

6176
export type CodegemmaModelInputModalitiesByName = {

packages/typescript/ai-ollama/src/meta/model-meta-codellama.ts

Lines changed: 35 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,8 @@
1-
import type { ChatRequest } from 'ollama'
2-
import type { DefaultOllamaModelMeta } from './models-meta'
1+
import type {
2+
OllamaChatRequest,
3+
OllamaChatRequestMessages,
4+
OllamaModelMeta,
5+
} from './models-meta'
36

47
const CODELLAMA_LATEST = {
58
name: 'codellama:latest',
@@ -10,7 +13,9 @@ const CODELLAMA_LATEST = {
1013
},
1114
size: '3.8gb',
1215
context: 16_000,
13-
} as const satisfies DefaultOllamaModelMeta<any>
16+
} as const satisfies OllamaModelMeta<
17+
OllamaChatRequest & OllamaChatRequestMessages
18+
>
1419

1520
const CODELLAMA_7b = {
1621
name: 'codellama:7b',
@@ -21,7 +26,9 @@ const CODELLAMA_7b = {
2126
},
2227
size: '3.8gb',
2328
context: 16_000,
24-
} as const satisfies DefaultOllamaModelMeta<any>
29+
} as const satisfies OllamaModelMeta<
30+
OllamaChatRequest & OllamaChatRequestMessages
31+
>
2532

2633
const CODELLAMA_13b = {
2734
name: 'codellama:13b',
@@ -32,7 +39,9 @@ const CODELLAMA_13b = {
3239
},
3340
size: '7.4gb',
3441
context: 16_000,
35-
} as const satisfies DefaultOllamaModelMeta<any>
42+
} as const satisfies OllamaModelMeta<
43+
OllamaChatRequest & OllamaChatRequestMessages
44+
>
3645

3746
const CODELLAMA_34b = {
3847
name: 'codellama:34b',
@@ -43,7 +52,9 @@ const CODELLAMA_34b = {
4352
},
4453
size: '19gb',
4554
context: 16_000,
46-
} as const satisfies DefaultOllamaModelMeta<any>
55+
} as const satisfies OllamaModelMeta<
56+
OllamaChatRequest & OllamaChatRequestMessages
57+
>
4758

4859
const CODELLAMA_70b = {
4960
name: 'codellama:70b',
@@ -54,7 +65,9 @@ const CODELLAMA_70b = {
5465
},
5566
size: '39gb',
5667
context: 2_000,
57-
} as const satisfies DefaultOllamaModelMeta<any>
68+
} as const satisfies OllamaModelMeta<
69+
OllamaChatRequest & OllamaChatRequestMessages
70+
>
5871

5972
export const CODELLAMA_MODELS = [
6073
CODELLAMA_LATEST.name,
@@ -77,11 +90,21 @@ export const CODELLAMA_MODELS = [
7790
// Manual type map for per-model provider options
7891
export type CodellamaChatModelProviderOptionsByName = {
7992
// Models with thinking and structured output support
80-
[CODELLAMA_LATEST.name]: ChatRequest
81-
[CODELLAMA_7b.name]: ChatRequest
82-
[CODELLAMA_13b.name]: ChatRequest
83-
[CODELLAMA_34b.name]: ChatRequest
84-
[CODELLAMA_70b.name]: ChatRequest
93+
[CODELLAMA_LATEST.name]: OllamaModelMeta<
94+
OllamaChatRequest & OllamaChatRequestMessages
95+
>
96+
[CODELLAMA_7b.name]: OllamaModelMeta<
97+
OllamaChatRequest & OllamaChatRequestMessages
98+
>
99+
[CODELLAMA_13b.name]: OllamaModelMeta<
100+
OllamaChatRequest & OllamaChatRequestMessages
101+
>
102+
[CODELLAMA_34b.name]: OllamaModelMeta<
103+
OllamaChatRequest & OllamaChatRequestMessages
104+
>
105+
[CODELLAMA_70b.name]: OllamaModelMeta<
106+
OllamaChatRequest & OllamaChatRequestMessages
107+
>
85108
}
86109

87110
export type CodellamaModelInputModalitiesByName = {

packages/typescript/ai-ollama/src/meta/model-meta-command-r-plus.ts

Lines changed: 27 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,10 @@
1-
import type { ChatRequest } from 'ollama'
2-
import type { DefaultOllamaModelMeta } from './models-meta'
1+
import type {
2+
OllamaChatRequest,
3+
OllamaChatRequestMessages,
4+
OllamaChatRequestTools,
5+
OllamaMessageTools,
6+
OllamaModelMeta,
7+
} from './models-meta'
38

49
const COMMAND_R_PLUS_LATEST = {
510
name: 'command-r-plus:latest',
@@ -10,7 +15,11 @@ const COMMAND_R_PLUS_LATEST = {
1015
},
1116
size: '59gb',
1217
context: 128_000,
13-
} as const satisfies DefaultOllamaModelMeta<any>
18+
} as const satisfies OllamaModelMeta<
19+
OllamaChatRequest &
20+
OllamaChatRequestMessages<OllamaMessageTools> &
21+
OllamaChatRequestTools
22+
>
1423

1524
const COMMAND_R_PLUS_104b = {
1625
name: 'command-r-plus:104b',
@@ -21,7 +30,11 @@ const COMMAND_R_PLUS_104b = {
2130
},
2231
size: '59gb',
2332
context: 128_000,
24-
} as const satisfies DefaultOllamaModelMeta<any>
33+
} as const satisfies OllamaModelMeta<
34+
OllamaChatRequest &
35+
OllamaChatRequestMessages<OllamaMessageTools> &
36+
OllamaChatRequestTools
37+
>
2538

2639
export const COMMAND_R_PLUS_MODELS = [
2740
COMMAND_R_PLUS_LATEST.name,
@@ -41,8 +54,16 @@ export const COMMAND_R_PLUS_MODELS = [
4154
// Manual type map for per-model provider options
4255
export type CommandRPlusChatModelProviderOptionsByName = {
4356
// Models with thinking and structured output support
44-
[COMMAND_R_PLUS_LATEST.name]: ChatRequest
45-
[COMMAND_R_PLUS_104b.name]: ChatRequest
57+
[COMMAND_R_PLUS_LATEST.name]: OllamaModelMeta<
58+
OllamaChatRequest &
59+
OllamaChatRequestMessages<OllamaMessageTools> &
60+
OllamaChatRequestTools
61+
>
62+
[COMMAND_R_PLUS_104b.name]: OllamaModelMeta<
63+
OllamaChatRequest &
64+
OllamaChatRequestMessages<OllamaMessageTools> &
65+
OllamaChatRequestTools
66+
>
4667
}
4768

4869
export type CommandRPlusModelInputModalitiesByName = {

packages/typescript/ai-ollama/src/meta/model-meta-command-r.ts

Lines changed: 27 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,10 @@
1-
import type { ChatRequest } from 'ollama'
2-
import type { DefaultOllamaModelMeta } from './models-meta'
1+
import type {
2+
OllamaChatRequest,
3+
OllamaChatRequestMessages,
4+
OllamaChatRequestTools,
5+
OllamaMessageTools,
6+
OllamaModelMeta,
7+
} from './models-meta'
38

49
const COMMAND_R_LATEST = {
510
name: 'command-r:latest',
@@ -10,7 +15,11 @@ const COMMAND_R_LATEST = {
1015
},
1116
size: '19gb',
1217
context: 128_000,
13-
} as const satisfies DefaultOllamaModelMeta<any>
18+
} as const satisfies OllamaModelMeta<
19+
OllamaChatRequest &
20+
OllamaChatRequestMessages<OllamaMessageTools> &
21+
OllamaChatRequestTools
22+
>
1423

1524
const COMMAND_R_35b = {
1625
name: 'command-r:35b',
@@ -21,7 +30,11 @@ const COMMAND_R_35b = {
2130
},
2231
size: '19gb',
2332
context: 128_000,
24-
} as const satisfies DefaultOllamaModelMeta<any>
33+
} as const satisfies OllamaModelMeta<
34+
OllamaChatRequest &
35+
OllamaChatRequestMessages<OllamaMessageTools> &
36+
OllamaChatRequestTools
37+
>
2538

2639
export const COMMAND_R_MODELS = [
2740
COMMAND_R_LATEST.name,
@@ -41,8 +54,16 @@ export const COMMAND_R_MODELS = [
4154
// Manual type map for per-model provider options
4255
export type CommandRChatModelProviderOptionsByName = {
4356
// Models with thinking and structured output support
44-
[COMMAND_R_LATEST.name]: ChatRequest
45-
[COMMAND_R_35b.name]: ChatRequest
57+
[COMMAND_R_LATEST.name]: OllamaModelMeta<
58+
OllamaChatRequest &
59+
OllamaChatRequestMessages<OllamaMessageTools> &
60+
OllamaChatRequestTools
61+
>
62+
[COMMAND_R_35b.name]: OllamaModelMeta<
63+
OllamaChatRequest &
64+
OllamaChatRequestMessages<OllamaMessageTools> &
65+
OllamaChatRequestTools
66+
>
4667
}
4768

4869
export type CommandRModelInputModalitiesByName = {

0 commit comments

Comments
 (0)