Skip to content

Commit 1a9f968

Browse files
Merge remote-tracking branch 'upstream/dev' into dev
2 parents 69abb81 + fec48b8 commit 1a9f968

24 files changed

Lines changed: 355 additions & 82 deletions

packages/core/script/generate-friendli.ts

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -48,7 +48,6 @@ const familyPatterns: [RegExp, string][] = [
4848
[/llama-4/i, "llama-4"],
4949
[/qwen3/i, "qwen3"],
5050
[/deepseek-r1/i, "deepseek-r1"],
51-
[/exaone/i, "exaone"],
5251
[/glm-4/i, "glm-4"],
5352
[/glm-5/i, "glm"],
5453
];

packages/core/script/generate-venice.ts

Lines changed: 3 additions & 30 deletions
Original file line numberDiff line numberDiff line change
@@ -3,30 +3,11 @@
33
import { z } from "zod";
44
import path from "node:path";
55
import { readdir } from "node:fs/promises";
6-
import * as readline from "node:readline";
76
import { ModelFamilyValues } from "../src/family.js";
87

98
// Venice API endpoint
109
const API_ENDPOINT = "https://api.venice.ai/api/v1/models?type=text";
1110

12-
async function promptForApiKey(): Promise<string | null> {
13-
const rl = readline.createInterface({
14-
input: process.stdin,
15-
output: process.stdout,
16-
});
17-
18-
return new Promise((resolve) => {
19-
rl.question(
20-
"Enter Venice API key to include alpha models (or press Enter to skip): ",
21-
(answer) => {
22-
rl.close();
23-
const trimmed = answer.trim();
24-
resolve(trimmed.length > 0 ? trimmed : null);
25-
},
26-
);
27-
});
28-
}
29-
3011
// Zod schemas for API response validation
3112
const Capabilities = z
3213
.object({
@@ -69,6 +50,7 @@ const ModelSpec = z
6950
.object({
7051
pricing: Pricing.optional(),
7152
availableContextTokens: z.number(),
53+
maxCompletionTokens: z.number().optional(),
7254
capabilities: Capabilities,
7355
constraints: z.any().optional(),
7456
name: z.string(),
@@ -257,11 +239,7 @@ function mergeModel(
257239
const caps = spec.capabilities;
258240

259241
const contextTokens = spec.availableContextTokens;
260-
const proposedOutputTokens = Math.floor(contextTokens / 4);
261-
const outputTokens =
262-
existing?.limit?.output !== undefined && existing.limit.output < proposedOutputTokens
263-
? existing.limit.output
264-
: proposedOutputTokens
242+
const outputTokens = spec.maxCompletionTokens ?? Math.floor(contextTokens / 4);
265243

266244
const openWeights = spec.modelSource
267245
? spec.modelSource.toLowerCase().includes("huggingface")
@@ -487,7 +465,7 @@ async function main() {
487465
"models",
488466
);
489467

490-
// Check for API key from CLI argument, environment, or prompt
468+
// Check for API key from CLI argument or environment variable
491469
let apiKey: string | null = null;
492470

493471
// Check CLI args for --api-key=xxx or --api-key xxx
@@ -506,11 +484,6 @@ async function main() {
506484
apiKey = process.env.VENICE_API_KEY ?? null;
507485
}
508486

509-
// Prompt if still no key
510-
if (!apiKey) {
511-
apiKey = await promptForApiKey();
512-
}
513-
514487
const includeAlpha = apiKey !== null;
515488

516489
if (dryRun) {

packages/core/src/family.ts

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -128,9 +128,6 @@ export const ModelFamilyValues = [
128128
"solar-mini",
129129
"solar-pro",
130130

131-
// Exaone
132-
"exaone",
133-
134131
// Step (StepFun)
135132
"step",
136133

Lines changed: 23 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,23 @@
1+
name = "GPT-5.3 Codex"
2+
family = "gpt-codex"
3+
release_date = "2026-02-24"
4+
last_updated = "2026-02-24"
5+
attachment = false
6+
reasoning = true
7+
temperature = false
8+
knowledge = "2025-08-31"
9+
tool_call = true
10+
open_weights = false
11+
12+
[cost]
13+
input = 1.75
14+
output = 14.00
15+
cache_read = 0.175
16+
17+
[limit]
18+
context = 400_000
19+
output = 128_000
20+
21+
[modalities]
22+
input = ["text", "image"]
23+
output = ["text"]
Lines changed: 23 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,23 @@
1+
name = "GPT-5.3 Codex"
2+
family = "gpt-codex"
3+
release_date = "2026-02-24"
4+
last_updated = "2026-02-24"
5+
attachment = false
6+
reasoning = true
7+
temperature = false
8+
knowledge = "2025-08-31"
9+
tool_call = true
10+
open_weights = false
11+
12+
[cost]
13+
input = 1.75
14+
output = 14.00
15+
cache_read = 0.175
16+
17+
[limit]
18+
context = 400_000
19+
output = 128_000
20+
21+
[modalities]
22+
input = ["text", "image"]
23+
output = ["text"]

providers/friendli/models/LGAI-EXAONE/EXAONE-4.0.1-32B.toml

Lines changed: 0 additions & 22 deletions
This file was deleted.

providers/friendli/models/LGAI-EXAONE/K-EXAONE-236B-A23B.toml

Lines changed: 0 additions & 22 deletions
This file was deleted.
Lines changed: 25 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,25 @@
1+
name = "MiniMax M2.5"
2+
family = "minimax"
3+
attachment = false
4+
reasoning = true
5+
tool_call = true
6+
structured_output = true
7+
temperature = true
8+
release_date = "2026-02-12"
9+
last_updated = "2026-02-12"
10+
open_weights = true
11+
12+
[interleaved]
13+
field = "reasoning_content"
14+
15+
[cost]
16+
input = 0.3
17+
output = 1.2
18+
19+
[limit]
20+
context = 196_608
21+
output = 196_608
22+
23+
[modalities]
24+
input = ["text"]
25+
output = ["text"]

providers/opencode-go/models/kimi-k2.5.toml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ field = "reasoning_content"
1515
[cost]
1616
input = 0.6
1717
output = 3.0
18-
cache_read = 0.08
18+
cache_read = 0.1
1919

2020
[limit]
2121
context = 262_144

providers/opencode-go/models/minimax-m2.5.toml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ open_weights = true
1212
[cost]
1313
input = 0.3
1414
output = 1.2
15-
cache_read = 0.06
15+
cache_read = 0.03
1616

1717
[limit]
1818
context = 204_800

0 commit comments

Comments
 (0)