Files
Moltbot/src/agents/usage.test.ts
Elarwei aa2826b5b1 fix(usage): parse Kimi K2 cached_tokens from prompt_tokens_details
Kimi K2 models use automatic prefix caching and return cache stats in
a nested field: usage.prompt_tokens_details.cached_tokens

This fixes issue #7073 where cacheRead was showing 0 for K2.5 users.

Also adds cached_tokens (top-level) for moonshot-v1 explicit caching API.

Closes #7073
2026-02-24 14:40:52 +00:00

179 lines
4.5 KiB
TypeScript

import { describe, expect, it } from "vitest";
import {
normalizeUsage,
hasNonzeroUsage,
derivePromptTokens,
deriveSessionTotalTokens,
} from "./usage.js";
describe("normalizeUsage", () => {
it("normalizes cache fields from provider response", () => {
const usage = normalizeUsage({
input: 1000,
output: 500,
cacheRead: 2000,
cacheWrite: 300,
});
expect(usage).toEqual({
input: 1000,
output: 500,
cacheRead: 2000,
cacheWrite: 300,
total: undefined,
});
});
it("normalizes cache fields from alternate naming", () => {
const usage = normalizeUsage({
input_tokens: 1000,
output_tokens: 500,
cache_read_input_tokens: 2000,
cache_creation_input_tokens: 300,
});
expect(usage).toEqual({
input: 1000,
output: 500,
cacheRead: 2000,
cacheWrite: 300,
total: undefined,
});
});
it("handles cache_read and cache_write naming variants", () => {
const usage = normalizeUsage({
input: 1000,
cache_read: 1500,
cache_write: 200,
});
expect(usage).toEqual({
input: 1000,
output: undefined,
cacheRead: 1500,
cacheWrite: 200,
total: undefined,
});
});
it("handles Moonshot/Kimi cached_tokens field", () => {
// Moonshot v1 returns cached_tokens instead of cache_read_input_tokens
const usage = normalizeUsage({
prompt_tokens: 30,
completion_tokens: 9,
total_tokens: 39,
cached_tokens: 19,
});
expect(usage).toEqual({
input: 30,
output: 9,
cacheRead: 19,
cacheWrite: undefined,
total: 39,
});
});
it("handles Kimi K2 prompt_tokens_details.cached_tokens field", () => {
// Kimi K2 uses automatic prefix caching and returns cached_tokens in prompt_tokens_details
const usage = normalizeUsage({
prompt_tokens: 1113,
completion_tokens: 5,
total_tokens: 1118,
prompt_tokens_details: { cached_tokens: 1024 },
});
expect(usage).toEqual({
input: 1113,
output: 5,
cacheRead: 1024,
cacheWrite: undefined,
total: 1118,
});
});
it("returns undefined when no valid fields are provided", () => {
const usage = normalizeUsage(null);
expect(usage).toBeUndefined();
});
it("handles undefined input", () => {
const usage = normalizeUsage(undefined);
expect(usage).toBeUndefined();
});
});
describe("hasNonzeroUsage", () => {
it("returns true when cache read is nonzero", () => {
const usage = { cacheRead: 100 };
expect(hasNonzeroUsage(usage)).toBe(true);
});
it("returns true when cache write is nonzero", () => {
const usage = { cacheWrite: 50 };
expect(hasNonzeroUsage(usage)).toBe(true);
});
it("returns true when both cache fields are nonzero", () => {
const usage = { cacheRead: 100, cacheWrite: 50 };
expect(hasNonzeroUsage(usage)).toBe(true);
});
it("returns false when cache fields are zero", () => {
const usage = { cacheRead: 0, cacheWrite: 0 };
expect(hasNonzeroUsage(usage)).toBe(false);
});
it("returns false for undefined usage", () => {
expect(hasNonzeroUsage(undefined)).toBe(false);
});
});
describe("derivePromptTokens", () => {
it("includes cache tokens in prompt total", () => {
const usage = {
input: 1000,
cacheRead: 500,
cacheWrite: 200,
};
const promptTokens = derivePromptTokens(usage);
expect(promptTokens).toBe(1700); // 1000 + 500 + 200
});
it("handles missing cache fields", () => {
const usage = {
input: 1000,
};
const promptTokens = derivePromptTokens(usage);
expect(promptTokens).toBe(1000);
});
it("returns undefined for empty usage", () => {
const promptTokens = derivePromptTokens({});
expect(promptTokens).toBeUndefined();
});
});
describe("deriveSessionTotalTokens", () => {
it("includes cache tokens in total calculation", () => {
const totalTokens = deriveSessionTotalTokens({
usage: {
input: 1000,
cacheRead: 500,
cacheWrite: 200,
},
contextTokens: 4000,
});
expect(totalTokens).toBe(1700); // 1000 + 500 + 200
});
it("prefers promptTokens override over derived total", () => {
const totalTokens = deriveSessionTotalTokens({
usage: {
input: 1000,
cacheRead: 500,
cacheWrite: 200,
},
contextTokens: 4000,
promptTokens: 2500, // Override
});
expect(totalTokens).toBe(2500);
});
});