-
Notifications
You must be signed in to change notification settings - Fork 2.3k
feat: update kimi-k2-0905-preview and kimi-k2-turbo-preview #7663
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from all commits
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -3,7 +3,7 @@ import type { ModelInfo } from "../model.js" | |
// https://platform.moonshot.ai/ | ||
export type MoonshotModelId = keyof typeof moonshotModels | ||
|
||
export const moonshotDefaultModelId: MoonshotModelId = "kimi-k2-0711-preview" | ||
export const moonshotDefaultModelId: MoonshotModelId = "kimi-k2-0905-preview" | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Is the new kimi-k2-0905-preview model already available in production? Changing the default model could break functionality for users if this model isn't accessible yet. Have you verified its availability? There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. |
||
|
||
export const moonshotModels = { | ||
"kimi-k2-0711-preview": { | ||
|
@@ -17,9 +17,20 @@ export const moonshotModels = { | |
cacheReadsPrice: 0.15, // $0.15 per million tokens (cache hit) | ||
description: `Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters.`, | ||
}, | ||
"kimi-k2-0905-preview": { | ||
maxTokens: 32_000, | ||
contextWindow: 262_144, | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. The context window has been doubled from 131,072 to 262,144 tokens. Has this been tested with large contexts? This significant increase could impact:
|
||
supportsImages: false, | ||
supportsPromptCache: true, | ||
inputPrice: 0.6, // $0.60 per million tokens (cache miss) | ||
outputPrice: 2.5, // $2.50 per million tokens | ||
cacheWritesPrice: 0, // $0 per million tokens (cache miss) | ||
cacheReadsPrice: 0.15, // $0.15 per million tokens (cache hit) | ||
description: `Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters.`, | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. The description is identical to the older kimi-k2-0711-preview model. If this is a newer version with improvements, could we update the description to reflect what's new or different about the 0905 version? |
||
}, | ||
"kimi-k2-turbo-preview": { | ||
maxTokens: 32_000, | ||
contextWindow: 131_072, | ||
contextWindow: 262_144, | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Similar concern here - the context window for kimi-k2-turbo-preview has been doubled. Has this change been tested for performance impact, especially given this is the 'turbo' variant optimized for speed? |
||
supportsImages: false, | ||
supportsPromptCache: true, | ||
inputPrice: 2.4, // $2.40 per million tokens (cache miss) | ||
|
Uh oh!
There was an error while loading. Please reload this page.