sy / public /scripts /chat-templates.js
吴松泽
main
c120a1c
import { t } from './i18n.js';
// the hash can be obtained from command line e.g. via: MODEL=path_to_model; python -c "import json, hashlib, sys; print(hashlib.sha256(json.load(open('"$MODEL"/tokenizer_config.json'))['chat_template'].encode()).hexdigest())"
// note that chat templates must be trimmed to match the llama.cpp metadata value
const hash_derivations = {
// Meta
'e10ca381b1ccc5cf9db52e371f3b6651576caee0a630b452e2816b2d404d4b65':
// Meta-Llama-3.1-8B-Instruct
// Meta-Llama-3.1-70B-Instruct
'Llama 3 Instruct'
,
'5816fce10444e03c2e9ee1ef8a4a1ea61ae7e69e438613f3b17b69d0426223a4':
// Llama-3.2-1B-Instruct
// Llama-3.2-3B-Instruct
'Llama 3 Instruct'
,
'73e87b1667d87ab7d7b579107f01151b29ce7f3ccdd1018fdc397e78be76219d':
// Nemotron 70B
'Llama 3 Instruct'
,
// Mistral
// Mistral Reference: https://github.com/mistralai/mistral-common
'e16746b40344d6c5b5265988e0328a0bf7277be86f1c335156eae07e29c82826':
// Mistral-Small-Instruct-2409
// Mistral-Large-Instruct-2407
'Mistral V2 & V3'
,
'26a59556925c987317ce5291811ba3b7f32ec4c647c400c6cc7e3a9993007ba7':
// Mistral-7B-Instruct-v0.3
'Mistral V2 & V3'
,
'e4676cb56dffea7782fd3e2b577cfaf1e123537e6ef49b3ec7caa6c095c62272':
// Mistral-Nemo-Instruct-2407
'Mistral V3-Tekken'
,
'3c4ad5fa60dd8c7ccdf82fa4225864c903e107728fcaf859fa6052cb80c92ee9':
// Mistral-Large-Instruct-2411
'Mistral V7'
,
'3934d199bfe5b6fab5cba1b5f8ee475e8d5738ac315f21cb09545b4e665cc005':
// Mistral Small 24B
'Mistral V7'
,
// Gemma
'ecd6ae513fe103f0eb62e8ab5bfa8d0fe45c1074fa398b089c93a7e70c15cfd6':
// gemma-2-9b-it
// gemma-2-27b-it
'Gemma 2'
,
'87fa45af6cdc3d6a9e4dd34a0a6848eceaa73a35dcfe976bd2946a5822a38bf3':
// gemma-2-2b-it
'Gemma 2'
,
'7de1c58e208eda46e9c7f86397df37ec49883aeece39fb961e0a6b24088dd3c4':
// gemma-3
'Gemma 2'
,
// Cohere
'3b54f5c219ae1caa5c0bb2cdc7c001863ca6807cf888e4240e8739fa7eb9e02e':
// command-r-08-2024
'Command R'
,
// Tulu
'ac7498a36a719da630e99d48e6ebc4409de85a77556c2b6159eeb735bcbd11df':
// Tulu-3-8B
// Tulu-3-70B
'Tulu'
,
// DeepSeek V2.5
'54d400beedcd17f464e10063e0577f6f798fa896266a912d8a366f8a2fcc0bca':
'DeepSeek-V2.5'
,
// DeepSeek R1
'b6835114b7303ddd78919a82e4d9f7d8c26ed0d7dfc36beeb12d524f6144eab1':
'DeepSeek-V2.5'
,
// THUDM-GLM 4
'854b703e44ca06bdb196cc471c728d15dbab61e744fe6cdce980086b61646ed1':
'GLM-4'
,
// Kimi K2, ...
'aab20feb9bc6881f941ea649356130ffbc4943b3c2577c0991e1fba90de5a0fc':
'Moonshot AI'
,
// gpt-oss (unsloth)
'70da0d2348e40aaf8dad05f04a316835fd10547bd7e3392ce337e4c79ba91c01':
'OpenAI Harmony'
,
// gpt-oss (ggml-org)
'a4c9919cbbd4acdd51ccffe22da049264b1b73e59055fa58811a99efbd7c8146':
'OpenAI Harmony'
,
};
const substr_derivations = [
['Moonshot AI', ['<|im_user|>user<|im_middle|>', '<|im_assistant|>assistant<|im_middle|>', '<|im_end|>']],
['OpenAI Harmony', ['<|start|>user<|message|>', '<|start|>assistant<|channel|>final<|message|>', '<|end|>']],
// Generic cases
['ChatML', ['<|im_start|>user', '<|im_start|>assistant', '<|im_end|>']],
];
const parse_derivation = derivation => (typeof derivation === 'string') ? {
'context': derivation,
'instruct': derivation,
} : derivation;
const not_found = { context: null, instruct: null };
export async function deriveTemplatesFromChatTemplate(chat_template, hash) {
if (chat_template.trim() === '') {
console.log('Missing chat template.');
return not_found;
}
if (hash in hash_derivations) {
return parse_derivation(hash_derivations[hash]);
}
// heuristics
for (const [derivation, substr] of substr_derivations) {
if ([substr].flat().every(str => chat_template.includes(str))) {
return parse_derivation(derivation);
}
}
console.warn(`Unknown chat template hash: ${hash} for [${chat_template}]`);
return not_found;
}
export async function bindModelTemplates(power_user, online_status) {
if (online_status === 'no_connection') {
return false;
}
const chatTemplateHash = power_user.chat_template_hash;
const bindModelTemplates = power_user.model_templates_mappings[online_status]
?? power_user.model_templates_mappings[chatTemplateHash]
?? {};
const bindingsMatch = bindModelTemplates
&& power_user.context.preset == bindModelTemplates['context']
&& (!power_user.instruct.enabled || power_user.instruct.preset === bindModelTemplates['instruct']);
const bound = [];
if (bindingsMatch) {
// unmap current preset
delete power_user.model_templates_mappings[chatTemplateHash];
delete power_user.model_templates_mappings[online_status];
toastr.info(t`Context preset for ${online_status} will use defaults when loaded the next time.`);
} else {
if (power_user.context_derived) {
if (power_user.context.preset !== bindModelTemplates['context']) {
bound.push(`${power_user.context.preset} context preset`);
// toastr.info(`Bound ${power_user.context.preset} preset to currently loaded model and all models that share its chat template.`);
// map current preset to current chat template hash
bindModelTemplates['context'] = power_user.context.preset;
}
} else {
toastr.warning(t`Note: Context derivation is disabled. Not including context preset.`);
}
if (power_user.instruct.enabled) {
if (power_user.instruct_derived) {
if (power_user.instruct.preset !== bindModelTemplates['instruct']) {
bound.push(`${power_user.instruct.preset} instruct preset`);
bindModelTemplates['instruct'] = power_user.instruct.preset;
}
} else {
toastr.warning(t`Note: Instruct derivation is disabled. Not including instruct preset.`);
}
}
if (bound.length == 0) {
toastr.warning(t`No applicable presets available.`);
return false;
}
toastr.info(t`Bound ${online_status} to ${bound.join(', ')}.`);
if (!online_status.startsWith('koboldcpp/ggml-model-')) {
power_user.model_templates_mappings[online_status] = bindModelTemplates;
}
if (chatTemplateHash !== '') {
power_user.model_templates_mappings[chatTemplateHash] = bindModelTemplates;
}
}
return true;
}