revert auto rope scaling for already-ropetuned models - just use their values
This commit is contained in:
parent
56a5fa7a60
commit
eb42c73953
3 changed files with 70 additions and 46 deletions
|
@ -941,19 +941,19 @@ ModelLoadResult gpttype_load_model(const load_model_inputs inputs, FileFormat in
|
|||
llamamodel->hparams.rope_freq_scale_train!=1.0f ||
|
||||
llamamodel->hparams.rope_scaling_type_train==2)
|
||||
{
|
||||
float ropemultiplier = 1.0f;
|
||||
if(llamamodel->hparams.rope_scaling_type_train!=2 &&
|
||||
llamamodel->hparams.n_ctx_train > 2048 && clamped_max_context_length > llamamodel->hparams.n_ctx_train)
|
||||
{
|
||||
ropemultiplier = (float)llamamodel->hparams.n_ctx_train / (float)clamped_max_context_length;
|
||||
llama_ctx_params.rope_freq_base = rope_freq_base = llamamodel->hparams.rope_freq_base_train;
|
||||
llama_ctx_params.rope_freq_scale = rope_freq_scale = ropemultiplier * llamamodel->hparams.rope_freq_scale_train;
|
||||
printf("Automatic RoPE Scaling: Using (scale:%.3f, base:%.1f).\n", rope_freq_scale, rope_freq_base);
|
||||
}
|
||||
else
|
||||
{
|
||||
// float ropemultiplier = 1.0f;
|
||||
// if(llamamodel->hparams.rope_scaling_type_train!=2 &&
|
||||
// llamamodel->hparams.n_ctx_train > 2048 && clamped_max_context_length > llamamodel->hparams.n_ctx_train)
|
||||
// {
|
||||
// ropemultiplier = (float)llamamodel->hparams.n_ctx_train / (float)clamped_max_context_length;
|
||||
// llama_ctx_params.rope_freq_base = rope_freq_base = llamamodel->hparams.rope_freq_base_train;
|
||||
// llama_ctx_params.rope_freq_scale = rope_freq_scale = ropemultiplier * llamamodel->hparams.rope_freq_scale_train;
|
||||
// printf("Automatic RoPE Scaling: Using (scale:%.3f, base:%.1f).\n", rope_freq_scale, rope_freq_base);
|
||||
// }
|
||||
// else
|
||||
// {
|
||||
printf("Automatic RoPE Scaling: Using model internal value.\n");
|
||||
}
|
||||
//}
|
||||
}
|
||||
else
|
||||
{
|
||||
|
|
90
klite.embd
90
klite.embd
|
@ -6,7 +6,7 @@ It requires no dependencies, installation or setup.
|
|||
Just copy this single static HTML file anywhere and open it in a browser, or from a webserver.
|
||||
Please go to https://github.com/LostRuins/lite.koboldai.net for updates on Kobold Lite.
|
||||
Kobold Lite is under the AGPL v3.0 License unless otherwise exempted. Please do not remove this line.
|
||||
Current version: 96
|
||||
Current version: 97
|
||||
-Concedo
|
||||
-->
|
||||
|
||||
|
@ -1654,18 +1654,9 @@ Current version: 96
|
|||
margin: 8px 0 8px;
|
||||
}
|
||||
|
||||
.cht_inp_hold input {
|
||||
|
||||
border: medium none;
|
||||
color: #bebebe;
|
||||
font-size: 15px;
|
||||
min-height: 36px;
|
||||
/* width: 100%; */
|
||||
outline:none;
|
||||
}
|
||||
|
||||
.cht_inp
|
||||
.cht_inp_bg
|
||||
{
|
||||
display: inline-block;
|
||||
width: calc(100% - 84px);
|
||||
background: #86868638 none repeat scroll 0 0;
|
||||
margin-top: 8px;
|
||||
|
@ -1673,8 +1664,23 @@ Current version: 96
|
|||
border-radius: 16px;
|
||||
padding-left: 10px;
|
||||
padding-right: 10px;
|
||||
padding-top: 7px;
|
||||
}
|
||||
.cht_inp.shorter
|
||||
.cht_inp_bg_inner
|
||||
{
|
||||
width: 100%;
|
||||
resize: none;
|
||||
overflow-y:scroll;
|
||||
overflow-x:hidden;
|
||||
background: #00000000 none repeat scroll 0 0;
|
||||
border: medium none;
|
||||
color: #bebebe;
|
||||
font-size: 15px;
|
||||
/* min-height: 36px; */
|
||||
/* width: 100%; */
|
||||
outline:none;
|
||||
}
|
||||
.cht_inp_bg.shorter
|
||||
{
|
||||
width: calc(100% - 114px);
|
||||
}
|
||||
|
@ -3238,7 +3244,7 @@ Current version: 96
|
|||
var image_db = {}; //stores a dictionary of pending images
|
||||
var completed_imgs_meta = {}; //stores temp info on completed images like alt text
|
||||
//key is ID, body is {done:false,queue:10,result:""}
|
||||
var stablemodels = []; //stored as {name,count}
|
||||
var stablemodels = [{"name": "stable_diffusion","count": 1}]; //stored as {name,count}
|
||||
var custom_kobold_endpoint = ""; //if set, does not use horde. Instead, attempts to use this sync endpoint
|
||||
var custom_oai_endpoint = "";
|
||||
var custom_oai_key = ""; //if set, uses the OpenAI API to generate
|
||||
|
@ -3599,11 +3605,18 @@ Current version: 96
|
|||
.setAttribute("content","width=device-width, initial-scale=1, maximum-scale=1");
|
||||
}
|
||||
|
||||
//fix for copy paste text in firefox
|
||||
//fix for copy paste text in firefox, and also to prevent pasting rich text
|
||||
{
|
||||
document.getElementById("gametext").addEventListener("paste", function(e) {
|
||||
e.preventDefault();
|
||||
var text = (e.originalEvent || e).clipboardData.getData('text/plain');
|
||||
|
||||
let text = e.clipboardData
|
||||
? (e.originalEvent || e).clipboardData.getData('text/plain')
|
||||
: // For IE
|
||||
window.clipboardData
|
||||
? window.clipboardData.getData('Text')
|
||||
: '';
|
||||
text = escapeHtml(text);
|
||||
text = text.replace(/\r?\n/g, '<br>');
|
||||
document.execCommand("insertHTML", false, text);
|
||||
});
|
||||
|
@ -7081,9 +7094,9 @@ Current version: 96
|
|||
{
|
||||
let sdmodelshtml = "";
|
||||
for (var i = 0; i < stablemodels.length; ++i) {
|
||||
sdmodelshtml += "<option value=\"" + stablemodels[i].name + " (" + stablemodels[i].count + ")\">";
|
||||
sdmodelshtml += "<option value=\"" + stablemodels[i].name + "\" "+(stablemodels[i].name==localsettings.generate_images_model?"selected":"")+">" + stablemodels[i].name + " (" + stablemodels[i].count + ")</option>";
|
||||
}
|
||||
document.getElementById("sdmodels").innerHTML = sdmodelshtml;
|
||||
document.getElementById("generate_images_model").innerHTML = sdmodelshtml;
|
||||
}
|
||||
|
||||
function toggle_preset() {
|
||||
|
@ -7118,13 +7131,10 @@ Current version: 96
|
|||
break;
|
||||
}
|
||||
}
|
||||
if (!matched && inputmodel != "*") {
|
||||
if (!matched) {
|
||||
document.getElementById("generate_images_model").value = defaultsettings.generate_images_model;
|
||||
}
|
||||
}
|
||||
function clear_sd_model() {
|
||||
document.getElementById("generate_images_model").value = "";
|
||||
}
|
||||
|
||||
function validate_samplers(savesetting = false) {
|
||||
let samplerstr = document.getElementById("sampler_order").value;
|
||||
|
@ -7665,6 +7675,7 @@ Current version: 96
|
|||
current_anotetemplate = "[Author's note: <|>]";
|
||||
document.getElementById("input_text").value = "";
|
||||
document.getElementById("cht_inp").value = "";
|
||||
chat_resize_input();
|
||||
image_db = {};
|
||||
completed_imgs_meta = {};
|
||||
localsettings.adventure_is_action = false;
|
||||
|
@ -7851,7 +7862,7 @@ Current version: 96
|
|||
let id_to_cancel = pending_response_id;
|
||||
|
||||
//flush any streaming text first
|
||||
if(is_using_custom_ep() && pending_response_id!="" && synchro_pending_stream!="")
|
||||
if(is_using_custom_ep() && pending_response_id != "" && (synchro_pending_stream != "" || synchro_polled_response != ""))
|
||||
{
|
||||
//apply a short delay of 1s before button reenables
|
||||
allow_reenable_submitbtn_timestamp = performance.now() + 500;
|
||||
|
@ -7859,7 +7870,10 @@ Current version: 96
|
|||
update_submit_button(true);
|
||||
}, 1000);
|
||||
|
||||
synchro_polled_response = synchro_pending_stream;
|
||||
if(synchro_pending_stream!="")
|
||||
{
|
||||
synchro_polled_response = synchro_pending_stream;
|
||||
}
|
||||
poll_in_progress = false;
|
||||
horde_poll_nearly_completed = false;
|
||||
poll_pending_response();
|
||||
|
@ -10219,10 +10233,10 @@ Current version: 96
|
|||
}
|
||||
if (localsettings.opmode == 3 && localsettings.chatopponent != "") {
|
||||
document.getElementById("chat_btnmode_chat").classList.remove("hidden");
|
||||
document.getElementById("cht_inp").classList.add("shorter");
|
||||
document.getElementById("cht_inp_bg").classList.add("shorter");
|
||||
} else {
|
||||
document.getElementById("chat_btnmode_chat").classList.add("hidden");
|
||||
document.getElementById("cht_inp").classList.remove("shorter");
|
||||
document.getElementById("cht_inp_bg").classList.remove("shorter");
|
||||
}
|
||||
|
||||
// Show the 'AI is typing' message if an answer is pending, and prevent the 'send button' from being clicked again.
|
||||
|
@ -10412,12 +10426,21 @@ Current version: 96
|
|||
}
|
||||
}
|
||||
}
|
||||
function chat_resize_input()
|
||||
{
|
||||
//resize chat inp
|
||||
let textarea = document.getElementById("cht_inp");
|
||||
let numberOfLineBreaks = (textarea.value.match(/\n/g) || []).length;
|
||||
numberOfLineBreaks = numberOfLineBreaks>4?4:numberOfLineBreaks;
|
||||
textarea.rows = numberOfLineBreaks+1;
|
||||
}
|
||||
function chat_submit_generation()
|
||||
{
|
||||
//easy solution is to just pump the text into the main box and submit it
|
||||
document.getElementById("input_text").value = document.getElementById("cht_inp").value;
|
||||
submit_generation();
|
||||
document.getElementById("cht_inp").value = "";
|
||||
chat_resize_input();
|
||||
}
|
||||
function chat_toggle_actionmenu()
|
||||
{
|
||||
|
@ -11430,15 +11453,16 @@ Current version: 96
|
|||
<div class="cht_inp_hold_outer">
|
||||
<div class="cht_inp_hold">
|
||||
<button onclick="show_groupchat_select()" id="chat_btnmode_chat" class="chat_btnmode_chat hidden" type="button"></button>
|
||||
<input id="cht_inp" type="text" name="chtchtinp" class="cht_inp" role="presentation" autocomplete="noppynop" spellcheck="true" placeholder="Type a message" value="" oninput="update_submit_button()" onkeypress="return chat_handle_typing(event)"/>
|
||||
|
||||
<div id="cht_inp_bg" class="cht_inp_bg">
|
||||
<textarea class="cht_inp_bg_inner" id="cht_inp" type="text" name="chtchtinp" role="presentation" autocomplete="noppynop" spellcheck="true" rows="1" wrap="off" placeholder="Type a message" value="" oninput="update_submit_button();chat_resize_input();" onkeypress="return chat_handle_typing(event)"/></textarea>
|
||||
</div>
|
||||
<button onclick="chat_submit_generation()" id="chat_msg_send_btn" class="chat_msg_send_btn" type="button"></button>
|
||||
<button onclick="abort_generation()" id="chat_msg_send_btn_abort" class="hidden chat_msg_send_btn_abort" type="button"></button>
|
||||
<button type="button" class="chat_msg_cust_btn" id="btn_chat_cust" onclick="chat_toggle_actionmenu()"></button>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="lastreq" id="lastreq2" style="padding-top: 4px; color:#999999"><span class="color_gray">Avoid sending privacy sensitive information. <a href="#" onclick="explain_horde()">Click here for more info</a>.</span></div>
|
||||
<div class="lastreq" id="lastreq2" style="padding-top: 2px; color:#999999"><span class="color_gray">Avoid sending privacy sensitive information. <a href="#" onclick="explain_horde()">Click here for more info</a>.</span></div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
|
@ -11701,6 +11725,7 @@ Current version: 96
|
|||
<option value="claude-instant-v1">claude-instant-v1</option>
|
||||
<option value="claude-instant-v1-100k">claude-instant-v1-100k</option>
|
||||
<option value="claude-2">claude-2</option>
|
||||
<option value="claude-2.1">claude-2.1</option>
|
||||
<option value="claude-2.0">claude-2.0</option>
|
||||
</select>
|
||||
<input type="checkbox" id="claudeaddversion" onchange="" checked>
|
||||
|
@ -12101,10 +12126,9 @@ Current version: 96
|
|||
<option value="2">Local A1111</option>
|
||||
<option value="3">OpenAI DALL-E</option>
|
||||
</select>
|
||||
<input list="sdmodels" class="form-control mdlpicker hidden" id="generate_images_model" style="font-size: 12px;height:20px;padding:2px;margin:0px 0 0;" placeholder="[Select Model]" value="" onblur="validate_sd_model()" onfocus="clear_sd_model()" title="Select a stable diffusion model to generate images with">
|
||||
<datalist id="sdmodels">
|
||||
<option value="stable_diffusion">
|
||||
</datalist>
|
||||
<select class="form-control" id="generate_images_model" style="font-size: 12px;height:20px;padding:2px;margin:0px 0 0;" onblur="validate_sd_model()" title="Select a stable diffusion model to generate images with">
|
||||
</select>
|
||||
|
||||
<div id="generate_images_local_model_container" class="settinglabel hidden">
|
||||
<select class="form-control" id="generate_images_local_model" style="height:20px;padding:0;margin:0px 0 0; width:calc(100% - 30px)">
|
||||
<option value="">[None]</option>
|
||||
|
|
|
@ -388,7 +388,7 @@ maxhordelen = 256
|
|||
modelbusy = threading.Lock()
|
||||
requestsinqueue = 0
|
||||
defaultport = 5001
|
||||
KcppVersion = "1.50.1"
|
||||
KcppVersion = "1.51"
|
||||
showdebug = True
|
||||
showsamplerwarning = True
|
||||
showmaxctxwarning = True
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue