diff --git a/examples/server/themes/mount-ai/index.html b/examples/server/themes/mount-ai/index.html
index afefd625e..251bdab17 100644
--- a/examples/server/themes/mount-ai/index.html
+++ b/examples/server/themes/mount-ai/index.html
@@ -257,7 +257,7 @@
console.log('already running...');
return;
}
- // just in case
+ // just in case (e.g. llama2)
const suffix = session.value.userMsgSuffix || "";
const prefix = session.value.userMsgPrefix || "";
const userMsg = prefix + msg + suffix;
@@ -285,7 +285,7 @@
await runLlama(prompt, {
...params.value,
slot_id: slot_id,
- stop: ["", "<|end|>", "<|eot_id|>", "<|end_of_text|>", "<|im_end|>", "<|EOT|>", "<|END_OF_TURN_TOKEN|>", "<|endoftext|>", template("{{char}}"), template("{{user}}")],
+ stop: ["", "<|end|>", "<|eot_id|>", "<|end_of_text|>", "<|im_end|>", "<|EOT|>", "<|END_OF_TURN_TOKEN|>", "<|end_of_turn|>", "<|endoftext|>", template("{{char}}"), template("{{user}}")],
}, "{{char}}");
}
@@ -642,6 +642,7 @@ const BoolField = ({ label, title, name, value }) => {
id="prompt"
class="persistent-input"
name="prompt"
+ placeholder="The following models do not support System Prompt by design: OpenChat, Orion, Phi-3, Starling"
value="${session.value.prompt}"
oninput=${updateSession}
>
@@ -663,16 +664,17 @@ const BoolField = ({ label, title, name, value }) => {
+