-
-
Notifications
You must be signed in to change notification settings - Fork 127
How to chat with Mesolitica chat
Please make sure login Github account to see the images.
We hosted Chat UI at https://chat.us.mesolitica.com/ to serve,
- https://huggingface.co/mesolitica/malaysian-mistral-7b-32k-instructions-v3.5
- https://huggingface.co/mesolitica/malaysian-tinyllama-1.1b-16k-instructions-v2
- https://huggingface.co/mesolitica/mallam-1.1b-20k-instructions-rag
- https://huggingface.co/mesolitica/mallam-5b-20k-instructions-v2
This order sorted by descending order in term of speed TPS (Tokens Per Second), https://huggingface.co/mesolitica/mallam-1.1b-20k-instructions-v2 is super fast.
In order to get the best response, please use Malaysian Mistral 7B v3.5 32k context length.
These models is a quick demonstration that the base model can be easily fine-tuned to achieve some performance, we are going to make it better in the future. It does have minimal moderation mechanisms, so please be patience to us.
We cannot tell how long we will open this Chat UI to the public due to GPU VM is super expensive, if you found out the chats crashed or no longer respond to you, probably the traffic is busy on that time, chillz ok.
Just press drop down button at Current Model
card, and it will list available models,
Personally I like Mistral the most, it can code, it can answer general QA related to the world and the most important is multiturn QA.
This will popup a new JSON page,
{
"note": "This is a preview of the prompt that will be sent to the model when retrying the message. It may differ from what was sent in the past if the parameters have been updated since",
"prompt": "<s>[INST] KWSP tu apa [/INST]",
"model": "Malaysian Mistral 7B 32k",
"parameters": {
"temperature": 0.3,
"truncate": 15360,
"max_new_tokens": 1024,
"stop": [
"</s>",
"[/INST]"
],
"top_p": 0.95,
"repetition_penalty": 1.15,
"top_k": 50,
"return_full_text": false
}
}
So you can try by yourself in any local.
If you are not satisfied with the output, you can always to regenerate,
If you want to make the models better, please help us on thumbs up or down,
Thumps up or down will be part of continuous human feedback loop learning process.
https://huggingface.co/spaces/mesolitica/malay-llm-leaderboard, currently we only focused on tatabahasa benchmark.
- Multi-turn Malaysian QA, it can answer follow up questions related to Malaysian context.
- Multi-turn General QA, it can answer follow up questions related to general or world context.
- Multi-turn Coding QA, it can answer follow up questions related to coding.
- Multi-turn Math and Science QA, it can answer mathematics and science follow up questions.
- Multi-turn RAG, you can give any document context and it can answer follow up questions based on the context.
You can ask about EPF.
You can ask about Palestine and Israel conflict.
All models should be able to do RAG including multiturn QA, you just provide context and it should able to answer, example,
I copied context from https://www.sinarharian.com.my/article/253799/berita/politik/isu-perkauman-kerajaan-didakwa-terdesak-kaitkan-pemimpin-pn and pasted into the chat and asked apa isu dengan konteks diatas
, and it replied,
Isu yang dibincangkan dalam artikel ini adalah tentang kenyataan berunsur perkauman yang dikaitkan dengan pemimpin Perikatan Nasional (PN). Artikel ini menyebut bahawa pemimpin PN, termasuk Presiden Pas dan Ketua Pemuda Pas Pulau Pinang, sedang disiasat atas kenyataan tersebut.
After that asked another follow up based on the printscreens.
Currently the UI does not able to upload the documents and we are not intended to add it for now.
This does not reflect our political view.
All these outputs from Malaysian Mistral 7B v3 32k context length
You ask to generate sajak buat sajak tentang nasi ayam
,
You ask about food at Penang,
Copy and pasted from https://iiumc.com/madu-atau-hempedu-ii-akhir/
**All these outputs from Malaysian Mistral 7B v3 32k context length
All these outputs from Malaysian Mistral 7B v3 32k context length
Always always verified the outputs.