Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Try a different LLM #10

Open
SakthiMahendran opened this issue May 8, 2024 · 1 comment
Open

Try a different LLM #10

SakthiMahendran opened this issue May 8, 2024 · 1 comment

Comments

@SakthiMahendran
Copy link

As people mentioned before LLaMa is very slow due it's huge size... So why don't we try different models for example
"phi3" from microsoft.... in the context of day to day uses llama and phi3 won't have big difference.... but phi3 is lightwieght and is much more quick than llama. It has 3.8B parameters.... If we wan't to make to useable then LLaMa won't be good choice...

try "ollama run phi3"... You will get an idea

@FotieMConstant
Copy link
Member

Hey there @SakthiMahendran thanks for the input on this. I really appreciate you suggestion. any thoughts on fine-tuning phi3? perhaps we can start from there. I have never fine-tuned it before and I suppose the fine-tuning processing is different from llama.

let me know. Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants