Replies: 2 comments 3 replies
-
|
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Here we can show off the bugs we find.
Note
The context should now work up to your machines limit or the models limit whichever comes first (131072 = 128k)
Warning
In the first beta release, llama.cpp did not support the rope scaling required for full context (limit is 8192). Soon this will be updated for full 128K functionality.
Setting context higher than 8192 will cause bad output.
Please list the following in your post.
Beta Was this translation helpful? Give feedback.
All reactions