{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"rwkv.cpp","owner":"RWKV","isFork":false,"description":"INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model","allTopics":["machine-learning","deep-learning","quantization","language-model","llm","rwkv","ggml"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":10,"issueCount":18,"starsCount":1133,"forksCount":82,"license":"MIT License","participation":[0,0,0,1,0,0,1,2,2,0,1,0,0,0,0,3,2,0,2,1,0,0,0,0,0,0,1,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,7,5,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-25T08:31:40.600Z"}},{"type":"Public","name":"llama.cpp","owner":"RWKV","isFork":true,"description":"LLM inference in C/C++","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":8956,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-26T10:52:49.934Z"}}],"repositoryCount":2,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"RWKV repositories"}