{"payload":{"pageCount":8,"repositories":[{"type":"Public","name":"safetensors","owner":"huggingface","isFork":false,"description":"Simple, safe way to store and distribute tensors","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":11,"starsCount":2598,"forksCount":163,"license":"Apache License 2.0","participation":[1,0,0,3,10,5,11,1,3,0,3,0,1,0,0,0,0,0,5,2,3,1,0,0,0,3,3,6,2,0,0,1,0,0,0,0,0,0,0,1,2,1,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-06T01:55:16.113Z"}},{"type":"Public","name":"diffusers","owner":"huggingface","isFork":false,"description":"🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch and FLAX.","allTopics":["deep-learning","pytorch","image-generation","flax","hacktoberfest","diffusion","text2image","image2image","jax","score-based-generative-modeling","stable-diffusion","stable-diffusion-diffusers","latent-diffusion-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":136,"issueCount":363,"starsCount":23938,"forksCount":4927,"license":"Apache License 2.0","participation":[41,32,50,43,33,27,44,45,25,50,34,48,31,26,26,39,37,39,46,25,51,38,15,50,41,43,31,38,27,28,44,33,30,45,42,37,35,31,26,23,19,22,35,21,26,21,30,27,26,33,39,33],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-06T01:45:32.572Z"}},{"type":"Public","name":"transformers","owner":"huggingface","isFork":false,"description":"🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.","allTopics":["python","seq2seq","flax","language-models","nlp-library","hacktoberfest","jax","pytorch-transformers","model-hub","nlp","machine-learning","natural-language-processing","deep-learning","tensorflow","pytorch","transformer","speech-recognition","pretrained-models","language-model","bert"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":261,"issueCount":906,"starsCount":128632,"forksCount":25514,"license":"Apache License 2.0","participation":[48,71,79,48,67,65,71,63,61,51,48,41,68,57,52,64,82,61,68,46,38,71,77,51,3,20,52,55,47,57,41,52,48,55,65,67,55,45,50,61,73,79,63,47,61,73,55,74,41,48,47,33],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-06T01:01:36.843Z"}},{"type":"Public","name":"dataset-viewer","owner":"huggingface","isFork":false,"description":"Lightweight web API for visualizing and exploring any dataset - computer vision, speech, text, and tabular - stored on the Hugging Face Hub","allTopics":["nlp","data","machine-learning","api-rest","datasets","huggingface"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":9,"issueCount":134,"starsCount":640,"forksCount":65,"license":"Apache License 2.0","participation":[12,13,32,41,26,25,10,12,32,13,17,38,33,27,20,13,19,25,15,14,7,15,15,7,0,7,22,15,9,16,20,17,28,17,17,24,11,13,9,26,17,10,8,10,29,17,15,12,13,17,19,4],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-06T00:37:39.373Z"}},{"type":"Public","name":"text-generation-inference","owner":"huggingface","isFork":false,"description":"Large Language Model Text Generation Inference","allTopics":["nlp","bloom","deep-learning","inference","pytorch","falcon","transformer","gpt","starcoder"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":22,"issueCount":125,"starsCount":8342,"forksCount":943,"license":"Apache License 2.0","participation":[15,19,13,13,13,15,3,8,13,3,2,20,14,3,5,4,0,4,2,7,6,2,18,7,0,0,6,2,20,13,7,12,8,13,0,3,13,4,5,18,7,18,20,2,21,14,14,25,13,11,24,30],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T22:07:34.588Z"}},{"type":"Public","name":"chat-ui","owner":"huggingface","isFork":false,"description":"Open source codebase powering the HuggingChat app","allTopics":["typescript","svelte","hacktoberfest","tailwindcss","huggingface","svelte-kit","sveltekit","llm","chatgpt"],"primaryLanguage":{"name":"TypeScript","color":"#3178c6"},"pullRequestCount":24,"issueCount":213,"starsCount":6766,"forksCount":947,"license":"Apache License 2.0","participation":[3,10,4,5,0,3,13,0,0,9,16,6,8,18,13,10,1,0,7,12,6,13,8,5,10,1,9,7,21,16,24,20,16,8,18,10,10,8,11,20,14,33,14,13,6,10,51,12,21,5,8,8],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T22:02:26.748Z"}},{"type":"Public","name":"lighteval","owner":"huggingface","isFork":false,"description":"LightEval is a lightweight LLM evaluation suite that Hugging Face has been using internally with the recently released LLM data processing library datatrove and LLM training library nanotron.","allTopics":["evaluation","evaluation-metrics","evaluation-framework","huggingface"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":13,"issueCount":48,"starsCount":471,"forksCount":55,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T20:00:36.874Z"}},{"type":"Public","name":"lerobot","owner":"huggingface","isFork":false,"description":"🤗 LeRobot: End-to-end Learning for Real-World Robotics in Pytorch","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":26,"issueCount":20,"starsCount":4348,"forksCount":348,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,5,0,4,10,34,47,72,55,66,41,73,56,6,22,22,12,10,15,7,8,8,2,4],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T18:18:29.282Z"}},{"type":"Public","name":"text-embeddings-inference","owner":"huggingface","isFork":false,"description":"A blazing fast inference solution for text embeddings models","allTopics":["ai","ml","embeddings","huggingface","llm"],"primaryLanguage":{"name":"Rust","color":"#dea584"},"pullRequestCount":5,"issueCount":55,"starsCount":2268,"forksCount":136,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,8,27,5,2,3,7,6,6,1,0,1,0,4,0,1,3,0,0,2,7,5,6,1,8,0,4,3,6,2,0,0,1,0,1,1,0,12,7,6],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T18:03:25.548Z"}},{"type":"Public","name":"ratchet","owner":"huggingface","isFork":false,"description":"A cross-platform browser ML framework.","allTopics":[],"primaryLanguage":{"name":"Rust","color":"#dea584"},"pullRequestCount":5,"issueCount":25,"starsCount":474,"forksCount":26,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,6,53,86,94,48,54,64,51,40,36,56,30,35,61,53,44,16,6,6,40,34,33,16,62,14,55],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T17:00:35.342Z"}},{"type":"Public","name":"blog","owner":"huggingface","isFork":false,"description":"Public repo for HF blog posts","allTopics":["hacktoberfest"],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":69,"issueCount":135,"starsCount":2138,"forksCount":656,"license":null,"participation":[6,31,19,11,13,3,32,21,19,33,16,21,11,5,14,15,9,18,3,3,6,21,11,8,4,8,4,14,11,22,9,8,29,26,5,13,24,6,18,17,29,14,8,12,17,21,16,10,12,7,18,12],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T16:42:46.409Z"}},{"type":"Public","name":"cookbook","owner":"huggingface","isFork":false,"description":"Open-source AI cookbook","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":18,"issueCount":16,"starsCount":1475,"forksCount":198,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T16:32:17.965Z"}},{"type":"Public","name":"accelerate","owner":"huggingface","isFork":false,"description":"🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":15,"issueCount":113,"starsCount":7343,"forksCount":875,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T16:01:06.450Z"}},{"type":"Public","name":"optimum-tpu","owner":"huggingface","isFork":false,"description":"Google TPU optimizations for transformers models","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":5,"starsCount":40,"forksCount":8,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T15:48:45.227Z"}},{"type":"Public","name":"datatrove","owner":"huggingface","isFork":false,"description":"Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":31,"starsCount":1756,"forksCount":105,"license":"Apache License 2.0","participation":[12,5,11,8,9,0,3,4,4,2,0,0,3,1,17,2,11,3,15,30,7,10,3,11,1,0,1,8,5,9,12,0,13,6,8,1,7,2,3,3,5,11,2,10,4,10,4,3,5,6,3,4],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T15:40:49.739Z"}},{"type":"Public","name":"optimum-quanto","owner":"huggingface","isFork":false,"description":"A pytorch quantization backend for optimum","allTopics":["pytorch","quantization","optimum"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":15,"starsCount":651,"forksCount":35,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,4,26,13,16,37,0,0,4,7,19,56,38,26,0,10,2,9,7,9,10,5,41,10,14,25,4,11,6,12,13,1,4,3,9,7,13,5,6,0,15,6],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T15:38:53.962Z"}},{"type":"Public","name":"optimum-neuron","owner":"huggingface","isFork":false,"description":"Easy, fast and very cheap training and inference on AWS Trainium and Inferentia chips.","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":15,"issueCount":83,"starsCount":176,"forksCount":53,"license":"Apache License 2.0","participation":[4,11,2,4,5,4,13,5,9,8,7,2,4,5,2,8,3,8,12,14,1,1,4,7,0,0,3,9,10,5,7,7,5,3,4,1,4,13,8,11,4,5,3,3,6,3,8,5,7,2,3,6],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T15:14:16.438Z"}},{"type":"Public","name":"leaderboards","owner":"huggingface","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":9,"forksCount":3,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T15:08:32.406Z"}},{"type":"Public","name":"peft","owner":"huggingface","isFork":false,"description":"🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.","allTopics":["python","adapter","transformers","pytorch","lora","diffusion","parameter-efficient-learning","llm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":13,"issueCount":19,"starsCount":14925,"forksCount":1419,"license":"Apache License 2.0","participation":[15,13,8,4,7,7,9,9,7,7,7,10,9,9,3,6,9,17,17,10,11,24,16,10,3,4,20,3,4,17,16,12,19,6,9,11,8,5,5,11,5,8,11,5,12,7,7,8,13,7,6,9],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T15:06:37.064Z"}},{"type":"Public","name":"trl","owner":"huggingface","isFork":false,"description":"Train transformer language models with reinforcement learning.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":19,"issueCount":54,"starsCount":8707,"forksCount":1067,"license":"Apache License 2.0","participation":[7,8,14,6,6,7,10,14,8,21,10,5,7,10,8,7,14,13,5,6,8,9,9,15,10,4,23,12,8,16,1,8,5,11,4,7,19,8,7,16,12,9,11,3,3,11,3,16,3,16,9,6],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T14:38:04.448Z"}},{"type":"Public","name":"optimum-intel","owner":"huggingface","isFork":false,"description":"🤗 Optimum Intel: Accelerate inference with Intel optimization tools","allTopics":["optimization","intel","transformers","pruning","distillation","onnx","openvino","diffusers","inference","quantization"],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":26,"issueCount":9,"starsCount":355,"forksCount":99,"license":"Apache License 2.0","participation":[2,6,15,18,2,2,6,1,6,2,5,6,3,4,3,2,4,10,0,0,3,3,6,1,5,4,8,11,11,13,9,9,4,5,12,14,16,15,2,12,11,25,15,10,16,5,10,17,8,10,5,12],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T14:37:45.277Z"}},{"type":"Public","name":"hub-docs","owner":"huggingface","isFork":false,"description":"Docs of the Hugging Face Hub","allTopics":["machine-learning","hacktoberfest"],"primaryLanguage":null,"pullRequestCount":19,"issueCount":71,"starsCount":251,"forksCount":214,"license":"Apache License 2.0","participation":[8,4,3,4,4,2,4,8,8,9,9,13,19,9,15,10,23,14,11,17,21,5,5,2,0,3,7,4,7,6,7,3,2,5,3,12,4,3,4,9,1,1,3,1,5,3,3,5,1,2,6,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T14:32:23.898Z"}},{"type":"Public","name":"huggingface.js","owner":"huggingface","isFork":false,"description":"Utilities to use the Hugging Face Hub API","allTopics":["machine-learning","inference","hub","api-client","huggingface"],"primaryLanguage":{"name":"TypeScript","color":"#3178c6"},"pullRequestCount":24,"issueCount":90,"starsCount":1281,"forksCount":160,"license":"MIT License","participation":[0,8,6,0,0,0,1,1,0,0,0,0,9,12,0,11,6,11,6,27,25,17,10,21,0,1,2,12,7,2,8,7,20,15,23,23,13,16,13,25,11,6,18,13,24,23,10,15,9,22,11,5],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T13:30:23.752Z"}},{"type":"Public","name":"llm-swarm","owner":"huggingface","isFork":false,"description":"Manage scalable open LLM inference endpoints in Slurm clusters","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":0,"starsCount":193,"forksCount":16,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,13,5,11,11,0,0,2,0,0,7,5,34,16,27,6,0,1,1,2,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T12:56:04.917Z"}},{"type":"Public","name":"hf-endpoints-documentation","owner":"huggingface","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":2,"issueCount":1,"starsCount":9,"forksCount":7,"license":null,"participation":[0,0,2,1,2,0,3,0,0,0,0,0,0,0,1,0,0,0,0,0,4,8,3,0,0,0,0,0,0,0,0,0,0,2,1,0,0,12,2,23,0,0,0,8,3,4,4,4,4,2,15,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T12:54:45.840Z"}},{"type":"Public","name":"nanotron","owner":"huggingface","isFork":false,"description":"Minimalistic large language model 3D-parallelism training","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":30,"issueCount":35,"starsCount":946,"forksCount":86,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,2,0,2,1,0,0,0,0,0,0,0,0,0,0,31,1,11,33,96,64,79,39,131,35,27,61,27,24,14,37,37,59,42,7,17,11,6,5,0,3,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T12:53:04.654Z"}},{"type":"Public","name":"optimum","owner":"huggingface","isFork":false,"description":"🚀 Accelerate training and inference of 🤗 Transformers and 🤗 Diffusers with easy to use hardware optimization tools","allTopics":["training","optimization","intel","transformers","tflite","onnxruntime","graphcore","habana","inference","pytorch","quantization","onnx"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":93,"issueCount":291,"starsCount":2306,"forksCount":403,"license":"Apache License 2.0","participation":[6,11,16,12,9,4,12,10,19,4,6,2,5,4,14,6,13,7,5,2,0,13,14,1,1,0,13,5,10,1,8,8,9,11,0,1,9,5,3,10,2,6,2,2,1,3,10,2,4,1,7,10],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T12:41:12.347Z"}},{"type":"Public","name":"candle","owner":"huggingface","isFork":false,"description":"Minimalist ML framework for Rust","allTopics":[],"primaryLanguage":{"name":"Rust","color":"#dea584"},"pullRequestCount":68,"issueCount":273,"starsCount":14335,"forksCount":808,"license":"Apache License 2.0","participation":[79,48,80,62,86,77,92,83,51,64,54,46,41,23,40,34,38,46,15,15,7,12,29,25,34,13,39,20,2,3,25,23,16,21,23,9,33,33,38,12,27,20,15,6,6,13,4,12,2,1,3,4],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T12:35:25.699Z"}},{"type":"Public","name":"optimum-habana","owner":"huggingface","isFork":false,"description":"Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)","allTopics":["transformers","bert","fine-tuning","hpu","habana"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":68,"issueCount":16,"starsCount":123,"forksCount":147,"license":"Apache License 2.0","participation":[4,8,13,2,5,19,10,16,3,16,13,20,8,13,5,13,14,12,12,21,15,14,9,10,5,5,7,10,16,9,4,12,14,7,14,17,16,9,12,9,11,22,14,21,5,11,10,29,17,7,3,5],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T11:59:04.055Z"}},{"type":"Public","name":"optimum-nvidia","owner":"huggingface","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":45,"starsCount":835,"forksCount":81,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T11:39:00.535Z"}}],"repositoryCount":232,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"huggingface repositories"}