{"payload":{"pageCount":8,"repositories":[{"type":"Public","name":"diffusers","owner":"huggingface","isFork":false,"description":"🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch and FLAX.","allTopics":["deep-learning","pytorch","image-generation","flax","hacktoberfest","diffusion","text2image","image2image","jax","score-based-generative-modeling","stable-diffusion","stable-diffusion-diffusers","latent-diffusion-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":129,"issueCount":338,"starsCount":23079,"forksCount":4760,"license":"Apache License 2.0","participation":[39,33,18,16,17,35,36,36,45,47,27,24,59,38,27,49,42,43,31,21,30,34,36,45,45,31,47,29,30,40,38,40,41,31,28,34,44,35,27,48,39,36,31,35,18,24,22,25,27,28,24,19],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T20:03:01.589Z"}},{"type":"Public","name":"chat-ui","owner":"huggingface","isFork":false,"description":"Open source codebase powering the HuggingChat app","allTopics":["typescript","svelte","hacktoberfest","tailwindcss","huggingface","svelte-kit","sveltekit","llm","chatgpt"],"primaryLanguage":{"name":"TypeScript","color":"#3178c6"},"pullRequestCount":26,"issueCount":188,"starsCount":6462,"forksCount":899,"license":"Apache License 2.0","participation":[4,5,5,5,0,1,4,9,4,5,0,10,6,0,0,9,17,6,12,13,14,9,1,0,13,6,6,17,6,3,11,5,4,14,15,23,18,19,16,8,19,12,9,5,20,12,24,23,19,8,7,22],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T20:00:56.289Z"}},{"type":"Public","name":"autotrain-advanced","owner":"huggingface","isFork":false,"description":"🤗 AutoTrain Advanced","allTopics":["natural-language-processing","natural-language-understanding","huggingface","autotrain","python","machine-learning","deep-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":18,"starsCount":3455,"forksCount":413,"license":"Apache License 2.0","participation":[8,5,2,7,0,5,5,7,27,5,11,21,13,20,1,4,2,9,4,3,11,9,3,3,1,14,4,30,35,36,3,0,0,1,4,4,6,15,11,10,20,16,15,10,5,7,8,38,7,29,11,20],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T19:50:15.600Z"}},{"type":"Public","name":"huggingface.js","owner":"huggingface","isFork":false,"description":"Utilities to use the Hugging Face Hub API","allTopics":["machine-learning","inference","hub","api-client","huggingface"],"primaryLanguage":{"name":"TypeScript","color":"#3178c6"},"pullRequestCount":25,"issueCount":87,"starsCount":1234,"forksCount":140,"license":"MIT License","participation":[5,7,2,0,5,1,0,8,6,0,0,0,2,0,0,0,0,5,6,10,0,13,5,12,11,24,21,17,22,9,0,1,2,13,6,3,8,8,23,20,13,26,12,14,21,20,11,8,16,14,23,21],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T19:46:30.182Z"}},{"type":"Public","name":"transformers","owner":"huggingface","isFork":false,"description":"🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.","allTopics":["python","seq2seq","flax","language-models","nlp-library","hacktoberfest","jax","pytorch-transformers","model-hub","nlp","machine-learning","natural-language-processing","deep-learning","tensorflow","pytorch","transformer","speech-recognition","pretrained-models","language-model","bert"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":235,"issueCount":843,"starsCount":126772,"forksCount":25098,"license":"Apache License 2.0","participation":[75,81,67,68,60,26,64,68,70,53,65,67,64,75,48,64,36,50,63,62,55,64,74,68,62,49,43,81,62,38,2,29,58,51,52,50,43,52,51,55,63,60,58,48,56,59,72,72,65,55,59,65],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T19:39:22.768Z"}},{"type":"Public","name":"lerobot","owner":"huggingface","isFork":false,"description":"🤗 LeRobot: State-of-the-art Machine Learning for Real-World Robotics in Pytorch","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":22,"issueCount":10,"starsCount":3264,"forksCount":235,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,3,1,5,22,41,65,37,76,44,56,66,45,18,21,16,13,5],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T19:21:00.812Z"}},{"type":"Public","name":"trl","owner":"huggingface","isFork":false,"description":"Train transformer language models with reinforcement learning.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":15,"issueCount":57,"starsCount":8378,"forksCount":1011,"license":"Apache License 2.0","participation":[7,17,7,14,2,10,8,13,9,6,6,8,8,14,10,21,8,7,7,9,9,7,17,8,7,4,10,12,4,17,8,10,19,11,8,15,1,10,3,12,5,7,21,4,11,13,16,5,10,3,3,12],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T18:26:34.779Z"}},{"type":"Public","name":"optimum-habana","owner":"huggingface","isFork":false,"description":"Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)","allTopics":["transformers","bert","fine-tuning","hpu","habana"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":51,"issueCount":15,"starsCount":113,"forksCount":129,"license":"Apache License 2.0","participation":[4,11,9,6,4,8,4,9,10,2,11,13,14,14,5,16,15,15,11,9,7,18,11,11,14,18,20,9,13,5,4,7,6,11,14,11,8,10,12,11,10,24,8,14,8,10,10,24,22,10,8,9],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T18:07:24.298Z"}},{"type":"Public","name":"text-generation-inference","owner":"huggingface","isFork":false,"description":"Large Language Model Text Generation Inference","allTopics":["nlp","bloom","deep-learning","inference","pytorch","falcon","transformer","gpt","starcoder"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":9,"issueCount":136,"starsCount":8111,"forksCount":893,"license":"Apache License 2.0","participation":[18,5,2,7,7,10,17,18,18,8,18,8,5,7,12,3,4,19,14,2,8,1,0,4,3,8,5,6,15,5,0,0,6,4,21,10,8,12,8,12,0,5,11,4,7,17,8,18,20,2,21,15],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T18:04:27.795Z"}},{"type":"Public","name":"ratchet","owner":"huggingface","isFork":false,"description":"A cross-platform browser ML framework.","allTopics":[],"primaryLanguage":{"name":"Rust","color":"#dea584"},"pullRequestCount":5,"issueCount":27,"starsCount":381,"forksCount":18,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,23,90,69,80,69,29,72,46,24,40,59,19,45,76,47,32,13,0,6,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T17:46:17.298Z"}},{"type":"Public","name":"api-inference-community","owner":"huggingface","isFork":false,"description":"","allTopics":["hacktoberfest"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":40,"issueCount":32,"starsCount":143,"forksCount":57,"license":"Apache License 2.0","participation":[6,1,4,4,0,3,1,1,6,8,1,0,1,0,0,0,0,1,3,2,1,2,0,3,1,0,3,7,0,1,3,0,2,0,0,4,1,0,0,5,0,1,1,2,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T17:08:26.643Z"}},{"type":"Public","name":"cookbook","owner":"huggingface","isFork":false,"description":"Open-source AI cookbook","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":17,"issueCount":15,"starsCount":1387,"forksCount":180,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,2,0,4,0,0,0,2,30,39,31,1,42,49,30,27,18,4,3,3,0,3,9,4,14,3,5],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T16:52:22.169Z"}},{"type":"Public","name":"optimum-neuron","owner":"huggingface","isFork":false,"description":"Easy, fast and very cheap training and inference on AWS Trainium and Inferentia chips.","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":22,"issueCount":81,"starsCount":168,"forksCount":46,"license":"Apache License 2.0","participation":[12,7,5,8,2,13,5,11,0,7,2,10,10,4,7,8,7,4,3,5,5,4,6,7,11,13,2,0,5,6,0,0,4,11,7,6,7,9,2,4,4,1,5,11,10,11,4,3,5,1,7,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T16:45:46.831Z"}},{"type":"Public","name":"optimum-intel","owner":"huggingface","isFork":false,"description":"🤗 Optimum Intel: Accelerate inference with Intel optimization tools","allTopics":["optimization","intel","transformers","pruning","distillation","onnx","openvino","diffusers","inference","quantization"],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":23,"issueCount":10,"starsCount":338,"forksCount":93,"license":"Apache License 2.0","participation":[10,11,7,4,2,11,2,5,25,8,3,1,7,0,6,3,5,5,3,4,4,1,10,4,0,0,3,4,6,2,3,8,7,10,11,13,7,9,4,8,12,14,15,13,7,8,13,27,16,4,17,5],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T16:19:29.901Z"}},{"type":"Public","name":"peft","owner":"huggingface","isFork":false,"description":"🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.","allTopics":["python","adapter","transformers","pytorch","lora","diffusion","parameter-efficient-learning","llm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":10,"issueCount":38,"starsCount":14321,"forksCount":1330,"license":"Apache License 2.0","participation":[15,4,15,10,13,5,20,11,5,5,6,8,9,8,7,9,6,10,12,7,1,7,9,18,16,10,17,21,17,5,3,5,19,6,3,15,19,18,12,5,12,9,7,4,8,9,4,14,5,6,11,9],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T16:13:50.961Z"}},{"type":"Public","name":"datatrove","owner":"huggingface","isFork":false,"description":"Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":17,"starsCount":1409,"forksCount":88,"license":"Apache License 2.0","participation":[0,0,14,19,23,33,9,5,14,4,9,3,0,4,4,2,0,0,3,4,16,1,10,3,23,22,11,6,11,3,1,0,1,9,7,11,7,7,7,6,7,1,9,0,3,5,12,2,5,7,4,13],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T16:11:49.032Z"}},{"type":"Public","name":"optimum","owner":"huggingface","isFork":false,"description":"🚀 Accelerate training and inference of 🤗 Transformers and 🤗 Diffusers with easy to use hardware optimization tools","allTopics":["training","optimization","intel","transformers","tflite","onnxruntime","graphcore","habana","inference","pytorch","quantization","onnx"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":85,"issueCount":272,"starsCount":2217,"forksCount":383,"license":"Apache License 2.0","participation":[6,3,13,2,16,7,6,14,13,12,9,3,14,13,14,5,5,2,7,7,9,8,14,4,6,1,2,11,15,0,1,3,10,6,9,1,9,10,8,9,0,4,9,2,7,7,1,7,1,2,1,6],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T16:03:09.037Z"}},{"type":"Public","name":"datasets","owner":"huggingface","isFork":false,"description":"🤗 The largest hub of ready-to-use datasets for ML models with fast, easy-to-use and efficient data manipulation tools","allTopics":["machine-learning","natural-language-processing","computer-vision","deep-learning","tensorflow","numpy","speech","pandas","datasets","hacktoberfest","nlp","pytorch"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":79,"issueCount":625,"starsCount":18588,"forksCount":2541,"license":"Apache License 2.0","participation":[12,6,13,5,6,6,13,9,17,3,3,8,6,9,9,5,3,6,1,7,14,4,7,2,14,8,11,5,11,15,4,1,6,1,11,10,10,5,3,20,5,6,6,1,5,9,9,2,7,3,8,7],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T16:01:19.395Z"}},{"type":"Public","name":"blog","owner":"huggingface","isFork":false,"description":"Public repo for HF blog posts","allTopics":["hacktoberfest"],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":66,"issueCount":133,"starsCount":2071,"forksCount":638,"license":null,"participation":[19,34,25,21,9,8,12,32,12,12,12,4,35,18,19,40,11,19,11,9,11,15,12,14,3,3,8,22,10,6,3,10,6,10,13,24,6,14,27,25,6,16,21,3,18,21,28,11,9,14,12,24],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T16:00:00.925Z"}},{"type":"Public","name":"optimum-tpu","owner":"huggingface","isFork":false,"description":"Google TPU optimizations for transformers models","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":1,"starsCount":31,"forksCount":5,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,2,0,2,1,2,11,3,1,4,7,0,3,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T15:19:46.526Z"}},{"type":"Public","name":"dataset-viewer","owner":"huggingface","isFork":false,"description":"Lightweight web API for visualizing and exploring any dataset - computer vision, speech, text, and tabular - stored on the Hugging Face Hub","allTopics":["nlp","data","machine-learning","api-rest","datasets","huggingface"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":144,"starsCount":627,"forksCount":62,"license":"Apache License 2.0","participation":[13,32,23,28,27,32,6,21,25,51,19,22,13,18,21,15,22,38,32,24,17,23,16,19,18,10,7,16,14,7,0,13,18,15,9,18,23,14,24,24,12,23,10,13,16,24,13,9,10,11,30,16],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T15:04:02.180Z"}},{"type":"Public","name":"accelerate","owner":"huggingface","isFork":false,"description":"🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":17,"issueCount":103,"starsCount":7153,"forksCount":833,"license":"Apache License 2.0","participation":[19,33,13,15,20,4,31,14,10,10,10,10,13,12,9,16,2,5,5,11,4,10,10,15,19,3,8,14,5,6,5,7,15,10,6,9,14,13,9,20,11,8,10,9,13,15,8,12,14,10,6,7],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T14:10:14.071Z"}},{"type":"Public","name":"scheduler-plugins","owner":"huggingface","isFork":true,"description":"Repository for out-of-tree scheduler plugins based on scheduler framework.","allTopics":[],"primaryLanguage":{"name":"Go","color":"#00ADD8"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":479,"license":"Apache License 2.0","participation":[3,0,1,5,3,0,2,3,11,9,2,2,1,6,6,2,0,0,2,0,0,2,9,4,17,4,4,1,6,0,1,3,1,1,5,1,0,0,0,2,5,0,0,0,0,18,5,0,5,5,8,9],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T14:08:25.075Z"}},{"type":"Public","name":"hub-docs","owner":"huggingface","isFork":false,"description":"Docs of the Hugging Face Hub","allTopics":["machine-learning","hacktoberfest"],"primaryLanguage":null,"pullRequestCount":18,"issueCount":69,"starsCount":238,"forksCount":202,"license":"Apache License 2.0","participation":[12,8,10,19,10,12,7,2,6,1,4,2,4,8,9,9,9,15,17,11,14,16,17,13,11,19,19,4,5,2,0,3,9,3,7,6,8,2,2,6,1,14,2,3,7,6,2,2,1,1,5,5],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T12:47:13.612Z"}},{"type":"Public","name":"huggingface_hub","owner":"huggingface","isFork":false,"description":"The official Python client for the Huggingface Hub.","allTopics":["machine-learning","natural-language-processing","deep-learning","models","pytorch","pretrained-models","hacktoberfest","model-hub"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":11,"issueCount":126,"starsCount":1745,"forksCount":449,"license":"Apache License 2.0","participation":[8,5,8,3,7,10,0,3,3,1,4,10,4,14,28,10,13,2,9,15,7,10,16,7,7,22,11,3,13,11,0,9,11,8,4,1,6,4,10,14,6,8,12,8,10,14,15,18,9,0,1,11],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T12:01:24.003Z"}},{"type":"Public","name":"dataspeech","owner":"huggingface","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":8,"starsCount":203,"forksCount":22,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,6,7,0,0,0,0,0,0,27,5,3,36,9,2,2,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T12:00:25.403Z"}},{"type":"Public","name":"nanotron","owner":"huggingface","isFork":false,"description":"Minimalistic large language model 3D-parallelism training","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":24,"issueCount":30,"starsCount":829,"forksCount":71,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,2,1,0,0,0,0,0,0,0,0,0,0,7,24,8,19,24,103,85,65,69,92,29,55,37,30,19,14,34,67,32,41,13,15,7,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T11:21:44.984Z"}},{"type":"Public","name":"lm-evaluation-harness","owner":"huggingface","isFork":true,"description":"A framework for few-shot evaluation of language models.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":0,"starsCount":2,"forksCount":1397,"license":"MIT License","participation":[27,77,83,98,82,77,118,51,24,82,153,84,71,64,43,71,60,9,40,22,40,8,53,18,36,67,102,33,42,27,5,9,16,15,12,13,13,6,19,15,16,10,10,5,5,0,4,5,14,10,2,12],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T10:08:47.653Z"}},{"type":"Public","name":"tgi-gaudi","owner":"huggingface","isFork":true,"description":"Large Language Model Text Generation Inference on Habana Gaudi","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":3,"starsCount":17,"forksCount":893,"license":"Apache License 2.0","participation":[18,5,2,7,7,10,17,18,18,8,18,8,5,7,12,3,4,19,14,2,8,1,0,4,3,8,5,11,15,7,0,1,10,11,22,10,8,16,18,27,2,10,13,9,11,20,0,5,6,0,4,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T10:05:10.151Z"}},{"type":"Public","name":"pytorch-image-models","owner":"huggingface","isFork":false,"description":"The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (ViT), MobileNet-V3/V2, RegNet, DPN, CSPNet, Swin Transformer, MaxViT, CoAtNet, ConvNeXt, and more","allTopics":["pytorch","imagenet","image-classification","resnet","pretrained-models","mixnet","pretrained-weights","distributed-training","dual-path-networks","mobilenet-v2","mobile-deep-learning","mobilenetv3","efficientnet","augmix","randaugment","nfnets","normalization-free-training","vision-transformer-models","convnext","maxvit"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":28,"issueCount":64,"starsCount":30206,"forksCount":4598,"license":"Apache License 2.0","participation":[3,2,7,1,1,1,0,1,10,19,11,22,18,18,0,2,4,8,5,2,21,6,10,2,12,17,1,8,2,1,5,3,4,3,9,5,16,8,0,0,2,7,8,0,3,17,4,1,18,16,17,4],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T05:10:59.449Z"}}],"repositoryCount":226,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}