{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"unsloth","owner":"unslothai","isFork":false,"description":"Finetune Llama 3, Mistral & Gemma LLMs 2-5x faster with 80% less memory","topicNames":["ai","llama","lora","gemma","mistral","fine-tuning","finetuning","llms","qlora","llama2"],"topicsNotShown":0,"allTopics":["ai","llama","lora","gemma","mistral","fine-tuning","finetuning","llms","qlora","llama2"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":12,"issueCount":211,"starsCount":9778,"forksCount":621,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-17T13:46:33.513Z"}},{"type":"Public","name":"transformers","owner":"unslothai","isFork":true,"description":"🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":24996,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-19T11:31:59.295Z"}},{"type":"Public","name":"gemma_pytorch","owner":"unslothai","isFork":true,"description":"The official PyTorch implementation of Google's Gemma models","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":478,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-02T06:10:39.775Z"}}],"repositoryCount":3,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}