{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"ipex-llm","owner":"intel-analytics","isFork":false,"description":"Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max). A PyTorch LLM library that seamlessly integrates with llama.cpp, Ollama, HuggingFace, LangChain, LlamaIndex, DeepSpeed, vLLM, FastChat, etc.","topicNames":["gpu","transformers","pytorch","llm"],"topicsNotShown":0,"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":245,"issueCount":799,"starsCount":6040,"forksCount":1204,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-11T06:17:10.914Z"}},{"type":"Public","name":"analytics-zoo","owner":"intel-analytics","isFork":false,"description":"Distributed Tensorflow, Keras and PyTorch on Apache Spark/Flink & Ray","topicNames":[],"topicsNotShown":0,"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":406,"starsCount":9,"forksCount":3,"license":"Apache License 2.0","participation":[0,0,0,0,3,0,0,1,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,2,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-07T08:59:07.065Z"}},{"type":"Public","name":"zoo-tutorials","owner":"intel-analytics","isFork":false,"description":"Tutorials for Analytics Zoo","topicNames":[],"topicsNotShown":0,"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":1,"issueCount":2,"starsCount":6,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-07T07:52:56.548Z"}},{"type":"Public","name":"BigDL-PPML-Azure-Occlum-Example","owner":"intel-analytics","isFork":false,"description":"","topicNames":[],"topicsNotShown":0,"primaryLanguage":{"name":"Shell","color":"#89e051"},"pullRequestCount":2,"issueCount":0,"starsCount":2,"forksCount":7,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-07T07:52:28.193Z"}},{"type":"Public","name":"text-generation-webui","owner":"intel-analytics","isFork":true,"description":"A Gradio Web UI for running local LLM on Intel GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max) using IPEX-LLM.","topicNames":[],"topicsNotShown":0,"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":4,"starsCount":13,"forksCount":4908,"license":"GNU Affero General Public License v3.0","participation":[57,68,76,54,31,62,51,35,50,44,45,51,50,56,41,47,1,79,93,67,13,25,24,51,16,74,36,45,19,36,67,66,27,29,43,12,22,13,30,29,34,0,2,3,6,7,0,5,4,3,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-05T20:19:43.135Z"}},{"type":"Public","name":"langchain","owner":"intel-analytics","isFork":true,"description":"🦜🔗 Build context-aware reasoning applications","topicNames":[],"topicsNotShown":0,"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":13056,"license":"MIT License","participation":[197,112,112,119,91,177,113,180,133,145,162,146,193,168,187,214,184,100,131,169,150,136,150,169,180,109,117,79,152,159,104,134,63,128,128,125,154,105,155,138,118,166,174,133,116,155,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-29T08:50:10.984Z"}},{"type":"Public","name":"BigDL-2.x","owner":"intel-analytics","isFork":false,"description":"BigDL: Distributed TensorFlow, Keras and PyTorch on Apache Spark/Flink & Ray","topicNames":["python","scala","apache-spark","pytorch","keras-tensorflow","bigdl","distributed-deep-learning","deep-neural-network","analytics-zoo"],"topicsNotShown":0,"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":2,"issueCount":0,"starsCount":2639,"forksCount":731,"license":"Apache License 2.0","participation":[17,11,21,41,36,23,38,33,59,49,27,41,33,25,19,33,41,30,38,36,8,60,38,46,51,27,41,29,34,41,47,47,34,25,32,33,51,39,69,0,82,66,32,49,47,5,4,2,0,1,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-26T02:36:51.492Z"}},{"type":"Public","name":"llama_index","owner":"intel-analytics","isFork":true,"description":"LlamaIndex is a data framework for your LLM applications","topicNames":[],"topicsNotShown":0,"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":4308,"license":"MIT License","participation":[29,28,40,49,32,40,42,51,51,42,39,38,34,41,52,54,50,51,44,74,66,58,30,64,67,94,65,72,48,71,66,73,33,58,76,84,66,44,49,90,347,97,81,70,70,102,79,24,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-25T03:11:27.885Z"}},{"type":"Public","name":"Langchain-Chatchat","owner":"intel-analytics","isFork":true,"description":"Knowledge Base QA using RAG pipeline on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max) with IPEX-LLM","topicNames":[],"topicsNotShown":0,"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":1,"starsCount":8,"forksCount":4888,"license":"Apache License 2.0","participation":[73,69,35,27,50,26,12,14,27,46,35,61,119,148,72,42,89,52,43,29,12,20,32,28,12,30,20,20,39,18,14,11,8,14,21,18,45,28,14,0,5,4,5,9,53,32,13,4,1,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-23T07:49:43.385Z"}},{"type":"Public","name":"private-gpt","owner":"intel-analytics","isFork":true,"description":"Interact with your documents using the power of GPT, 100% privately, no data leaks","topicNames":[],"topicsNotShown":0,"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":0,"forksCount":6955,"license":"Apache License 2.0","participation":[43,17,3,2,11,0,0,1,1,4,1,0,0,4,1,3,0,2,0,5,0,0,10,6,5,11,16,11,12,4,6,9,2,0,1,3,1,1,2,3,2,0,2,10,8,0,11,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-18T03:11:04.353Z"}},{"type":"Public","name":"FastChat","owner":"intel-analytics","isFork":true,"description":"An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.","topicNames":[],"topicsNotShown":0,"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":0,"forksCount":4227,"license":"Apache License 2.0","participation":[14,17,2,22,18,30,22,25,26,31,11,20,20,15,22,12,11,12,20,7,11,17,10,8,9,11,6,15,7,0,11,3,24,3,8,9,15,6,9,10,0,2,0,1,1,2,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-26T09:19:11.892Z"}},{"type":"Public","name":"ipex-llm-tutorial","owner":"intel-analytics","isFork":false,"description":"Accelerate LLM with low-bit (FP4 / INT4 / FP8 / INT8) optimizations using ipex-llm","topicNames":[],"topicsNotShown":0,"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":4,"issueCount":8,"starsCount":111,"forksCount":30,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,4,36,48,33,1,4,14,27,55,25,0,0,1,0,0,2,0,0,0,0,8,2,0,0,1,2,0,1,1,0,0,0,0,0,0,1,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-26T07:09:50.060Z"}},{"type":"Public","name":"BigDL-core","owner":"intel-analytics","isFork":false,"description":"Core HW bindings and optimizations for BigDL","topicNames":[],"topicsNotShown":0,"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":3,"issueCount":11,"starsCount":33,"forksCount":36,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-05T04:40:42.756Z"}}],"repositoryCount":13,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"mirror","text":"Mirrors"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}