{"payload":{"pageCount":41,"repositories":[{"type":"Public","name":"ai-containers","owner":"intel","isFork":false,"description":"This repository contains Dockerfiles, scripts, yaml files, Helm charts, etc. used to scale out AI containers with versions of TensorFlow and PyTorch that have been optimized for Intel platforms. Scaling is done with python, Docker, kubernetes, kubeflow, cnvrg.io, Helm, and other container orchestration frameworks for use in the cloud and on-premise","allTopics":["docker","kubernetes","docker-compose","tensorflow","scikit-learn","intel","pytorch"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":9,"issueCount":0,"starsCount":14,"forksCount":10,"license":"Apache License 2.0","participation":[0,1,0,6,0,2,1,6,3,7,10,16,1,0,3,2,0,5,0,2,2,1,2,4,2,4,2,2,0,1,7,2,0,0,3,7,3,11,3,4,1,2,4,3,8,0,0,13,13,17,13,45],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T17:13:54.149Z"}},{"type":"Public","name":"llvm","owner":"intel","isFork":false,"description":"Intel staging area for llvm.org contribution. Home for Intel LLVM-based projects.","allTopics":["sycl","oneapi","llvm","intel"],"primaryLanguage":null,"pullRequestCount":356,"issueCount":478,"starsCount":1182,"forksCount":700,"license":"Other","participation":[847,867,863,818,797,937,1036,916,768,912,855,847,679,835,783,822,737,632,585,859,814,881,761,724,643,914,796,841,540,224,660,775,948,888,850,773,703,720,796,870,828,859,674,764,726,853,791,739,733,752,371,67],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T17:12:35.624Z"}},{"type":"Public","name":"scikit-learn-intelex","owner":"intel","isFork":false,"description":"Intel(R) Extension for Scikit-learn is a seamless way to speed up your Scikit-learn application","allTopics":["big-data","analytics","gpu","machine-learning-algorithms","intel","data-analysis","ai-training","oneapi","ai-inference","swrepo","ai-machine-learning","python","machine-learning","scikit-learn"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":32,"issueCount":44,"starsCount":1169,"forksCount":169,"license":"Apache License 2.0","participation":[4,0,6,6,3,6,6,7,4,7,6,9,9,17,12,9,14,4,5,4,9,7,3,8,4,6,5,5,6,1,4,8,6,14,9,9,11,12,7,5,7,6,6,4,9,12,8,5,7,4,1,6],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T17:07:59.938Z"}},{"type":"Public","name":"compute-runtime","owner":"intel","isFork":false,"description":"Intel® Graphics Compute Runtime for oneAPI Level Zero and OpenCL™ Driver","allTopics":["gpu","gpgpu","compute","intel-hd-graphics","opencl","intel"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":5,"issueCount":73,"starsCount":1086,"forksCount":226,"license":"MIT License","participation":[24,45,53,50,56,48,59,56,56,39,24,58,63,44,56,53,54,71,57,45,46,60,50,47,41,83,57,44,56,25,34,50,56,52,37,54,59,61,65,79,67,64,63,41,65,52,49,28,42,65,59,48],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T17:06:22.018Z"}},{"type":"Public","name":"fpga-runtime-for-opencl","owner":"intel","isFork":false,"description":"Intel® FPGA Runtime for OpenCL™ Software Technology","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":15,"issueCount":21,"starsCount":29,"forksCount":65,"license":"Other","participation":[1,0,1,0,1,3,0,1,2,4,2,3,0,0,0,1,4,1,0,2,1,0,1,1,1,0,1,4,3,0,3,1,0,0,0,0,0,2,2,1,1,0,1,3,7,0,0,1,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T17:05:09.222Z"}},{"type":"Public","name":"ittapi","owner":"intel","isFork":false,"description":"Intel® Instrumentation and Tracing Technology (ITT) and Just-In-Time (JIT) API","allTopics":[],"primaryLanguage":{"name":"Rust","color":"#dea584"},"pullRequestCount":2,"issueCount":11,"starsCount":77,"forksCount":42,"license":null,"participation":[2,0,0,8,2,0,5,4,0,0,0,0,0,0,1,5,0,6,0,2,3,2,0,2,0,0,2,3,2,0,5,3,0,0,0,5,0,0,0,0,0,0,0,0,0,2,0,1,0,1,1,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T16:48:23.203Z"}},{"type":"Public","name":"intel-xpu-backend-for-triton","owner":"intel","isFork":false,"description":"OpenAI Triton backend for Intel® GPUs","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":21,"issueCount":143,"starsCount":85,"forksCount":26,"license":"MIT License","participation":[14,11,23,22,26,27,18,5,6,40,33,28,23,17,37,28,25,22,13,12,29,23,17,18,9,25,19,22,7,6,43,44,65,59,62,40,62,55,84,76,72,58,66,96,72,79,59,65,70,54,69,37],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T16:33:58.840Z"}},{"type":"Public","name":"ccc-linux-guest-hardening-docs","owner":"intel","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"HTML","color":"#e34c26"},"pullRequestCount":0,"issueCount":1,"starsCount":6,"forksCount":9,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,0,0,0,0,0,4,0,0,0,2,0,1,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T16:26:20.704Z"}},{"type":"Public","name":"intel-inb-manageability","owner":"intel","isFork":false,"description":"The Intel® In-Band Manageability Framework enables an administrator to perform critical Device Management operations over-the-air remotely from the cloud. It also facilitates the publishing of telemetry and critical events and logs from an IoT device to the cloud enabling the administrator to take corrective actions if, and when necessary. The f…","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":1,"starsCount":21,"forksCount":20,"license":"Other","participation":[5,3,14,7,0,1,1,1,3,2,5,4,4,2,1,0,1,3,2,3,2,2,8,5,4,7,6,0,2,0,3,4,7,4,5,9,2,1,2,1,0,2,0,0,2,0,5,1,1,1,1,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T16:22:06.947Z"}},{"type":"Public","name":"ipmctl","owner":"intel","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"C","color":"#555555"},"pullRequestCount":2,"issueCount":58,"starsCount":177,"forksCount":61,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,1,2,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T16:16:43.065Z"}},{"type":"Public","name":"torch-xpu-ops","owner":"intel","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":34,"issueCount":67,"starsCount":11,"forksCount":6,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,8,7,20,10,12,16,10,11,4,11,4,16,17,19,9],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T16:16:13.850Z"}},{"type":"Public","name":"dffml","owner":"intel","isFork":false,"description":"The easiest way to use Machine Learning. Mix and match underlying ML libraries and data set sources. Generate new datasets or modify existing ones with ease.","allTopics":["analytics","libraries","models","pipelines","data-flow","asyncio","flow-based-programming","datasets","dag","frameworks","event-based","dataflows","ai-training","hyperautomation","dffml","ai-inference","swrepo","ai-machine-learning","python","machine-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":26,"issueCount":381,"starsCount":243,"forksCount":135,"license":"MIT License","participation":[6,7,1,4,2,1,0,4,3,0,1,0,0,7,0,0,0,0,1,0,0,1,2,1,4,1,5,16,0,10,36,2,0,1,0,0,1,0,0,0,1,11,1,2,12,0,1,7,1,1,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T16:14:44.649Z"}},{"type":"Public","name":"intel-npu-acceleration-library","owner":"intel","isFork":false,"description":"Intel® NPU Acceleration Library","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":11,"starsCount":327,"forksCount":19,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,6,0,0,0,0,0,0,4,1,0,0,1,3,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T16:13:47.574Z"}},{"type":"Public","name":"gits","owner":"intel","isFork":false,"description":"API capture-replay tool for Vulkan, OpenCL, Intel oneAPI Level Zero and OpenGL","allTopics":["opengl","opencl","vulkan-api","oneapi","levelzero"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":1,"starsCount":35,"forksCount":7,"license":"MIT License","participation":[6,14,5,8,8,1,0,1,0,6,3,2,9,1,7,7,9,13,4,3,2,4,19,22,6,35,0,8,5,0,0,6,1,4,5,0,4,3,0,1,4,5,5,4,4,6,16,0,7,1,17,20],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T16:12:55.338Z"}},{"type":"Public","name":"gmmlib","owner":"intel","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":5,"issueCount":8,"starsCount":158,"forksCount":81,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T16:07:05.899Z"}},{"type":"Public","name":"userspace-cni-network-plugin","owner":"intel","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Go","color":"#00ADD8"},"pullRequestCount":18,"issueCount":22,"starsCount":234,"forksCount":68,"license":"Apache License 2.0","participation":[5,0,2,0,16,16,0,3,0,1,1,0,0,0,7,1,4,3,0,0,0,12,3,8,5,4,2,9,0,0,19,1,1,1,0,0,5,0,0,0,0,0,0,5,0,0,0,0,4,7,1,4],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T16:07:02.837Z"}},{"type":"Public archive","name":"kubevirt-tdx","owner":"intel","isFork":false,"description":"TDX Enabling for Kubevirt","allTopics":["kubevirt","tdx"],"primaryLanguage":{"name":"Go","color":"#00ADD8"},"pullRequestCount":5,"issueCount":3,"starsCount":4,"forksCount":1,"license":"Apache License 2.0","participation":[45,17,16,7,0,3,0,0,0,0,0,7,5,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T15:48:26.613Z"}},{"type":"Public","name":"web-ai-showcase","owner":"intel","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":1,"issueCount":0,"starsCount":2,"forksCount":1,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T15:33:50.010Z"}},{"type":"Public","name":"onnxruntime","owner":"intel","isFork":true,"description":"ONNX Runtime: cross-platform, high performance scoring engine for ML models","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":33,"issueCount":2,"starsCount":54,"forksCount":2689,"license":"MIT License","participation":[33,39,44,36,37,55,41,43,57,63,68,52,43,41,50,35,40,38,54,26,47,48,57,59,31,53,41,45,30,16,35,92,39,42,52,26,29,52,41,35,58,41,43,31,30,51,59,27,32,26,31,32],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T15:29:12.020Z"}},{"type":"Public","name":"intel-lpmd","owner":"intel","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"C","color":"#555555"},"pullRequestCount":2,"issueCount":2,"starsCount":16,"forksCount":9,"license":"GNU General Public License v2.0","participation":[0,0,0,0,0,0,0,0,1,1,0,0,0,8,1,0,1,13,4,1,0,0,0,0,0,0,0,8,0,0,0,0,0,12,1,3,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T15:14:19.068Z"}},{"type":"Public","name":"intel-extension-for-openxla","owner":"intel","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":3,"issueCount":3,"starsCount":28,"forksCount":6,"license":"Apache License 2.0","participation":[0,0,3,2,2,5,0,9,2,6,5,5,4,5,5,1,3,5,14,7,4,8,3,2,5,5,5,2,3,4,3,4,4,3,10,0,4,6,3,6,4,4,3,6,6,6,13,5,4,3,4,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T15:09:19.638Z"}},{"type":"Public","name":"ai-visual-inference-samples","owner":"intel","isFork":false,"description":"Intel® AI Visual Inference Samples are easy to use script implementations of workloads with processing pipelines consisting of media and AI inference elements. Samples are built on top of open source media and AI inference frameworks to demonstrate functionality and performance of Intel® Data Center GPU Flex Series for AI visual inference use cases","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":0,"starsCount":3,"forksCount":0,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T15:04:34.186Z"}},{"type":"Public","name":"cluster-management-toolkit","owner":"intel","isFork":false,"description":"Toolkit for managing and monitoring Kubernetes clusters; includes a Curses-based console UI as well as a few command-line tools.","allTopics":["kubernetes"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":11,"starsCount":6,"forksCount":2,"license":"MIT License","participation":[11,9,1,14,15,5,2,9,7,4,4,0,8,8,5,1,13,13,38,20,28,11,14,10,11,5,8,18,10,5,21,8,21,33,51,25,42,36,45,35,22,30,25,23,23,9,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T14:59:14.353Z"}},{"type":"Public","name":"neural-compressor","owner":"intel","isFork":false,"description":"SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX Runtime","allTopics":["sparsity","pruning","quantization","knowledge-distillation","auto-tuning","int8","low-precision","quantization-aware-training","post-training-quantization","awq","int4","large-language-models","gptq","smoothquant","sparsegpt","fp4","mxformat"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":18,"issueCount":25,"starsCount":2022,"forksCount":243,"license":"Apache License 2.0","participation":[15,29,15,17,14,12,13,12,19,6,17,23,14,31,27,12,12,7,10,9,12,12,4,16,20,14,13,17,14,9,7,10,20,18,14,7,1,12,20,11,14,8,13,5,9,9,16,4,9,14,14,8],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T14:35:28.863Z"}},{"type":"Public","name":"intel-extension-for-transformers","owner":"intel","isFork":false,"description":"⚡ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Platforms⚡","allTopics":["retrieval","chatbot","rag","habana","large-language-model","chatpdf","llm-inference","4-bits","speculative-decoding","llm-cpu","streamingllm","intel-optimized-llamacpp","neural-chat","neural-chat-7b","autoround","gaudi3"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":27,"issueCount":43,"starsCount":1999,"forksCount":190,"license":"Apache License 2.0","participation":[22,23,9,32,30,29,17,19,21,16,11,14,39,56,41,48,34,6,38,45,43,46,40,37,60,68,47,35,82,10,14,19,37,27,46,10,10,25,20,15,28,27,16,12,16,20,10,3,15,6,17,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T13:57:05.815Z"}},{"type":"Public","name":"kubernetes-power-manager","owner":"intel","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Go","color":"#00ADD8"},"pullRequestCount":9,"issueCount":3,"starsCount":76,"forksCount":18,"license":"Apache License 2.0","participation":[0,2,0,2,0,0,3,3,12,0,0,1,0,3,0,0,0,0,0,2,0,0,0,1,2,0,0,0,2,0,0,0,0,0,0,0,0,3,2,0,0,0,0,0,1,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T13:47:32.530Z"}},{"type":"Public","name":"auto-round","owner":"intel","isFork":false,"description":"SOTA Weight-only Quantization Algorithm for LLMs. This is official implementation of \"Optimize Weight Rounding via Signed Gradient Descent for the Quantization of LLMs\"","allTopics":["rounding","quantization","awq","int4","gptq","neural-compressor","weight-only"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":10,"starsCount":84,"forksCount":11,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,18,13,34,14,5,3,12,14,7,12,8,7,3,5,3,9,3,2,5,10,7],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T12:13:56.303Z"}},{"type":"Public","name":"vpl-gpu-rt","owner":"intel","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":14,"issueCount":33,"starsCount":96,"forksCount":87,"license":"MIT License","participation":[10,6,12,10,8,4,6,4,4,9,4,10,15,6,7,2,4,2,4,11,10,7,9,10,7,5,8,13,5,2,8,7,3,5,8,0,1,3,7,6,3,8,8,6,3,12,5,2,4,6,5,5],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T11:20:20.872Z"}},{"type":"Public","name":"intel-application-migration-tool-for-openacc-to-openmp","owner":"intel","isFork":false,"description":"OpenACC* to OpenMP* API assisting migration tool","allTopics":["openmp","openacc"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":32,"forksCount":5,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,6,0,0,0,0,0,0,0,1,0,0,0,2,0,0,1,0,1,1,0,3,0,1,4,1,0,0,0,0,0,3,2,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T10:56:04.591Z"}},{"type":"Public","name":"xFasterTransformer","owner":"intel","isFork":false,"description":"","allTopics":["intel","inference","transformer","xeon","llama","model-serving","llm","chatglm","qwen"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":9,"issueCount":10,"starsCount":247,"forksCount":47,"license":"Apache License 2.0","participation":[0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,2,0,5,12,5,11,8,15,8,14,6,12,0,13,4,9,11,11,1,4,17,5,9,2,6,9,14,14,9,12,3,8,11,4,6],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T10:54:13.804Z"}}],"repositoryCount":1210,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}