{"payload":{"header_redesign_enabled":false,"results":[{"id":"66823715","archived":false,"color":"#3572A5","followers":6048,"has_funding_file":false,"hl_name":"intel-analytics/ipex-llm","hl_trunc_description":"Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, etc.) on Intel CPU and GPU (e.g.,…","language":"Python","mirror":false,"owned_by_organization":true,"public":true,"repo":{"repository":{"id":66823715,"name":"ipex-llm","owner_id":10941215,"owner_login":"intel-analytics","updated_at":"2024-05-11T14:09:12.760Z","has_issues":true}},"sponsorable":false,"topics":["gpu","transformers","pytorch","llm"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":11,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":87,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253Aintel-analytics%252Fipex-llm%2B%2Blanguage%253APython","metadata":null,"csrf_tokens":{"/intel-analytics/ipex-llm/star":{"post":"7UyOGbIfirc1XGUhYDTtTjJu4j8359HVLHeneWwKXrCm2rpthu-t4AIKu8Ch9uFxY1ln8T7q4GRnYkxwFbHssg"},"/intel-analytics/ipex-llm/unstar":{"post":"yq-IE7hJJD23WA_eXpjTgugkVNEvX2A9NOV6v1LLn5d-9jN7fXGvoX1xVM2IwqPQRb1SewK7pdrfFPBuG7Nvsg"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"-mMdonM4havMvMDOdukURvxz7xzCm1ST3TrFWSkqmHom1KWT0f0oKGcXJG-gG4zBNPj_3umLO1rX37I8YL4zqg"}}},"title":"Repository search results"}