{"payload":{"feedbackUrl":"https://github.com/orgs/community/discussions/53140","repo":{"id":66823715,"defaultBranch":"main","name":"ipex-llm","ownerLogin":"intel-analytics","currentUserCanPush":false,"isFork":false,"isEmpty":false,"createdAt":"2016-08-29T07:59:50.000Z","ownerAvatar":"https://avatars.githubusercontent.com/u/10941215?v=4","public":true,"private":false,"isOrgOwned":true},"refInfo":{"name":"","listCacheKey":"v0:1714111389.0","currentOid":""},"activityList":{"items":[{"before":"9f6358e4c2522f8f95db4825fab68dc8c0592b0f","after":"f8dd2e52ad6b2c0ec71330fb380289c8e248799a","ref":"refs/heads/main","pushedAt":"2024-05-11T06:40:37.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"hkvision","name":"Kai Huang","path":"/hkvision","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/21994046?s=80&v=4"},"commit":{"message":"Fix Langchain upstream ut (#10985)\n\n* Fix Langchain upstream ut\r\n\r\n* Small fix\r\n\r\n* Install bigdl-llm\r\n\r\n* Update run-langchain-upstream-tests.sh\r\n\r\n* Update run-langchain-upstream-tests.sh\r\n\r\n* Update llm_unit_tests.yml\r\n\r\n* Update run-langchain-upstream-tests.sh\r\n\r\n* Update llm_unit_tests.yml\r\n\r\n* Update run-langchain-upstream-tests.sh\r\n\r\n* fix git checkout\r\n\r\n* fix\r\n\r\n---------\r\n\r\nCo-authored-by: Zhangky11 <2321096202@qq.com>\r\nCo-authored-by: Keyan (Kyrie) Zhang <79576162+Zhangky11@users.noreply.github.com>","shortMessageHtmlLink":"Fix Langchain upstream ut (#10985)"}},{"before":"5e0872073e9372d648918f32ff55aec0bb558db0","after":"9f6358e4c2522f8f95db4825fab68dc8c0592b0f","ref":"refs/heads/main","pushedAt":"2024-05-11T04:33:35.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"Oscilloscope98","name":"Yuwen Hu","path":"/Oscilloscope98","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/54161268?s=80&v=4"},"commit":{"message":"Deprecate support for pytorch 2.0 on Linux for `ipex-llm >= 2.1.0b20240511` (#10986)\n\n* Remove xpu_2.0 option in setup.py\r\n\r\n* Disable xpu_2.0 test in UT and nightly\r\n\r\n* Update docs for deprecated pytorch 2.0\r\n\r\n* Small doc update","shortMessageHtmlLink":"Deprecate support for pytorch 2.0 on Linux for `ipex-llm >= 2.1.0b202…"}},{"before":"ad96f32ce01ef969df60c5972f93536cb8ebe8f2","after":"5e0872073e9372d648918f32ff55aec0bb558db0","ref":"refs/heads/main","pushedAt":"2024-05-11T01:20:31.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"rnwang04","name":"Ruonan Wang","path":"/rnwang04","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/105281011?s=80&v=4"},"commit":{"message":"add version for llama.cpp and ollama (#10982)\n\n* add version for cpp\r\n\r\n* meet review","shortMessageHtmlLink":"add version for llama.cpp and ollama (#10982)"}},{"before":"cfed76b2ed90bb67bf882655389ada3a734e910f","after":"ad96f32ce01ef969df60c5972f93536cb8ebe8f2","ref":"refs/heads/main","pushedAt":"2024-05-10T09:33:46.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"MeouSker77","name":"Yishuo Wang","path":"/MeouSker77","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34214113?s=80&v=4"},"commit":{"message":"optimize phi3 1st token performance (#10981)","shortMessageHtmlLink":"optimize phi3 1st token performance (#10981)"}},{"before":"f9615f12d1f3977455f6b42730f5f8115d3b50b7","after":"cfed76b2ed90bb67bf882655389ada3a734e910f","ref":"refs/heads/main","pushedAt":"2024-05-10T08:40:16.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"lalalapotter","name":"Cengguang Zhang","path":"/lalalapotter","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/27332689?s=80&v=4"},"commit":{"message":"LLM: add long-context support for Qwen1.5-7B/Baichuan2-7B/Mistral-7B. (#10937)\n\n* LLM: add split tensor support for baichuan2-7b and qwen1.5-7b.\r\n\r\n* fix style.\r\n\r\n* fix style.\r\n\r\n* fix style.\r\n\r\n* add support for mistral and fix condition threshold.\r\n\r\n* fix style.\r\n\r\n* fix comments.","shortMessageHtmlLink":"LLM: add long-context support for Qwen1.5-7B/Baichuan2-7B/Mistral-7B. ("}},{"before":"a6342cc068a331ce84ed793686ce74c362d74acc","after":"f9615f12d1f3977455f6b42730f5f8115d3b50b7","ref":"refs/heads/main","pushedAt":"2024-05-10T07:02:58.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"plusbang","name":"binbin Deng","path":"/plusbang","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/108676127?s=80&v=4"},"commit":{"message":"Add driver related packages version check in env script (#10977)","shortMessageHtmlLink":"Add driver related packages version check in env script (#10977)"}},{"before":"e7531258804a72b6dbba9502f9b0dfd1fc31a627","after":"a6342cc068a331ce84ed793686ce74c362d74acc","ref":"refs/heads/main","pushedAt":"2024-05-09T11:50:04.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"hkvision","name":"Kai Huang","path":"/hkvision","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/21994046?s=80&v=4"},"commit":{"message":"Empty cache after phi first attention to support 4k input (#10972)\n\n* empty cache\r\n\r\n* fix style","shortMessageHtmlLink":"Empty cache after phi first attention to support 4k input (#10972)"}},{"before":"b7f7d05a7e5b3f5da9d571e43703180390a30068","after":"e7531258804a72b6dbba9502f9b0dfd1fc31a627","ref":"refs/heads/main","pushedAt":"2024-05-09T09:02:59.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"MeouSker77","name":"Yishuo Wang","path":"/MeouSker77","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34214113?s=80&v=4"},"commit":{"message":"use fp16_sdp when head_dim=96 (#10976)","shortMessageHtmlLink":"use fp16_sdp when head_dim=96 (#10976)"}},{"before":"697ca79ecad55f5496a83bec4ffdd94bde704cce","after":"b7f7d05a7e5b3f5da9d571e43703180390a30068","ref":"refs/heads/main","pushedAt":"2024-05-09T08:44:12.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"rnwang04","name":"Ruonan Wang","path":"/rnwang04","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/105281011?s=80&v=4"},"commit":{"message":"update llama.cpp usage of llama3 (#10975)\n\n* update llama.cpp usage of llama3\r\n\r\n* fix","shortMessageHtmlLink":"update llama.cpp usage of llama3 (#10975)"}},{"before":"e3159c45e4a809f33c1b0e36bb1fa0f0fbb01b35","after":"697ca79ecad55f5496a83bec4ffdd94bde704cce","ref":"refs/heads/main","pushedAt":"2024-05-09T07:16:18.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"MeouSker77","name":"Yishuo Wang","path":"/MeouSker77","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34214113?s=80&v=4"},"commit":{"message":"use quantize kv and sdp in phi3-mini (#10973)","shortMessageHtmlLink":"use quantize kv and sdp in phi3-mini (#10973)"}},{"before":"459b764406286b3f451c99a2a70018fc6fba6659","after":"e3159c45e4a809f33c1b0e36bb1fa0f0fbb01b35","ref":"refs/heads/main","pushedAt":"2024-05-09T05:57:45.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"shane-huang","name":"Shengsheng Huang","path":"/shane-huang","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1995599?s=80&v=4"},"commit":{"message":"update private gpt quickstart and a small fix for dify (#10969)","shortMessageHtmlLink":"update private gpt quickstart and a small fix for dify (#10969)"}},{"before":"11df5f9773a6cf8348ead9b1c63844e9d90da8b8","after":"459b764406286b3f451c99a2a70018fc6fba6659","ref":"refs/heads/main","pushedAt":"2024-05-09T02:40:34.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"hzjane","name":"Wang, Jian4","path":"/hzjane","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/61138589?s=80&v=4"},"commit":{"message":"Remove munually_build_for_test push outside (#10968)","shortMessageHtmlLink":"Remove munually_build_for_test push outside (#10968)"}},{"before":"37820e1d8696694da08f8c68ea56829701f9fcd4","after":"11df5f9773a6cf8348ead9b1c63844e9d90da8b8","ref":"refs/heads/main","pushedAt":"2024-05-08T13:18:20.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"shane-huang","name":"Shengsheng Huang","path":"/shane-huang","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1995599?s=80&v=4"},"commit":{"message":"revise private GPT quickstart and a few fixes for other quickstart (#10967)","shortMessageHtmlLink":"revise private GPT quickstart and a few fixes for other quickstart (#…"}},{"before":"f4c615b1ee270cf07dd41a4fe8be3aef6b2719b5","after":"37820e1d8696694da08f8c68ea56829701f9fcd4","ref":"refs/heads/main","pushedAt":"2024-05-08T12:48:00.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"shane-huang","name":"Shengsheng Huang","path":"/shane-huang","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1995599?s=80&v=4"},"commit":{"message":"Add privateGPT quickstart (#10932)\n\n* Add privateGPT quickstart\r\n\r\n* Update privateGPT_quickstart.md\r\n\r\n* Update _toc.yml\r\n\r\n* Update _toc.yml\r\n\r\n---------\r\n\r\nCo-authored-by: Shengsheng Huang ","shortMessageHtmlLink":"Add privateGPT quickstart (#10932)"}},{"before":"7e7d969dcb1988a38cdf64fd9afedb3519728e1d","after":"f4c615b1ee270cf07dd41a4fe8be3aef6b2719b5","ref":"refs/heads/main","pushedAt":"2024-05-08T09:19:59.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"hzjane","name":"Wang, Jian4","path":"/hzjane","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/61138589?s=80&v=4"},"commit":{"message":"Add cohere example (#10954)\n\n* add link first\r\n\r\n* add_cpu_example\r\n\r\n* add GPU example","shortMessageHtmlLink":"Add cohere example (#10954)"}},{"before":"3209d6b0576595c528322cdfad5a218f52862946","after":"7e7d969dcb1988a38cdf64fd9afedb3519728e1d","ref":"refs/heads/main","pushedAt":"2024-05-08T09:12:50.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"glorysdj","name":"Dongjie Shi","path":"/glorysdj","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/2161573?s=80&v=4"},"commit":{"message":"a experimental for workflow abuse step1 fix a typo (#10965)\n\n* Update llm_unit_tests.yml\r\n\r\n* Update README.md\r\n\r\n* Update llm_unit_tests.yml\r\n\r\n* Update llm_unit_tests.yml","shortMessageHtmlLink":"a experimental for workflow abuse step1 fix a typo (#10965)"}},{"before":"02870dc3851c182af783ff234ac319c955d80457","after":"3209d6b0576595c528322cdfad5a218f52862946","ref":"refs/heads/main","pushedAt":"2024-05-08T09:09:48.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"hzjane","name":"Wang, Jian4","path":"/hzjane","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/61138589?s=80&v=4"},"commit":{"message":"Fix spculative llama3 no stop error (#10963)\n\n* fix normal\r\n\r\n* add eos_tokens_id on sp and add list if\r\n\r\n* update\r\n\r\n* no none","shortMessageHtmlLink":"Fix spculative llama3 no stop error (#10963)"}},{"before":"2ebec0395cc2d9fedd8b44bf5a839d0f77fe2246","after":"02870dc3851c182af783ff234ac319c955d80457","ref":"refs/heads/main","pushedAt":"2024-05-08T08:55:24.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"xiangyuT","name":"Xiangyu Tian","path":"/xiangyuT","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/109123695?s=80&v=4"},"commit":{"message":"LLM: Refine README of AutoTP-FastAPI example (#10960)","shortMessageHtmlLink":"LLM: Refine README of AutoTP-FastAPI example (#10960)"}},{"before":"dfa314727804954cdf570ee4250996f06f67562e","after":"2ebec0395cc2d9fedd8b44bf5a839d0f77fe2246","ref":"refs/heads/main","pushedAt":"2024-05-08T08:33:17.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"MeouSker77","name":"Yishuo Wang","path":"/MeouSker77","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34214113?s=80&v=4"},"commit":{"message":"optimize phi-3-mini-128 (#10959)","shortMessageHtmlLink":"optimize phi-3-mini-128 (#10959)"}},{"before":"5973d6c75338ee0cf2cc01ecaba6283fc6b4cf05","after":"dfa314727804954cdf570ee4250996f06f67562e","ref":"refs/heads/main","pushedAt":"2024-05-08T06:28:05.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"qiuxin2012","name":"Xin Qiu","path":"/qiuxin2012","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/4495653?s=80&v=4"},"commit":{"message":"update (#10944)","shortMessageHtmlLink":"update (#10944)"}},{"before":"15ee3fd54240ac6c9a5a543195e63005d49cce83","after":"5973d6c75338ee0cf2cc01ecaba6283fc6b4cf05","ref":"refs/heads/main","pushedAt":"2024-05-08T06:27:51.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"qiuxin2012","name":"Xin Qiu","path":"/qiuxin2012","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/4495653?s=80&v=4"},"commit":{"message":"make gemma's output better (#10943)","shortMessageHtmlLink":"make gemma's output better (#10943)"}},{"before":"0d6e12036f1bde76b2240ed55251f0e6d1e32373","after":"15ee3fd54240ac6c9a5a543195e63005d49cce83","ref":"refs/heads/main","pushedAt":"2024-05-08T06:16:43.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"Oscilloscope98","name":"Yuwen Hu","path":"/Oscilloscope98","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/54161268?s=80&v=4"},"commit":{"message":"Update igpu perf internlm (#10958)","shortMessageHtmlLink":"Update igpu perf internlm (#10958)"}},{"before":"164e6957af875a6c4b0f1c4eabfe726080d980c2","after":"0d6e12036f1bde76b2240ed55251f0e6d1e32373","ref":"refs/heads/main","pushedAt":"2024-05-08T02:46:20.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"leonardozcm","name":"Zhao Changmin","path":"/leonardozcm","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/26900100?s=80&v=4"},"commit":{"message":"Disable fast_init_ in load_low_bit (#10945)\n\n* fast_init_ disable","shortMessageHtmlLink":"Disable fast_init_ in load_low_bit (#10945)"}},{"before":"c801c37bc6ab12e6edd2c81293391c4355a7f1ba","after":"164e6957af875a6c4b0f1c4eabfe726080d980c2","ref":"refs/heads/main","pushedAt":"2024-05-08T01:34:03.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"qiyuangong","name":"Qiyuan Gong","path":"/qiyuangong","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/3848789?s=80&v=4"},"commit":{"message":"Refine axolotl quickstart (#10957)\n\n* Add default accelerate config for axolotl quickstart.\r\n* Fix requirement link.\r\n* Upgrade peft to 0.10.0 in requirement.","shortMessageHtmlLink":"Refine axolotl quickstart (#10957)"}},{"before":"aa2fa9fde18e424c96b3ce34d5337fd28c43c0eb","after":"c801c37bc6ab12e6edd2c81293391c4355a7f1ba","ref":"refs/heads/main","pushedAt":"2024-05-07T09:26:20.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"MeouSker77","name":"Yishuo Wang","path":"/MeouSker77","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34214113?s=80&v=4"},"commit":{"message":"optimize phi3 again: use quantize kv if possible (#10953)","shortMessageHtmlLink":"optimize phi3 again: use quantize kv if possible (#10953)"}},{"before":"c11170b96ff50a219008083aa224e58cd29ffe5d","after":"aa2fa9fde18e424c96b3ce34d5337fd28c43c0eb","ref":"refs/heads/main","pushedAt":"2024-05-07T07:53:09.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"MeouSker77","name":"Yishuo Wang","path":"/MeouSker77","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34214113?s=80&v=4"},"commit":{"message":"optimize phi3 again: use sdp if possible (#10951)","shortMessageHtmlLink":"optimize phi3 again: use sdp if possible (#10951)"}},{"before":"d7ca5d935b4b827946712bff208464fd1cc57c69","after":"c11170b96ff50a219008083aa224e58cd29ffe5d","ref":"refs/heads/main","pushedAt":"2024-05-07T07:12:26.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"qiyuangong","name":"Qiyuan Gong","path":"/qiyuangong","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/3848789?s=80&v=4"},"commit":{"message":"Upgrade Peft to 0.10.0 in finetune examples and docker (#10930)\n\n* Upgrade Peft to 0.10.0 in finetune examples.\r\n* Upgrade Peft to 0.10.0 in docker.","shortMessageHtmlLink":"Upgrade Peft to 0.10.0 in finetune examples and docker (#10930)"}},{"before":"0efe26c3b67c6c200fc25de5d044b72929210bb2","after":"d7ca5d935b4b827946712bff208464fd1cc57c69","ref":"refs/heads/main","pushedAt":"2024-05-07T07:09:14.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"qiyuangong","name":"Qiyuan Gong","path":"/qiyuangong","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/3848789?s=80&v=4"},"commit":{"message":"Upgrade Peft version to 0.10.0 for LLM finetune (#10886)\n\n* Upgrade Peft version to 0.10.0\r\n* Upgrade Peft version in ARC unit test and HF-Peft example.","shortMessageHtmlLink":"Upgrade Peft version to 0.10.0 for LLM finetune (#10886)"}},{"before":"245c7348bc8654d15dda67c24e4f6843625c93a7","after":"0efe26c3b67c6c200fc25de5d044b72929210bb2","ref":"refs/heads/main","pushedAt":"2024-05-07T05:48:39.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"Oscilloscope98","name":"Yuwen Hu","path":"/Oscilloscope98","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/54161268?s=80&v=4"},"commit":{"message":"Change order of chatglm2-6b and chatglm3-6b in iGPU perf test for more stable performance (#10948)","shortMessageHtmlLink":"Change order of chatglm2-6b and chatglm3-6b in iGPU perf test for mor…"}},{"before":"08ad40b2516a2e297324447ebc5b434e6a296a67","after":"245c7348bc8654d15dda67c24e4f6843625c93a7","ref":"refs/heads/main","pushedAt":"2024-05-07T05:35:42.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"Oscilloscope98","name":"Yuwen Hu","path":"/Oscilloscope98","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/54161268?s=80&v=4"},"commit":{"message":"Add codegemma example (#10884)\n\n* add codegemma example in GPU/HF-Transformers-AutoModels/\r\n\r\n* add README of codegemma example in GPU/HF-Transformers-AutoModels/\r\n\r\n* add codegemma example in GPU/PyTorch-Models/\r\n\r\n* add readme of codegemma example in GPU/PyTorch-Models/\r\n\r\n* add codegemma example in CPU/HF-Transformers-AutoModels/\r\n\r\n* add readme of codegemma example in CPU/HF-Transformers-AutoModels/\r\n\r\n* add codegemma example in CPU/PyTorch-Models/\r\n\r\n* add readme of codegemma example in CPU/PyTorch-Models/\r\n\r\n* fix typos\r\n\r\n* fix filename typo\r\n\r\n* add codegemma in tables\r\n\r\n* add comments of lm_head\r\n\r\n* remove comments of use_cache","shortMessageHtmlLink":"Add codegemma example (#10884)"}}],"hasNextPage":true,"hasPreviousPage":false,"activityType":"all","actor":null,"timePeriod":"all","sort":"DESC","perPage":30,"cursor":"djE6ks8AAAAERx9GMAA","startCursor":null,"endCursor":null}},"title":"Activity · intel-analytics/ipex-llm"}