Skip to content

Pull requests: AutoGPTQ/AutoGPTQ

Author
Filter by author
Label
Filter by label
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Milestones
Filter by milestone
Reviews
Assignee
Filter by who’s assigned
Sort

Pull requests list

Fix transformers 4.38.0 seq_len
#668 opened May 11, 2024 by randoentity Loading…
Support QBits kernel for CPU device
#660 opened Apr 29, 2024 by PenghuiCheng Loading…
[BUG/DEPRECATION] Remove fused attention/mlp
#659 opened Apr 28, 2024 by Qubitium Loading…
4 tasks done
[DEPRECATION] Remove triton v1
#658 opened Apr 28, 2024 by Qubitium Loading…
4 tasks done
[BUG] Fix H100 crash/compat with Marlin
#654 opened Apr 27, 2024 by Qubitium Loading…
3 tasks done
Extend support for Phi-3 models
#651 opened Apr 27, 2024 by davidgxue Loading…
[FEATURE] Allow loading of quantized lm_head
#648 opened Apr 25, 2024 by Qubitium Loading…
1 task done
Save sharded checkpoints
#645 opened Apr 19, 2024 by LaaZa Loading…
[PERFORMANCE] Fix Packing thread regression in code
#642 opened Apr 16, 2024 by Qubitium Loading…
2 tasks done
[BUG/FEATURE] Fix Sym=False, new checkpoint_format = gptq_v2
#640 opened Apr 12, 2024 by Qubitium Loading…
29 tasks done
Add support for DBRX
#623 opened Mar 28, 2024 by LaaZa Draft
Add support for InternLM-XComposer2-VL
#619 opened Mar 27, 2024 by LaaZa Loading…
Add Qwen2MoE
#593 opened Mar 14, 2024 by bozheng-hit Loading…
Format logging output in a form of a table
#569 opened Feb 28, 2024 by Andrei-Aksionov Loading…
Fix for fields being lists
#563 opened Feb 23, 2024 by stevelaskaridis Loading…
Support CPU only quantization.
#558 opened Feb 20, 2024 by miaojinc Loading…
Fix unnessary vram usage while injecting fused attn
#453 opened Nov 28, 2023 by lszxb Loading…
Save and Load sharded gptq checkpoint
#364 opened Oct 7, 2023 by PanQiWei Loading…
Fix g_idx in fused kernel
#305 opened Aug 30, 2023 by chu-tianxiang Loading…
ProTip! Add no:assignee to see everything that’s not assigned.