Skip to content

Pull requests: InternLM/lmdeploy

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Sort

Pull requests list

Fix torch_dtype in lite
#2956 opened Dec 25, 2024 by AllentDan Loading…
Bump version to v0.6.5
#2955 opened Dec 25, 2024 by lvhan028 Loading…
[maca] support deepseekv2 for maca backend. enhancement New feature or request
#2918 opened Dec 18, 2024 by Reinerzhou Draft
Remove threadsafe improvement
#2907 opened Dec 17, 2024 by grimoire Loading…
Support moe w8a8
#2894 opened Dec 13, 2024 by grimoire Loading…
[WIP]: use weights iterator while loading
#2886 opened Dec 12, 2024 by RunningLeon Loading…
support Turbomind ep enhancement New feature or request
#2883 opened Dec 12, 2024 by irexyc Loading…
Support Medusa speculative decoding enhancement New feature or request
#2859 opened Dec 5, 2024 by AllentDan Loading…
[Feature] Support llava onevision enhancement New feature or request
#2783 opened Nov 21, 2024 by deepindeed2022 Loading…
support qwen2-vl with turbomind backend enhancement New feature or request
#2720 opened Nov 6, 2024 by irexyc Loading…
update pre-commit config
#2683 opened Oct 30, 2024 by lvhan028 Loading…
support release pipeline improvement
#2581 opened Oct 11, 2024 by irexyc Loading…
[Feature] support qqq(w4a8) for lmdeploy
#2274 opened Aug 9, 2024 by HandH1998 Loading…
6 tasks done
[Feature] Support XTuner Lite Llava enhancement New feature or request
#2191 opened Jul 31, 2024 by pppppM Loading…
Add prefix cache stats to usage
#2018 opened Jul 13, 2024 by ispobock Loading…
ProTip! Follow long discussions with comments:>50.