Skip to content

Pull requests: InternLM/lmdeploy

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Sort

Pull requests list

[maca] add cudagraph support on maca backend.
#2834 opened Nov 29, 2024 by Reinerzhou Loading…
Update internvl chat template
#2832 opened Nov 29, 2024 by AllentDan Loading…
add openssh-server installation in dockerfile
#2830 opened Nov 28, 2024 by lvhan028 Loading…
profile throughput without new threads improvement
#2826 opened Nov 28, 2024 by grimoire Loading…
better kv allocate improvement
#2814 opened Nov 26, 2024 by grimoire Loading…
Refactor VLM modules improvement
#2810 opened Nov 25, 2024 by lvhan028 Loading…
[Feature] Support llava onevision enhancement New feature or request
#2783 opened Nov 21, 2024 by deepindeed2022 Loading…
[ascend]feat: support kv int8 enhancement New feature or request
#2736 opened Nov 11, 2024 by yao-fengchen Loading…
support qwen2-vl with turbomind backend enhancement New feature or request
#2720 opened Nov 6, 2024 by irexyc Loading…
update pre-commit config
#2683 opened Oct 30, 2024 by lvhan028 Loading…
support release pipeline improvement
#2581 opened Oct 11, 2024 by irexyc Loading…
Torchrun launching multiple api_server
#2402 opened Aug 30, 2024 by AllentDan Loading…
More w8a8 models
#2373 opened Aug 26, 2024 by AllentDan Draft
[Feature] support qqq(w4a8) for lmdeploy
#2274 opened Aug 9, 2024 by HandH1998 Loading…
6 tasks done
[Feature] Support XTuner Lite Llava enhancement New feature or request
#2191 opened Jul 31, 2024 by pppppM Loading…
Add prefix cache stats to usage
#2018 opened Jul 13, 2024 by ispobock Loading…
Add Jetson platform support (by docker)
#1820 opened Jun 21, 2024 by BestAnHongjun Loading…
support vl benchmark
#1662 opened May 27, 2024 by AllentDan Loading…
ProTip! Follow long discussions with comments:>50.