Ecosyste.ms: Timeline
Browse the timeline of events for every public repo on GitHub. Data updated hourly from GH Archive.
sixsixcoder pushed 413 commits to cogagent sixsixcoder/vllm
- [Model] Support `is_causal` HF config field for Qwen2 model (#10621) Signed-off-by: DarkLight1337 <tlleungac@connect... ed46f14
- Super tiny little typo fix (#10633) 2b0879b
- [Bug]: Authorization ignored when root_path is set (#10606) Signed-off-by: chaunceyjiang <[email protected]> d04b13a
- [Bugfix] Fix chunked prefill with model dtype float32 on Turing Devices (#9850) Signed-off-by: Wallas Santos <wallas... c27df94
- [Docs] Add Snowflake Slides (#10641) Signed-off-by: simon-mo <[email protected]> 452a4e8
- [Model]: Add support for Aria model (#10514) Signed-off-by: xffxff <[email protected]> Co-authored-by: Isotr0py <203... b1d9205
- [Model] Enable optional prefix when loading embedding models (#10639) Signed-off-by: DarkLight1337 <tlleungac@connec... cf73f0c
- [Doc] Fix typos in docs (#10636) Signed-off-by: DarkLight1337 <[email protected]> 1b583cf
- [Model] Add OLMo November 2024 model (#10503) 9db713a
- [misc] do not read HOST_IP (#10644) Signed-off-by: youkaichao <[email protected]> 6e9ff05
- [bugfix] fix aria model and add torch.compile (#10645) Signed-off-by: youkaichao <[email protected]> 45ac4ff
- [Feature] vLLM ARM Enablement for AARCH64 CPUs (#9228) Signed-off-by: Sanket Kale <[email protected]> Co-aut... a6760f6
- [v1] EngineArgs for better config handling for v1 (#10382) Signed-off-by: rickyx <[email protected]> 519e8e4
- custom allreduce + torch.compile (#10121) Signed-off-by: youkaichao <[email protected]> Co-authored-by: youkaich... 9a88f89
- [Misc] Remove outdated init protocols (#10655) Signed-off-by: DarkLight1337 <[email protected]> 9406353
- [ci] add vllm_test_utils (#10659) Signed-off-by: youkaichao <[email protected]> 334d64d
- [V1] Enable profile for LLMEngine (#10665) 1f6584e
- [Bugfix] Fix for Spec model TP + Chunked Prefill (#10232) Signed-off-by: andoorve <[email protected]... db66e01
- [Hardware][NVIDIA] Add non-NVML CUDA mode for Jetson (#9735) Signed-off-by: Conroy Cheers <[email protected]> f5792c7
- [Bugfix] Fix using `-O[0,3]` with LLM entrypoint (#10677) Signed-off-by: mgoin <[email protected]> 9a99273
- and 393 more ...
sixsixcoder created a branch on sixsixcoder/vllm
cogagent - A high-throughput and memory-efficient inference and serving engine for LLMs