Ecosyste.ms: Timeline

Browse the timeline of events for every public repo on GitHub. Data updated hourly from GH Archive.

huggingface/text-generation-inference

mfuntowicz pushed 2 commits to trtllm-executor-thread huggingface/text-generation-inference
  • chore(trtllm): move dockerfile to right place c8d4721
  • chore(trtllm): disable tokenizer parallelism by default 7f02f49

View on GitHub

mfuntowicz pushed 2 commits to feat-backend-llamacpp huggingface/text-generation-inference

View on GitHub

xlg-go starred huggingface/text-generation-inference
mottoslo created a comment on an issue on huggingface/text-generation-inference
gentle ping @drbh is this issue being handled internally ? any feedback would be great !

View on GitHub

xxz-wow starred huggingface/text-generation-inference
kyc159357 starred huggingface/text-generation-inference
drbh pushed 1 commit to pr-2634-ci-branch huggingface/text-generation-inference
  • fix: adjust default when json tool choice is 193ad66

View on GitHub

remorses starred huggingface/text-generation-inference
kindero starred huggingface/text-generation-inference
Lawlers starred huggingface/text-generation-inference
SMAntony opened an issue on huggingface/text-generation-inference
Distributed Inference failing for Llama-3.1-70b-Instruct
### System Info text-generation-inference docker: sha-5e0fb46 (latest) OS: Ubuntu 22.04 Model: meta-llama/Llama-3.1-70B-Instruct GPU Used: 4 `nvidia-smi`: ``` +-----------------------------...
sywangyi closed a pull request on huggingface/text-generation-inference
add gptq and awq int4 support in intel platform
# What does this PR do? <!-- Congratulations! You've made it this far! You're not quite done yet though. Once merged, your PR is going to appear in the release notes with the title you set, ...
nimishbongale created a comment on an issue on huggingface/text-generation-inference
Same issue!

View on GitHub

RiantRedwood starred huggingface/text-generation-inference
KerwinKai starred huggingface/text-generation-inference
danieldk deleted a branch huggingface/text-generation-inference

maintenance/reciprocal-handling

danieldk pushed 1 commit to main huggingface/text-generation-inference
  • Make handling of FP8 scales more consisent (#2666) Change `fp8_quantize` so that we can pass around reciprocals ever... 5e0fb46

View on GitHub

danieldk created a review comment on a pull request on huggingface/text-generation-inference
Fixed.

View on GitHub

danieldk created a review on a pull request on huggingface/text-generation-inference

View on GitHub

SMAntony created a comment on an issue on huggingface/text-generation-inference
@danieldk could you please have a look?

View on GitHub

ShuteLee starred huggingface/text-generation-inference
JirenJin starred huggingface/text-generation-inference
sam-ulrich1 opened an issue on huggingface/text-generation-inference
Getting 2 different responses from the same HTTP call with seed set depending on what machine calls
### System Info tag:2.3.1 docker image running on nvidia 4090 on top of 20.04 Ubuntu ``` 2024-10-18T19:25:04.160854Z INFO text_generation_launcher: Args { model_id: "Qwen/Qwen2.5-Coder-1.5...
cswaney created a comment on an issue on huggingface/text-generation-inference
Solved. Apptainer automatically mounts `$HOME`, which happened to have `boto3` installed.

View on GitHub

cswaney closed an issue on huggingface/text-generation-inference
ModuleNotFoundError: No module named 'jmespath'
### System Info Image: `ghcr.io/huggingface/text-generation-inference:2.3.0` - Arch: `amd64` - Revision: `169178b937d0c4173b0fdcd6bf10a858cfe4f428` Model: `meta-llama/Llama-3.2-1B` - R...
cswaney opened an issue on huggingface/text-generation-inference
ModuleNotFoundError: No module named 'jmespath'
### System Info Image: `ghcr.io/huggingface/text-generation-inference:2.3.0` - Arch: `amd64` - Revision: `169178b937d0c4173b0fdcd6bf10a858cfe4f428` Model: `meta-llama/Llama-3.2-1B` - R...
drbh created a review comment on a pull request on huggingface/text-generation-inference
`let tools_to_use = match tool_choice` was updated to return a `vec![]` rather than a `Vec::with_capacity(0)`, and the conditional was moved below `tools_to_use` to return none in both cases (no to...

View on GitHub

drbh created a review on a pull request on huggingface/text-generation-inference

View on GitHub

drbh opened a pull request on huggingface/text-generation-inference
feat: return streaming errors as an event formatted for openai's client
This PR adds the functionality to emit errors as formatted events. Note the changes look large because of a new indent however the main logic that changed is the addition of ```rust Err(err...
Load more