Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

launch TGI with the argument --max-input-tokens smaller than sliding_window=4096 (got here max_input_tokens=16384) #2730

Open
1 of 4 tasks
ashwincv0112 opened this issue Nov 7, 2024 · 0 comments

Comments

@ashwincv0112
Copy link

ashwincv0112 commented Nov 7, 2024

System Info

We are using a EC2 instance with T4 machine in AWS (g4dn.2xlarge) for deploying our fine-tuned model.

Information

  • Docker
  • The CLI directly

Tasks

  • An officially supported command
  • My own modifications

Reproduction

Please be informed that we are trying to deploy our fine-tuned version of starcoder2-3B model to an EC2 instance (T4 machine).

We are using the below command for deploying the model.

export volume=$PWD/data && docker run --gpus all -d -p 8080:80 -v $volume:/data -e USE_FLASH_ATTENTION=false ghcr.io/huggingface/text-generation-inference:2.3.0 --model-id=/data/ --max-input-tokens 16384 --max-total-tokens 26048

the deployment is failing with the below error message:

ValueError: The backend cuda does not support sliding window attention that is used by the model type starcoder2. To use this model nonetheless with the cuda backend, please launch TGI with the argument `--max-input-tokens` smaller than sliding_window=4096 (got here max_input_tokens=16384).

Could you please let us know if there is any way we can change the max_input_tokens parameter value while deploying the model.

Thanks.

Expected behavior

We would be able to change the --max-input-tokens parameter value to 16384 while deploying the model.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant