Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

TGI Server should be installable via pip #2696

Open
cdoern opened this issue Oct 27, 2024 · 0 comments
Open

TGI Server should be installable via pip #2696

cdoern opened this issue Oct 27, 2024 · 0 comments

Comments

@cdoern
Copy link

cdoern commented Oct 27, 2024

Feature request

pip install text-generation-inference pip install text-generation-launcher. Having these as published packages will allow projects to use the serving and client functionality.

I am hitting a wall here, and currently am trying to clone TGI and use the serving functionality. This is proving to be pretty hard because the way the project is built and more focused on being run in a containerized application.

Motivation

Having a local LLM server that can handle batched requests would be amazing. llama-cpp-python is the only comparable embedded llm serving tool and it isn't built to really handle concurrent request like TGI is.

We are looking to use this in instructlab as a serving backend option if published to pypi or even locally installable via pip.

Your contribution

I could add the pyproject.toml and setup.py with some help around what parts of the project we'd want built into a python package!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant