Skip to content
This repository has been archived by the owner on Dec 6, 2023. It is now read-only.

Added petals models #104

Merged
merged 4 commits into from
Sep 20, 2023
Merged

Conversation

filopedraz
Copy link
Collaborator

Description

Added Petals services

  • Stable Beluga 2
  • Code Llama 34B

Missing

  • Falcon 180B

@biswaroop1547
Copy link
Collaborator

LGTM, do we plan to add something on the daemon side to choose swarm or run it in petals swarm by default?

biswaroop1547
biswaroop1547 previously approved these changes Sep 19, 2023
@filopedraz
Copy link
Collaborator Author

We use by default always the public swarm (for now).

cht-petals/build.sh Outdated Show resolved Hide resolved
cht-petals/build.sh Outdated Show resolved Hide resolved
@casperdcl
Copy link
Contributor

Sorry about the force-pushes - was debugging this questionable GHA CI behaviour 🤯

@casperdcl casperdcl temporarily deployed to internal September 20, 2023 10:28 — with GitHub Actions Inactive
@casperdcl casperdcl temporarily deployed to internal September 20, 2023 10:31 — with GitHub Actions Inactive
@casperdcl casperdcl temporarily deployed to internal September 20, 2023 11:06 — with GitHub Actions Inactive
@casperdcl casperdcl temporarily deployed to internal September 20, 2023 11:11 — with GitHub Actions Inactive
@casperdcl
Copy link
Contributor

casperdcl commented Sep 20, 2023

Are premai-io/CodeLlama-34b-Instruct-hf and petals-team/StableBeluga2 only compatible with arm64 and not amd64?

Dockerfile:16
> RUN python3 download.py --model $MODEL_ID
Floating point exception (core dumped)

Or maybe it's a Petals bug.

@casperdcl casperdcl temporarily deployed to internal September 20, 2023 11:25 — with GitHub Actions Inactive
@casperdcl casperdcl temporarily deployed to internal September 20, 2023 11:30 — with GitHub Actions Inactive
TODO: revert when Petals fixes this
@casperdcl casperdcl temporarily deployed to internal September 20, 2023 11:33 — with GitHub Actions Inactive
@casperdcl casperdcl temporarily deployed to internal September 20, 2023 11:54 — with GitHub Actions Inactive
@filopedraz
Copy link
Collaborator Author

It's a Petals bug. I will investigate it further. Wanna have the arm images to be done so that I can create a PR to the registry and test the entire flow on Mac :)

Copy link
Contributor

@casperdcl casperdcl left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

lgtm

@filopedraz
Copy link
Collaborator Author

Falcon requires an HF token in order to fetch the weights due to the terms of services. I am investigating here how to upload the weights to Torrent and then download them with a CLI that we can use inside a dockerfile. :)

@casperdcl
Copy link
Contributor

casperdcl commented Sep 20, 2023

... or just add the HF token to GH repo secrets :)

@filopedraz filopedraz merged commit 0e3fdd4 into premAI-io:main Sep 20, 2023
5 checks passed
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
No open projects
Status: Done
Development

Successfully merging this pull request may close these issues.

3 participants