Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The instructions to install Llama3 is horrible #1131

Open
Eyesun23 opened this issue Jun 21, 2024 · 2 comments
Open

The instructions to install Llama3 is horrible #1131

Eyesun23 opened this issue Jun 21, 2024 · 2 comments

Comments

@Eyesun23
Copy link

Eyesun23 commented Jun 21, 2024

I followed the steps of getting access to the models; I received a link. But I am getting this error after I ran:

torchrun --nproc_per_node=1 example_chat_completion.py \ --ckpt_dir Meta-Llama-3-8B-Instruct/ \ --tokenizer_path Meta-Llama-3-8B-Instruct/tokenizer.model \ --max_seq_len 512 --max_batch_size 6

/opt/anaconda3/envs/llama3/bin/python: can't open file 'example_chat_completion.py': [Errno 2] No such file or directory
[2024-06-21 16:30:35,995] torch.distributed.elastic.multiprocessing.api: [ERROR] failed (exitcode: 2) local_rank: 0 (pid: 97659) of binary: /opt/anaconda3/envs/llama3/bin/python
Traceback (most recent call last):
  File "/opt/anaconda3/envs/llama3/bin/torchrun", line 8, in <module>
    sys.exit(main())
  File "/opt/anaconda3/envs/llama3/lib/python3.8/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 347, in wrapper
    return f(*args, **kwargs)
  File "/opt/anaconda3/envs/llama3/lib/python3.8/site-packages/torch/distributed/run.py", line 812, in main
    run(args)
  File "/opt/anaconda3/envs/llama3/lib/python3.8/site-packages/torch/distributed/run.py", line 803, in run
    elastic_launch(
  File "/opt/anaconda3/envs/llama3/lib/python3.8/site-packages/torch/distributed/launcher/api.py", line 135, in __call__
    return launch_agent(self._config, self._entrypoint, list(args))
  File "/opt/anaconda3/envs/llama3/lib/python3.8/site-packages/torch/distributed/launcher/api.py", line 268, in launch_agent
    raise ChildFailedError(
torch.distributed.elastic.multiprocessing.errors.ChildFailedError: 
============================================================
example_chat_completion.py FAILED
------------------------------------------------------------
Failures:
  <NO_OTHER_FAILURES>
------------------------------------------------------------
Root Cause (first observed failure):
[0]:
  time      : 2024-06-21_16:30:35
  host      : aysuns-mbp.attlocal.net
  rank      : 0 (local_rank: 0)
  exitcode  : 2 (pid: 97659)
  error_file: <N/A>
  traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html```
@ENDGU
Copy link

ENDGU commented Jun 24, 2024

I suspect the issue lies in the first line: ‘’/opt/anaconda3/envs/llama3/bin/python: can't open file 'example_chat_completion.py': [Errno 2] No such file or directory‘’,you can check if there is any problem with the file path when you run this command.

@elnukakujo
Copy link

Hi, I also followed this README.md procedure using a python virtual environment and I also got stuck at the PYTHONPATH=$(git rev-parse --show-toplevel) torchrun llama_models/scripts/example_chat_completion.py $CHECKPOINT_DIR since there is no such llama_models/ folder created.

Additionally I tried to follow this other installation tutorial but I also got stuck because 403 Forbidden

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants