Closed
Description
Hi, I believe there are docs missing on how to actually run the model once you download it? E.g. I followed the instructions and downloaded the 3.1 8B (base) model into the models/llama3_1/Meta-Llama-3.1-8B/
directory, but it's not clear what to do next. I'm guessing you'd want to load the params.json
, init the ModelArgs
with it, init the Transformer
, load the params from consolidated.00.pth
and torchrun
that?
I'm guessing it would be along the lines of what exists in the llama3 repo (e.g. example_text_completion.py), which I am a bit hesitatant to build on given the notice about it being deprecated.
Metadata
Metadata
Assignees
Labels
No labels