Skip to content

How to run the model? #82

Closed
Closed
@karpathy

Description

@karpathy

Hi, I believe there are docs missing on how to actually run the model once you download it? E.g. I followed the instructions and downloaded the 3.1 8B (base) model into the models/llama3_1/Meta-Llama-3.1-8B/ directory, but it's not clear what to do next. I'm guessing you'd want to load the params.json, init the ModelArgs with it, init the Transformer, load the params from consolidated.00.pth and torchrun that?

I'm guessing it would be along the lines of what exists in the llama3 repo (e.g. example_text_completion.py), which I am a bit hesitatant to build on given the notice about it being deprecated.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions