I've been exploring the implementation of models like Llama in Hugging Face’s transformers library, for example:
Hugging Face's Llama model implementation.
I’m curious about how these implementations work:
- Are the model codes in Hugging Face’s
transformerslibrary (e.g., for Llama) written directly by the creators of the models, or is it Hugging Face’s attempt to reproduce the functionality based on documentation or papers released by the model's original creators? - How closely do these implementations follow the specifications or official codebases (if available) provided by the model authors?