Skip to content
GitHub
View on GitHub

Llama2-7B

API reference for Llama2_7B

from modal_training_gym.common.models.llama2_7b import Llama2_7B

Llama 2 7B from Meta.

Inherits from: HFModelConfiguration, ModelConfiguration

FieldTypeDefaultDescription
model_namestr"meta-llama/Llama-2-7b-hf"HuggingFace repo ID or other model identifier. Default "".
model_path`strNone`None
architecture`ModelArchitectureNone`None
training`ModelTrainingConfigNone`ModelTrainingConfig(gpu_type='H100', n_nodes=1, tensor_model_parallel_size=1, pipeline_model_parallel_size=1, context_parallel_size=1, sequence_parallel=False, expert_model_parallel_size=1, moe_permute_fusion=False, moe_grouped_gemm=False, moe_shared_expert_overlap=False, moe_aux_loss_coeff=0.0, lora_rank=8, lora_alpha=16, target_modules='all-linear', merge_lora=False)

Download or materialize weights into the model volume.

Source: modal_training_gym/common/models/llama2_7b.py