--- license: llama3 base_model: meta-llama/Llama-4-Scout-17B-16E tags: - llama4 - moe - torchtitan - custom --- # Llama 4 Debug Model - Trained with TorchTitan Custom-trained Llama 4 model using TorchTitan framework. ## Model Details - **Training Framework**: TorchTitan - **Training Steps**: 10,000 - **Model Size**: ~220 MB - **Precision**: bfloat16 ## Usage from transformers import AutoModelForCausalLM, AutoTokenizer model = AutoModelForCausalLM.from_pretrained("lakhera2023/llama4-debugmodel-10k") tokenizer = AutoTokenizer.from_pretrained("lakhera2023/llama4-debugmodel-10k") prompt = "Once upon a time" inputs = tokenizer(prompt, return_tensors="pt") outputs = model.generate(**inputs, max_new_tokens=100) print(tokenizer.decode(outputs[0]))