affine-lora / training_metadata.json
affine
Upload folder using huggingface_hub
f78ae2e verified
{
"base_model": "openai/gpt-oss-120b",
"method": "8GPU_Model_Parallelism_SFTTrainer",
"quantization": "Mxfp4",
"model_parallelism": true,
"gpu_count": 8,
"lora_config": {
"r": 16,
"alpha": 32,
"target_modules": [
"q_proj",
"k_proj",
"v_proj",
"o_proj",
"gate_proj",
"up_proj",
"down_proj",
"router"
],
"optimized_for": "Project_Euler_Mathematical_Reasoning"
},
"training_config": {
"batch_size": 2,
"gradient_accumulation": 16,
"effective_batch_size": 32,
"learning_rate": 0.0001,
"max_length": 1024,
"epochs": 1
},
"dataset_type": "Project_Euler_Mathematical_Problems",
"training_date": "2025-08-26T14:45:11.908481",
"layer_distribution": "36_layers_across_8_H200_GPUs"
}