{ "config": { "model_dtype": "torch.float16", "model_name": "BrewInteractive/fikri-3.1-8B-Instruct" }, "results": { "MMLU": { "metric_name": 0.39170302447681726 }, "Truthful_qa": { "metric_name": 0.47062528701279294 }, "ARC": { "metric_name": 0.4061433447098976 }, "HellaSwag": { "metric_name": 0.45229761770351135 }, "GSM8K": { "metric_name": 0.10250569476082004 }, "GSM1K": { "metric_name": 0.10250569476082004 }, "Winogrande": { "metric_name": 0.5497630331753555 } } }