{"model_name": "olmo1124_13b_4k_finetune_epoch_2_7.5e-06", "model_type": "sft", "datasets": ["ai2-adapt-dev/oasst1_converted", "ai2-adapt-dev/flan_v2_converted", "ai2-adapt-dev/tulu_hard_coded_repeated_10", "ai2-adapt-dev/no_robots_converted", "ai2-adapt-dev/tulu_v3.9_wildchat_100k", "ai2-adapt-dev/personahub_math_v5_regen_149960", "allenai/tulu-3-sft-personas-math-grade", "ai2-adapt-dev/tulu_v3.9_open_math_2_gsm8k_50k", "ai2-adapt-dev/numinamath_tir_math_decontaminated", "ai2-adapt-dev/tulu_v3.9_personahub_math_interm_algebra_20k", "ai2-adapt-dev/personahub_code_v2_34999", "ai2-adapt-dev/evol_codealpaca_heval_decontaminated", "ai2-adapt-dev/personahub_ifdata_manual_seed_v3_29980", "ai2-adapt-dev/coconot_converted", "ai2-adapt-dev/tulu_v3.9_wildjailbreak_decontaminated_50k", "ai2-adapt-dev/tulu_v3.9_synthetic_finalresp_wildguardmixtrain_decontaminated_50k", "ai2-adapt-dev/tulu_v3.9_sciriff_10k", "ai2-adapt-dev/tulu_v3.9_table_gpt_5k", "ai2-adapt-dev/tulu_v3.9_aya_100k"], "base_model": "/weka/oe-training-default/ai2-llm/checkpoints/OLMo-medium/peteish13-anneal-from-596057-300B-legal-whammy-2-soup/step35773_olmo1124", "wandb_path": "https://wandb.ai/ai2-llm/open_instruct_internal/runs/qripf17k", "beaker_experiment": "https://beaker.org/ex/01JDE13EXAPVQ50BCA8A6Z4KCT/", "beaker_datasets": ["https://beaker.org/ds/01JDE13EYTJNZMWM8G2FM69BNM", "https://beaker.org/ds/01JDE13F3XXZBA0SRMV4WG7FAC", "https://beaker.org/ds/01JDE13F930PQ6XMJ8XYMGBA7M", "https://beaker.org/ds/01JDE13FE6SRK1C8E04QWRT8H9"]}