InterpBench / benchmark_metadata.json
cybershiptrooper's picture
Update metadata (#5)
c0a69c0 verified
raw
history blame
61.2 kB
{
"name": "InterpBench",
"version": "1.0.0",
"description": "A benchmark of transformers with known circuits for evaluating mechanistic interpretability techniques.",
"license": "https://creativecommons.org/licenses/by/4.0/",
"url": "https://huggingface.co/cybershiptrooper/InterpBench",
"cases": [
{
"case_id": "11",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/11",
"task_description": "Counts the number of words in a sequence based on their length.",
"vocab": [
"J",
"LB",
"TPSI",
"V",
"b",
"no",
"oCLrZaW",
"poiVg"
],
"max_seq_len": 10,
"min_seq_len": 4,
"files": [
{
"file_name": "edges.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/edges.pkl"
},
{
"file_name": "ll_model.pth",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/ll_model.pth"
},
{
"file_name": "ll_model_cfg.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/ll_model_cfg.pkl"
},
{
"file_name": "meta.json",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/meta.json"
}
],
"transformer_cfg": {
"n_layers": 2,
"d_model": 12,
"n_ctx": 10,
"d_head": 3,
"model_name": "custom",
"n_heads": 4,
"d_mlp": 48,
"act_fn": "gelu",
"d_vocab": 10,
"eps": 1e-05,
"use_attn_result": true,
"use_attn_scale": true,
"use_split_qkv_input": true,
"use_hook_mlp_in": true,
"use_attn_in": false,
"use_local_attn": false,
"original_architecture": null,
"from_checkpoint": false,
"checkpoint_index": null,
"checkpoint_label_type": null,
"checkpoint_value": null,
"tokenizer_name": null,
"window_size": null,
"attn_types": null,
"init_mode": "gpt2",
"normalization_type": null,
"n_devices": 1,
"attention_dir": "causal",
"attn_only": false,
"seed": 0,
"initializer_range": 0.1460593486680443,
"init_weights": true,
"scale_attn_by_inverse_layer_idx": false,
"positional_embedding_type": "standard",
"final_rms": false,
"d_vocab_out": 5,
"parallel_attn_mlp": false,
"rotary_dim": null,
"n_params": 3456,
"use_hook_tokens": false,
"gated_mlp": false,
"default_prepend_bos": true,
"dtype": "torch.float32",
"tokenizer_prepends_bos": null,
"n_key_value_heads": null,
"post_embedding_ln": false,
"rotary_base": 10000,
"trust_remote_code": false,
"rotary_adjacent_pairs": false
},
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/ll_model_cfg.pkl",
"training_args": {
"atol": 0.05,
"lr": 0.01,
"use_single_loss": false,
"iit_weight": 1.0,
"behavior_weight": 1.0,
"strict_weight": 0.4,
"epochs": 500,
"act_fn": "gelu",
"clip_grad_norm": 1.0,
"lr_scheduler": ""
},
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/meta.json",
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/ll_model.pth",
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/edges.pkl"
},
{
"case_id": "13",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/13",
"task_description": "Analyzes the trend (increasing, decreasing, constant) of numeric tokens.",
"vocab": [
0,
1,
2
],
"max_seq_len": 10,
"min_seq_len": 4,
"files": [
{
"file_name": "edges.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/edges.pkl"
},
{
"file_name": "ll_model.pth",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/ll_model.pth"
},
{
"file_name": "ll_model_cfg.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/ll_model_cfg.pkl"
},
{
"file_name": "meta.json",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/meta.json"
}
],
"transformer_cfg": {
"n_layers": 2,
"d_model": 20,
"n_ctx": 10,
"d_head": 5,
"model_name": "custom",
"n_heads": 4,
"d_mlp": 80,
"act_fn": "gelu",
"d_vocab": 5,
"eps": 1e-05,
"use_attn_result": true,
"use_attn_scale": true,
"use_split_qkv_input": true,
"use_hook_mlp_in": true,
"use_attn_in": false,
"use_local_attn": false,
"original_architecture": null,
"from_checkpoint": false,
"checkpoint_index": null,
"checkpoint_label_type": null,
"checkpoint_value": null,
"tokenizer_name": null,
"window_size": null,
"attn_types": null,
"init_mode": "gpt2",
"normalization_type": null,
"n_devices": 1,
"attention_dir": "bidirectional",
"attn_only": false,
"seed": 0,
"initializer_range": 0.1460593486680443,
"init_weights": true,
"scale_attn_by_inverse_layer_idx": false,
"positional_embedding_type": "standard",
"final_rms": false,
"d_vocab_out": 3,
"parallel_attn_mlp": false,
"rotary_dim": null,
"n_params": 9600,
"use_hook_tokens": false,
"gated_mlp": false,
"default_prepend_bos": true,
"dtype": "torch.float32",
"tokenizer_prepends_bos": null,
"n_key_value_heads": null,
"post_embedding_ln": false,
"rotary_base": 10000,
"trust_remote_code": false,
"rotary_adjacent_pairs": false
},
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/ll_model_cfg.pkl",
"training_args": {
"atol": 0.05,
"lr": 0.01,
"use_single_loss": false,
"iit_weight": 1.0,
"behavior_weight": 1.0,
"strict_weight": 0.4,
"epochs": 500,
"act_fn": "gelu",
"clip_grad_norm": 1.0,
"lr_scheduler": ""
},
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/meta.json",
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/ll_model.pth",
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/edges.pkl"
},
{
"case_id": "18",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/18",
"task_description": "Classify each token based on its frequency as 'rare', 'common', or 'frequent'.",
"vocab": [
"a",
"b",
"c",
"d",
"e"
],
"max_seq_len": 10,
"min_seq_len": 4,
"files": [
{
"file_name": "edges.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/edges.pkl"
},
{
"file_name": "ll_model.pth",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/ll_model.pth"
},
{
"file_name": "ll_model_cfg.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/ll_model_cfg.pkl"
},
{
"file_name": "meta.json",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/meta.json"
}
],
"transformer_cfg": {
"n_layers": 2,
"d_model": 26,
"n_ctx": 10,
"d_head": 6,
"model_name": "custom",
"n_heads": 4,
"d_mlp": 104,
"act_fn": "gelu",
"d_vocab": 7,
"eps": 1e-05,
"use_attn_result": true,
"use_attn_scale": true,
"use_split_qkv_input": true,
"use_hook_mlp_in": true,
"use_attn_in": false,
"use_local_attn": false,
"original_architecture": null,
"from_checkpoint": false,
"checkpoint_index": null,
"checkpoint_label_type": null,
"checkpoint_value": null,
"tokenizer_name": null,
"window_size": null,
"attn_types": null,
"init_mode": "gpt2",
"normalization_type": null,
"n_devices": 1,
"attention_dir": "bidirectional",
"attn_only": false,
"seed": 0,
"initializer_range": 0.12344267996967354,
"init_weights": true,
"scale_attn_by_inverse_layer_idx": false,
"positional_embedding_type": "standard",
"final_rms": false,
"d_vocab_out": 3,
"parallel_attn_mlp": false,
"rotary_dim": null,
"n_params": 15808,
"use_hook_tokens": false,
"gated_mlp": false,
"default_prepend_bos": true,
"dtype": "torch.float32",
"tokenizer_prepends_bos": null,
"n_key_value_heads": null,
"post_embedding_ln": false,
"rotary_base": 10000,
"trust_remote_code": false,
"rotary_adjacent_pairs": false
},
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/ll_model_cfg.pkl",
"training_args": {
"atol": 0.05,
"lr": 0.001,
"use_single_loss": false,
"iit_weight": 1.0,
"behavior_weight": 1.0,
"strict_weight": 1.0,
"epochs": 2000,
"act_fn": "gelu",
"clip_grad_norm": 0.1,
"lr_scheduler": "",
"model_pair": "strict"
},
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/meta.json",
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/ll_model.pth",
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/edges.pkl"
},
{
"case_id": "19",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/19",
"task_description": "Removes consecutive duplicate tokens from a sequence.",
"vocab": [
"a",
"b",
"c"
],
"max_seq_len": 15,
"min_seq_len": 4,
"files": [
{
"file_name": "edges.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/edges.pkl"
},
{
"file_name": "ll_model.pth",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/ll_model.pth"
},
{
"file_name": "ll_model_cfg.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/ll_model_cfg.pkl"
},
{
"file_name": "meta.json",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/meta.json"
}
],
"transformer_cfg": {
"n_layers": 2,
"d_model": 32,
"n_ctx": 15,
"d_head": 8,
"model_name": "custom",
"n_heads": 4,
"d_mlp": 128,
"act_fn": "gelu",
"d_vocab": 5,
"eps": 1e-05,
"use_attn_result": true,
"use_attn_scale": true,
"use_split_qkv_input": true,
"use_hook_mlp_in": true,
"use_attn_in": false,
"use_local_attn": false,
"original_architecture": null,
"from_checkpoint": false,
"checkpoint_index": null,
"checkpoint_label_type": null,
"checkpoint_value": null,
"tokenizer_name": null,
"window_size": null,
"attn_types": null,
"init_mode": "gpt2",
"normalization_type": null,
"n_devices": 1,
"attention_dir": "causal",
"attn_only": false,
"seed": 0,
"initializer_range": 0.15689290811054724,
"init_weights": true,
"scale_attn_by_inverse_layer_idx": false,
"positional_embedding_type": "standard",
"final_rms": false,
"d_vocab_out": 3,
"parallel_attn_mlp": false,
"rotary_dim": null,
"n_params": 24576,
"use_hook_tokens": false,
"gated_mlp": false,
"default_prepend_bos": true,
"dtype": "torch.float32",
"tokenizer_prepends_bos": null,
"n_key_value_heads": null,
"post_embedding_ln": false,
"rotary_base": 10000,
"trust_remote_code": false,
"rotary_adjacent_pairs": false
},
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/ll_model_cfg.pkl",
"training_args": {
"atol": 0.05,
"lr": 0.001,
"use_single_loss": false,
"iit_weight": 1.0,
"behavior_weight": 1.0,
"strict_weight": 0.4,
"epochs": 2000,
"act_fn": "gelu",
"clip_grad_norm": 0.1,
"lr_scheduler": ""
},
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/meta.json",
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/ll_model.pth",
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/edges.pkl"
},
{
"case_id": "20",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/20",
"task_description": "Detect spam messages based on appearance of spam keywords.",
"vocab": [
"J",
"LB",
"TPSI",
"V",
"b",
"click",
"no",
"now",
"oCLrZaW",
"offer",
"poiVg",
"spam"
],
"max_seq_len": 10,
"min_seq_len": 4,
"files": [
{
"file_name": "edges.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/edges.pkl"
},
{
"file_name": "ll_model.pth",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/ll_model.pth"
},
{
"file_name": "ll_model_cfg.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/ll_model_cfg.pkl"
},
{
"file_name": "meta.json",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/meta.json"
}
],
"transformer_cfg": {
"n_layers": 2,
"d_model": 13,
"n_ctx": 10,
"d_head": 3,
"model_name": "custom",
"n_heads": 4,
"d_mlp": 52,
"act_fn": "gelu",
"d_vocab": 14,
"eps": 1e-05,
"use_attn_result": true,
"use_attn_scale": true,
"use_split_qkv_input": true,
"use_hook_mlp_in": true,
"use_attn_in": false,
"use_local_attn": false,
"original_architecture": null,
"from_checkpoint": false,
"checkpoint_index": null,
"checkpoint_label_type": null,
"checkpoint_value": null,
"tokenizer_name": null,
"window_size": null,
"attn_types": null,
"init_mode": "gpt2",
"normalization_type": null,
"n_devices": 1,
"attention_dir": "causal",
"attn_only": false,
"seed": 0,
"initializer_range": 0.16,
"init_weights": true,
"scale_attn_by_inverse_layer_idx": false,
"positional_embedding_type": "standard",
"final_rms": false,
"d_vocab_out": 2,
"parallel_attn_mlp": false,
"rotary_dim": null,
"n_params": 3952,
"use_hook_tokens": false,
"gated_mlp": false,
"default_prepend_bos": true,
"dtype": "torch.float32",
"tokenizer_prepends_bos": null,
"n_key_value_heads": null,
"post_embedding_ln": false,
"rotary_base": 10000,
"trust_remote_code": false,
"rotary_adjacent_pairs": false
},
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/ll_model_cfg.pkl",
"training_args": {
"atol": 0.05,
"lr": 0.01,
"use_single_loss": false,
"iit_weight": 1.0,
"behavior_weight": 1.0,
"strict_weight": 0.4,
"epochs": 2000,
"act_fn": "gelu",
"clip_grad_norm": 0.1,
"lr_scheduler": "",
"model_pair": "strict"
},
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/meta.json",
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/ll_model.pth",
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/edges.pkl"
},
{
"case_id": "21",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/21",
"task_description": "Extract unique tokens from a string",
"vocab": [
"a",
"b",
"c"
],
"max_seq_len": 10,
"min_seq_len": 4,
"files": [
{
"file_name": "edges.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/edges.pkl"
},
{
"file_name": "ll_model.pth",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/ll_model.pth"
},
{
"file_name": "ll_model_cfg.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/ll_model_cfg.pkl"
},
{
"file_name": "meta.json",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/meta.json"
}
],
"transformer_cfg": {
"n_layers": 4,
"d_model": 50,
"n_ctx": 10,
"d_head": 12,
"model_name": "custom",
"n_heads": 4,
"d_mlp": 200,
"act_fn": "gelu",
"d_vocab": 5,
"eps": 1e-05,
"use_attn_result": true,
"use_attn_scale": true,
"use_split_qkv_input": true,
"use_hook_mlp_in": true,
"use_attn_in": false,
"use_local_attn": false,
"original_architecture": null,
"from_checkpoint": false,
"checkpoint_index": null,
"checkpoint_label_type": null,
"checkpoint_value": null,
"tokenizer_name": null,
"window_size": null,
"attn_types": null,
"init_mode": "gpt2",
"normalization_type": null,
"n_devices": 1,
"attention_dir": "causal",
"attn_only": false,
"seed": 0,
"initializer_range": 0.09847319278346618,
"init_weights": true,
"scale_attn_by_inverse_layer_idx": false,
"positional_embedding_type": "standard",
"final_rms": false,
"d_vocab_out": 3,
"parallel_attn_mlp": false,
"rotary_dim": null,
"n_params": 118400,
"use_hook_tokens": false,
"gated_mlp": false,
"default_prepend_bos": true,
"dtype": "torch.float32",
"tokenizer_prepends_bos": null,
"n_key_value_heads": null,
"post_embedding_ln": false,
"rotary_base": 10000,
"trust_remote_code": false,
"rotary_adjacent_pairs": false
},
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/ll_model_cfg.pkl",
"training_args": {
"atol": 0.05,
"lr": 0.0005,
"use_single_loss": false,
"iit_weight": 1.0,
"behavior_weight": 1.0,
"strict_weight": 0.5,
"epochs": 2000,
"act_fn": "gelu",
"clip_grad_norm": 0.1,
"lr_scheduler": "",
"model_pair": "strict"
},
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/meta.json",
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/ll_model.pth",
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/edges.pkl"
},
{
"case_id": "26",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/26",
"task_description": "Creates a cascading effect by repeating each token in sequence incrementally.",
"vocab": [
"a",
"b",
"c"
],
"max_seq_len": 10,
"min_seq_len": 4,
"files": [
{
"file_name": "edges.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/edges.pkl"
},
{
"file_name": "ll_model.pth",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/ll_model.pth"
},
{
"file_name": "ll_model_cfg.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/ll_model_cfg.pkl"
},
{
"file_name": "meta.json",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/meta.json"
}
],
"transformer_cfg": {
"n_layers": 2,
"d_model": 21,
"n_ctx": 10,
"d_head": 5,
"model_name": "custom",
"n_heads": 4,
"d_mlp": 84,
"act_fn": "gelu",
"d_vocab": 5,
"eps": 1e-05,
"use_attn_result": true,
"use_attn_scale": true,
"use_split_qkv_input": true,
"use_hook_mlp_in": true,
"use_attn_in": false,
"use_local_attn": false,
"original_architecture": null,
"from_checkpoint": false,
"checkpoint_index": null,
"checkpoint_label_type": null,
"checkpoint_value": null,
"tokenizer_name": null,
"window_size": null,
"attn_types": null,
"init_mode": "gpt2",
"normalization_type": null,
"n_devices": 1,
"attention_dir": "causal",
"attn_only": false,
"seed": 0,
"initializer_range": 0.12344267996967354,
"init_weights": true,
"scale_attn_by_inverse_layer_idx": false,
"positional_embedding_type": "standard",
"final_rms": false,
"d_vocab_out": 27,
"parallel_attn_mlp": false,
"rotary_dim": null,
"n_params": 10416,
"use_hook_tokens": false,
"gated_mlp": false,
"default_prepend_bos": true,
"dtype": "torch.float32",
"tokenizer_prepends_bos": null,
"n_key_value_heads": null,
"post_embedding_ln": false,
"rotary_base": 10000,
"trust_remote_code": false,
"rotary_adjacent_pairs": false
},
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/ll_model_cfg.pkl",
"training_args": {
"atol": 0.05,
"lr": 0.01,
"use_single_loss": false,
"iit_weight": 1.0,
"behavior_weight": 1.0,
"strict_weight": 0.4,
"epochs": 2000,
"act_fn": "gelu",
"clip_grad_norm": 0.1,
"lr_scheduler": "",
"model_pair": "strict"
},
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/meta.json",
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/ll_model.pth",
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/edges.pkl"
},
{
"case_id": "29",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/29",
"task_description": "Creates abbreviations for each token in the sequence.",
"vocab": [
"J",
"LB",
"TPSI",
"V",
"b",
"no",
"oCLrZaW",
"poiVg"
],
"max_seq_len": 10,
"min_seq_len": 4,
"files": [
{
"file_name": "edges.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/edges.pkl"
},
{
"file_name": "ll_model.pth",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/ll_model.pth"
},
{
"file_name": "ll_model_cfg.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/ll_model_cfg.pkl"
},
{
"file_name": "meta.json",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/meta.json"
}
],
"transformer_cfg": {
"n_layers": 2,
"d_model": 13,
"n_ctx": 10,
"d_head": 3,
"model_name": "custom",
"n_heads": 4,
"d_mlp": 52,
"act_fn": "gelu",
"d_vocab": 10,
"eps": 1e-05,
"use_attn_result": true,
"use_attn_scale": true,
"use_split_qkv_input": true,
"use_hook_mlp_in": true,
"use_attn_in": false,
"use_local_attn": false,
"original_architecture": null,
"from_checkpoint": false,
"checkpoint_index": null,
"checkpoint_label_type": null,
"checkpoint_value": null,
"tokenizer_name": null,
"window_size": null,
"attn_types": null,
"init_mode": "gpt2",
"normalization_type": null,
"n_devices": 1,
"attention_dir": "causal",
"attn_only": false,
"seed": 0,
"initializer_range": 0.1539600717839002,
"init_weights": true,
"scale_attn_by_inverse_layer_idx": false,
"positional_embedding_type": "standard",
"final_rms": false,
"d_vocab_out": 8,
"parallel_attn_mlp": false,
"rotary_dim": null,
"n_params": 3952,
"use_hook_tokens": false,
"gated_mlp": false,
"default_prepend_bos": true,
"dtype": "torch.float32",
"tokenizer_prepends_bos": null,
"n_key_value_heads": null,
"post_embedding_ln": false,
"rotary_base": 10000,
"trust_remote_code": false,
"rotary_adjacent_pairs": false
},
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/ll_model_cfg.pkl",
"training_args": {
"atol": 0.05,
"lr": 0.01,
"use_single_loss": false,
"iit_weight": 1.0,
"behavior_weight": 1.0,
"strict_weight": 0.4,
"epochs": 2000,
"act_fn": "gelu",
"clip_grad_norm": 0.1,
"lr_scheduler": "",
"model_pair": "strict"
},
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/meta.json",
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/ll_model.pth",
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/edges.pkl"
},
{
"case_id": "3",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/3",
"task_description": "Returns the fraction of 'x' in the input up to the i-th position for all i.",
"vocab": [
"a",
"b",
"c",
"x"
],
"max_seq_len": 5,
"min_seq_len": 4,
"files": [
{
"file_name": "edges.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/edges.pkl"
},
{
"file_name": "ll_model.pth",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/ll_model.pth"
},
{
"file_name": "ll_model_cfg.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/ll_model_cfg.pkl"
},
{
"file_name": "meta.json",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/meta.json"
}
],
"transformer_cfg": {
"n_layers": 2,
"d_model": 12,
"n_ctx": 5,
"d_head": 3,
"model_name": "custom",
"n_heads": 4,
"d_mlp": 48,
"act_fn": "gelu",
"d_vocab": 6,
"eps": 1e-05,
"use_attn_result": true,
"use_attn_scale": true,
"use_split_qkv_input": true,
"use_hook_mlp_in": true,
"use_attn_in": false,
"use_local_attn": false,
"original_architecture": null,
"from_checkpoint": false,
"checkpoint_index": null,
"checkpoint_label_type": null,
"checkpoint_value": null,
"tokenizer_name": null,
"window_size": null,
"attn_types": null,
"init_mode": "gpt2",
"normalization_type": null,
"n_devices": 1,
"attention_dir": "causal",
"attn_only": false,
"seed": 0,
"initializer_range": 0.22188007849009167,
"init_weights": true,
"scale_attn_by_inverse_layer_idx": false,
"positional_embedding_type": "standard",
"final_rms": false,
"d_vocab_out": 1,
"parallel_attn_mlp": false,
"rotary_dim": null,
"n_params": 3456,
"use_hook_tokens": false,
"gated_mlp": false,
"default_prepend_bos": true,
"dtype": "torch.float32",
"tokenizer_prepends_bos": null,
"n_key_value_heads": null,
"post_embedding_ln": false,
"rotary_base": 10000,
"trust_remote_code": false,
"rotary_adjacent_pairs": false
},
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/ll_model_cfg.pkl",
"training_args": {
"atol": 0.05,
"lr": 0.001,
"use_single_loss": false,
"iit_weight": 1.0,
"behavior_weight": 1.0,
"strict_weight": 10.0,
"epochs": 2000,
"act_fn": "gelu",
"clip_grad_norm": 0.1,
"lr_scheduler": ""
},
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/meta.json",
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/ll_model.pth",
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/edges.pkl"
},
{
"case_id": "33",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/33",
"task_description": "Checks if each token's length is odd or even.",
"vocab": [
"J",
"LB",
"TPSI",
"V",
"b",
"no",
"oCLrZaW",
"poiVg"
],
"max_seq_len": 10,
"min_seq_len": 4,
"files": [
{
"file_name": "edges.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/edges.pkl"
},
{
"file_name": "ll_model.pth",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/ll_model.pth"
},
{
"file_name": "ll_model_cfg.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/ll_model_cfg.pkl"
},
{
"file_name": "meta.json",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/meta.json"
}
],
"transformer_cfg": {
"n_layers": 2,
"d_model": 4,
"n_ctx": 10,
"d_head": 1,
"model_name": "custom",
"n_heads": 4,
"d_mlp": 16,
"act_fn": "gelu",
"d_vocab": 10,
"eps": 1e-05,
"use_attn_result": true,
"use_attn_scale": true,
"use_split_qkv_input": true,
"use_hook_mlp_in": true,
"use_attn_in": false,
"use_local_attn": false,
"original_architecture": null,
"from_checkpoint": false,
"checkpoint_index": null,
"checkpoint_label_type": null,
"checkpoint_value": null,
"tokenizer_name": null,
"window_size": null,
"attn_types": null,
"init_mode": "gpt2",
"normalization_type": null,
"n_devices": 1,
"attention_dir": "causal",
"attn_only": false,
"seed": 0,
"initializer_range": 0.17457431218879393,
"init_weights": true,
"scale_attn_by_inverse_layer_idx": false,
"positional_embedding_type": "standard",
"final_rms": false,
"d_vocab_out": 2,
"parallel_attn_mlp": false,
"rotary_dim": null,
"n_params": 384,
"use_hook_tokens": false,
"gated_mlp": false,
"default_prepend_bos": true,
"dtype": "torch.float32",
"tokenizer_prepends_bos": null,
"n_key_value_heads": null,
"post_embedding_ln": false,
"rotary_base": 10000,
"trust_remote_code": false,
"rotary_adjacent_pairs": false
},
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/ll_model_cfg.pkl",
"training_args": {
"atol": 0.05,
"lr": 0.001,
"use_single_loss": false,
"iit_weight": 1.0,
"behavior_weight": 1.0,
"strict_weight": 0.4,
"epochs": 2000,
"act_fn": "gelu",
"clip_grad_norm": 0.1,
"lr_scheduler": ""
},
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/meta.json",
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/ll_model.pth",
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/edges.pkl"
},
{
"case_id": "34",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/34",
"task_description": "Calculate the ratio of vowels to consonants in each word.",
"vocab": [
"J",
"LB",
"TPSI",
"V",
"b",
"no",
"oCLrZaW",
"poiVg"
],
"max_seq_len": 10,
"min_seq_len": 4,
"files": [
{
"file_name": "edges.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/edges.pkl"
},
{
"file_name": "ll_model.pth",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/ll_model.pth"
},
{
"file_name": "ll_model_cfg.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/ll_model_cfg.pkl"
},
{
"file_name": "meta.json",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/meta.json"
}
],
"transformer_cfg": {
"n_layers": 2,
"d_model": 16,
"n_ctx": 10,
"d_head": 4,
"model_name": "custom",
"n_heads": 4,
"d_mlp": 64,
"act_fn": "gelu",
"d_vocab": 10,
"eps": 1e-05,
"use_attn_result": true,
"use_attn_scale": true,
"use_split_qkv_input": true,
"use_hook_mlp_in": true,
"use_attn_in": false,
"use_local_attn": false,
"original_architecture": null,
"from_checkpoint": false,
"checkpoint_index": null,
"checkpoint_label_type": null,
"checkpoint_value": null,
"tokenizer_name": null,
"window_size": null,
"attn_types": null,
"init_mode": "gpt2",
"normalization_type": null,
"n_devices": 1,
"attention_dir": "causal",
"attn_only": false,
"seed": 0,
"initializer_range": 0.16329931618554522,
"init_weights": true,
"scale_attn_by_inverse_layer_idx": false,
"positional_embedding_type": "standard",
"final_rms": false,
"d_vocab_out": 5,
"parallel_attn_mlp": false,
"rotary_dim": null,
"n_params": 6144,
"use_hook_tokens": false,
"gated_mlp": false,
"default_prepend_bos": true,
"dtype": "torch.float32",
"tokenizer_prepends_bos": null,
"n_key_value_heads": null,
"post_embedding_ln": false,
"rotary_base": 10000,
"trust_remote_code": false,
"rotary_adjacent_pairs": false
},
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/ll_model_cfg.pkl",
"training_args": {
"atol": 0.05,
"lr": 0.01,
"use_single_loss": false,
"iit_weight": 1.0,
"behavior_weight": 1.0,
"strict_weight": 1.0,
"epochs": 2000,
"act_fn": "gelu",
"clip_grad_norm": 0.1,
"lr_scheduler": "",
"model_pair": "strict"
},
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/meta.json",
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/ll_model.pth",
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/edges.pkl"
},
{
"case_id": "35",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/35",
"task_description": "Alternates capitalization of each character in words.",
"vocab": [
"J",
"LB",
"TPSI",
"V",
"b",
"no",
"oCLrZaW",
"poiVg"
],
"max_seq_len": 10,
"min_seq_len": 4,
"files": [
{
"file_name": "edges.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/edges.pkl"
},
{
"file_name": "ll_model.pth",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/ll_model.pth"
},
{
"file_name": "ll_model_cfg.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/ll_model_cfg.pkl"
},
{
"file_name": "meta.json",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/meta.json"
}
],
"transformer_cfg": {
"n_layers": 2,
"d_model": 9,
"n_ctx": 10,
"d_head": 2,
"model_name": "custom",
"n_heads": 4,
"d_mlp": 36,
"act_fn": "gelu",
"d_vocab": 10,
"eps": 1e-05,
"use_attn_result": true,
"use_attn_scale": true,
"use_split_qkv_input": true,
"use_hook_mlp_in": true,
"use_attn_in": false,
"use_local_attn": false,
"original_architecture": null,
"from_checkpoint": false,
"checkpoint_index": null,
"checkpoint_label_type": null,
"checkpoint_value": null,
"tokenizer_name": null,
"window_size": null,
"attn_types": null,
"init_mode": "gpt2",
"normalization_type": null,
"n_devices": 1,
"attention_dir": "causal",
"attn_only": false,
"seed": 0,
"initializer_range": 0.1539600717839002,
"init_weights": true,
"scale_attn_by_inverse_layer_idx": false,
"positional_embedding_type": "standard",
"final_rms": false,
"d_vocab_out": 8,
"parallel_attn_mlp": false,
"rotary_dim": null,
"n_params": 1872,
"use_hook_tokens": false,
"gated_mlp": false,
"default_prepend_bos": true,
"dtype": "torch.float32",
"tokenizer_prepends_bos": null,
"n_key_value_heads": null,
"post_embedding_ln": false,
"rotary_base": 10000,
"trust_remote_code": false,
"rotary_adjacent_pairs": false
},
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/ll_model_cfg.pkl",
"training_args": {
"atol": 0.05,
"lr": 0.01,
"use_single_loss": false,
"iit_weight": 1.0,
"behavior_weight": 1.0,
"strict_weight": 1.0,
"epochs": 2000,
"act_fn": "gelu",
"clip_grad_norm": 0.1,
"lr_scheduler": "",
"model_pair": "strict"
},
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/meta.json",
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/ll_model.pth",
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/edges.pkl"
},
{
"case_id": "36",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/36",
"task_description": "Classifies each token as 'positive', 'negative', or 'neutral' based on emojis.",
"vocab": [
"\ud83d\udcd8",
"\ud83d\ude0a",
"\ud83d\ude22"
],
"max_seq_len": 10,
"min_seq_len": 4,
"files": [
{
"file_name": "edges.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/edges.pkl"
},
{
"file_name": "ll_model.pth",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/ll_model.pth"
},
{
"file_name": "ll_model_cfg.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/ll_model_cfg.pkl"
},
{
"file_name": "meta.json",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/meta.json"
}
],
"transformer_cfg": {
"n_layers": 2,
"d_model": 6,
"n_ctx": 10,
"d_head": 1,
"model_name": "custom",
"n_heads": 4,
"d_mlp": 24,
"act_fn": "gelu",
"d_vocab": 5,
"eps": 1e-05,
"use_attn_result": true,
"use_attn_scale": true,
"use_split_qkv_input": true,
"use_hook_mlp_in": true,
"use_attn_in": false,
"use_local_attn": false,
"original_architecture": null,
"from_checkpoint": false,
"checkpoint_index": null,
"checkpoint_label_type": null,
"checkpoint_value": null,
"tokenizer_name": null,
"window_size": null,
"attn_types": null,
"init_mode": "gpt2",
"normalization_type": null,
"n_devices": 1,
"attention_dir": "causal",
"attn_only": false,
"seed": 0,
"initializer_range": 0.19402850002906638,
"init_weights": true,
"scale_attn_by_inverse_layer_idx": false,
"positional_embedding_type": "standard",
"final_rms": false,
"d_vocab_out": 3,
"parallel_attn_mlp": false,
"rotary_dim": null,
"n_params": 768,
"use_hook_tokens": false,
"gated_mlp": false,
"default_prepend_bos": true,
"dtype": "torch.float32",
"tokenizer_prepends_bos": null,
"n_key_value_heads": null,
"post_embedding_ln": false,
"rotary_base": 10000,
"trust_remote_code": false,
"rotary_adjacent_pairs": false
},
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/ll_model_cfg.pkl",
"training_args": {
"atol": 0.05,
"lr": 0.01,
"use_single_loss": false,
"iit_weight": 1.0,
"behavior_weight": 1.0,
"strict_weight": 1.0,
"epochs": 2000,
"act_fn": "gelu",
"clip_grad_norm": 0.1,
"lr_scheduler": "",
"model_pair": "strict"
},
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/meta.json",
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/ll_model.pth",
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/edges.pkl"
},
{
"case_id": "37",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/37",
"task_description": "Reverses each word in the sequence except for specified exclusions.",
"vocab": [
"J",
"LB",
"TPSI",
"V",
"b",
"no",
"oCLrZaW",
"poiVg"
],
"max_seq_len": 10,
"min_seq_len": 4,
"files": [
{
"file_name": "edges.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/edges.pkl"
},
{
"file_name": "ll_model.pth",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/ll_model.pth"
},
{
"file_name": "ll_model_cfg.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/ll_model_cfg.pkl"
},
{
"file_name": "meta.json",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/meta.json"
}
],
"transformer_cfg": {
"n_layers": 2,
"d_model": 12,
"n_ctx": 10,
"d_head": 3,
"model_name": "custom",
"n_heads": 4,
"d_mlp": 48,
"act_fn": "gelu",
"d_vocab": 10,
"eps": 1e-05,
"use_attn_result": true,
"use_attn_scale": true,
"use_split_qkv_input": true,
"use_hook_mlp_in": true,
"use_attn_in": false,
"use_local_attn": false,
"original_architecture": null,
"from_checkpoint": false,
"checkpoint_index": null,
"checkpoint_label_type": null,
"checkpoint_value": null,
"tokenizer_name": null,
"window_size": null,
"attn_types": null,
"init_mode": "gpt2",
"normalization_type": null,
"n_devices": 1,
"attention_dir": "causal",
"attn_only": false,
"seed": 0,
"initializer_range": 0.1539600717839002,
"init_weights": true,
"scale_attn_by_inverse_layer_idx": false,
"positional_embedding_type": "standard",
"final_rms": false,
"d_vocab_out": 8,
"parallel_attn_mlp": false,
"rotary_dim": null,
"n_params": 3456,
"use_hook_tokens": false,
"gated_mlp": false,
"default_prepend_bos": true,
"dtype": "torch.float32",
"tokenizer_prepends_bos": null,
"n_key_value_heads": null,
"post_embedding_ln": false,
"rotary_base": 10000,
"trust_remote_code": false,
"rotary_adjacent_pairs": false
},
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/ll_model_cfg.pkl",
"training_args": {
"atol": 0.05,
"lr": 0.01,
"use_single_loss": false,
"iit_weight": 1.0,
"behavior_weight": 1.0,
"strict_weight": 1.0,
"epochs": 2000,
"act_fn": "gelu",
"clip_grad_norm": 0.1,
"lr_scheduler": "",
"model_pair": "strict"
},
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/meta.json",
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/ll_model.pth",
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/edges.pkl"
},
{
"case_id": "4",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/4",
"task_description": "Return fraction of previous open tokens minus the fraction of close tokens.",
"vocab": [
"(",
")",
"a",
"b",
"c"
],
"max_seq_len": 10,
"min_seq_len": 4,
"files": [
{
"file_name": "edges.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/edges.pkl"
},
{
"file_name": "ll_model.pth",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/ll_model.pth"
},
{
"file_name": "ll_model_cfg.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/ll_model_cfg.pkl"
},
{
"file_name": "meta.json",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/meta.json"
}
],
"transformer_cfg": {
"n_layers": 2,
"d_model": 20,
"n_ctx": 10,
"d_head": 5,
"model_name": "custom",
"n_heads": 4,
"d_mlp": 80,
"act_fn": "gelu",
"d_vocab": 7,
"eps": 1e-05,
"use_attn_result": true,
"use_attn_scale": true,
"use_split_qkv_input": true,
"use_hook_mlp_in": true,
"use_attn_in": false,
"use_local_attn": false,
"original_architecture": null,
"from_checkpoint": false,
"checkpoint_index": null,
"checkpoint_label_type": null,
"checkpoint_value": null,
"tokenizer_name": null,
"window_size": null,
"attn_types": null,
"init_mode": "gpt2",
"normalization_type": null,
"n_devices": 1,
"attention_dir": "causal",
"attn_only": false,
"seed": 0,
"initializer_range": 0.17056057308448835,
"init_weights": true,
"scale_attn_by_inverse_layer_idx": false,
"positional_embedding_type": "standard",
"final_rms": false,
"d_vocab_out": 1,
"parallel_attn_mlp": false,
"rotary_dim": null,
"n_params": 9600,
"use_hook_tokens": false,
"gated_mlp": false,
"default_prepend_bos": true,
"dtype": "torch.float32",
"tokenizer_prepends_bos": null,
"n_key_value_heads": null,
"post_embedding_ln": false,
"rotary_base": 10000,
"trust_remote_code": false,
"rotary_adjacent_pairs": false
},
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/ll_model_cfg.pkl",
"training_args": {
"atol": 0.05,
"lr": 0.001,
"use_single_loss": false,
"iit_weight": 1.0,
"behavior_weight": 1.0,
"strict_weight": 0.4,
"epochs": 2000,
"act_fn": "gelu",
"clip_grad_norm": 0.1,
"lr_scheduler": ""
},
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/meta.json",
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/ll_model.pth",
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/edges.pkl"
},
{
"case_id": "8",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/8",
"task_description": "Fills gaps between tokens with a specified filler.",
"vocab": [
"J",
"LB",
"TPSI",
"V",
"b",
"no",
"oCLrZaW",
"poiVg"
],
"max_seq_len": 10,
"min_seq_len": 4,
"files": [
{
"file_name": "edges.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/8/edges.pkl"
},
{
"file_name": "ll_model.pth",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/8/ll_model.pth"
},
{
"file_name": "ll_model_cfg.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/8/ll_model_cfg.pkl"
},
{
"file_name": "meta.json",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/8/meta.json"
}
],
"transformer_cfg": {
"n_layers": 2,
"d_model": 20,
"n_ctx": 10,
"d_head": 5,
"model_name": "custom",
"n_heads": 4,
"d_mlp": 80,
"act_fn": "gelu",
"d_vocab": 10,
"eps": 1e-05,
"use_attn_result": true,
"use_attn_scale": true,
"use_split_qkv_input": true,
"use_hook_mlp_in": true,
"use_attn_in": false,
"use_local_attn": false,
"original_architecture": null,
"from_checkpoint": false,
"checkpoint_index": null,
"checkpoint_label_type": null,
"checkpoint_value": null,
"tokenizer_name": null,
"window_size": null,
"attn_types": null,
"init_mode": "gpt2",
"normalization_type": null,
"n_devices": 1,
"attention_dir": "causal",
"attn_only": false,
"seed": 0,
"initializer_range": 0.13333333333333333,
"init_weights": true,
"scale_attn_by_inverse_layer_idx": false,
"positional_embedding_type": "standard",
"final_rms": false,
"d_vocab_out": 8,
"parallel_attn_mlp": false,
"rotary_dim": null,
"n_params": 9600,
"use_hook_tokens": false,
"gated_mlp": false,
"default_prepend_bos": true,
"dtype": "torch.float32",
"tokenizer_prepends_bos": null,
"n_key_value_heads": null,
"post_embedding_ln": false,
"rotary_base": 10000,
"trust_remote_code": false,
"rotary_adjacent_pairs": false
},
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/8/ll_model_cfg.pkl",
"training_args": {
"atol": 0.05,
"lr": 0.01,
"use_single_loss": false,
"iit_weight": 1.0,
"behavior_weight": 1.0,
"strict_weight": 0.4,
"epochs": 500,
"act_fn": "gelu",
"clip_grad_norm": 1.0,
"lr_scheduler": ""
},
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/8/meta.json",
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/8/ll_model.pth",
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/8/edges.pkl"
},
{
"case_id": "ioi",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/ioi",
"task_description": "Indirect object identification",
"max_seq_len": 16,
"min_seq_len": 16,
"files": [
{
"file_name": "edges.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/edges.pkl"
},
{
"file_name": "ll_model.pth",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/ll_model.pth"
},
{
"file_name": "ll_model_cfg.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/ll_model_cfg.pkl"
},
{
"file_name": "meta.json",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/meta.json"
}
],
"transformer_cfg": {
"n_layers": 6,
"d_model": 64,
"n_ctx": 1024,
"d_head": 16,
"model_name": "gpt2",
"n_heads": 4,
"d_mlp": 3072,
"act_fn": "gelu_new",
"d_vocab": 50257,
"eps": 1e-05,
"use_attn_result": false,
"use_attn_scale": true,
"use_split_qkv_input": false,
"use_hook_mlp_in": false,
"use_attn_in": false,
"use_local_attn": false,
"original_architecture": "GPT2LMHeadModel",
"from_checkpoint": false,
"checkpoint_index": null,
"checkpoint_label_type": null,
"checkpoint_value": null,
"tokenizer_name": "gpt2",
"window_size": null,
"attn_types": null,
"init_mode": "gpt2",
"normalization_type": "LNPre",
"n_devices": 1,
"attention_dir": "causal",
"attn_only": false,
"seed": null,
"initializer_range": 0.02886751345948129,
"init_weights": true,
"scale_attn_by_inverse_layer_idx": false,
"positional_embedding_type": "standard",
"final_rms": false,
"d_vocab_out": 50257,
"parallel_attn_mlp": false,
"rotary_dim": null,
"n_params": 2457600,
"use_hook_tokens": false,
"gated_mlp": false,
"default_prepend_bos": true,
"dtype": "torch.float32",
"tokenizer_prepends_bos": false,
"n_key_value_heads": null,
"post_embedding_ln": false,
"rotary_base": 10000,
"trust_remote_code": false,
"rotary_adjacent_pairs": false
},
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/ll_model_cfg.pkl",
"training_args": {
"next_token": true,
"non_ioi_thresh": 0.65,
"use_per_token_check": false,
"batch_size": 256,
"lr": 0.001,
"num_workers": 0,
"early_stop": true,
"lr_scheduler": null,
"scheduler_val_metric": [
"val/accuracy",
"val/IIA"
],
"scheduler_mode": "max",
"clip_grad_norm": 1.0,
"atol": 0.05,
"use_single_loss": false,
"iit_weight": 1.0,
"behavior_weight": 1.0,
"strict_weight": 0.4
},
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/meta.json",
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/ll_model.pth",
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/edges.pkl"
},
{
"case_id": "ioi_next_token",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/ioi_next_token",
"task_description": "Indirect object identification",
"max_seq_len": 16,
"min_seq_len": 16,
"files": [
{
"file_name": "edges.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/edges.pkl"
},
{
"file_name": "ll_model.pth",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/ll_model.pth"
},
{
"file_name": "ll_model_cfg.pkl",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/ll_model_cfg.pkl"
},
{
"file_name": "meta.json",
"url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/meta.json"
}
],
"transformer_cfg": {
"n_layers": 6,
"d_model": 64,
"n_ctx": 1024,
"d_head": 16,
"model_name": "gpt2",
"n_heads": 4,
"d_mlp": 3072,
"act_fn": "gelu_new",
"d_vocab": 50257,
"eps": 1e-05,
"use_attn_result": false,
"use_attn_scale": true,
"use_split_qkv_input": false,
"use_hook_mlp_in": false,
"use_attn_in": false,
"use_local_attn": false,
"original_architecture": "GPT2LMHeadModel",
"from_checkpoint": false,
"checkpoint_index": null,
"checkpoint_label_type": null,
"checkpoint_value": null,
"tokenizer_name": "gpt2",
"window_size": null,
"attn_types": null,
"init_mode": "gpt2",
"normalization_type": "LNPre",
"n_devices": 1,
"attention_dir": "causal",
"attn_only": false,
"seed": null,
"initializer_range": 0.02886751345948129,
"init_weights": true,
"scale_attn_by_inverse_layer_idx": false,
"positional_embedding_type": "standard",
"final_rms": false,
"d_vocab_out": 50257,
"parallel_attn_mlp": false,
"rotary_dim": null,
"n_params": 2457600,
"use_hook_tokens": false,
"gated_mlp": false,
"default_prepend_bos": true,
"dtype": "torch.float32",
"tokenizer_prepends_bos": false,
"n_key_value_heads": null,
"post_embedding_ln": false,
"rotary_base": 10000,
"trust_remote_code": false,
"rotary_adjacent_pairs": false
},
"transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/ll_model_cfg.pkl",
"training_args": {
"next_token": true,
"non_ioi_thresh": 0.65,
"use_per_token_check": false,
"batch_size": 256,
"lr": 0.001,
"num_workers": 0,
"early_stop": true,
"lr_scheduler": null,
"scheduler_val_metric": [
"val/accuracy",
"val/IIA"
],
"scheduler_mode": "max",
"clip_grad_norm": 1.0,
"atol": 0.05,
"use_single_loss": false,
"iit_weight": 1.0,
"behavior_weight": 1.0,
"strict_weight": 0.4
},
"training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/meta.json",
"weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/ll_model.pth",
"circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/edges.pkl"
}
]
}