asahi417 commited on
Commit
a5f4f1e
·
1 Parent(s): 4a04b75
Files changed (2) hide show
  1. main_s2s.sh +7 -0
  2. tokenize_dataset_s2s.py +1 -1
main_s2s.sh CHANGED
@@ -18,6 +18,13 @@ do
18
  echo ${LINE_NO_START}
19
  python fetch_dataset_s2s.py
20
  done
 
 
 
 
 
 
 
21
 
22
  ####################
23
  # enA-viA: 740_598 #
 
18
  echo ${LINE_NO_START}
19
  python fetch_dataset_s2s.py
20
  done
21
+ # tokenize
22
+ for i in $(seq 1 144);
23
+ do
24
+ export DATASET_ID=${i}
25
+ export DIRECTION="enA-jaA"
26
+ python tokenize_dataset_s2s.py
27
+ done
28
 
29
  ####################
30
  # enA-viA: 740_598 #
tokenize_dataset_s2s.py CHANGED
@@ -12,7 +12,7 @@ batch_size = int(os.getenv("BATCH_SIZE", 64))
12
  num_proc = int(os.getenv("NUM_PROC", 1))
13
  hf_org = os.getenv("HF_ORG", "asahi417")
14
  hf_dataset = f"seamless-align-{direction}"
15
- dataset = load_dataset(f"{hf_org}/{hf_dataset}", config_name=f"subset_{dataset_id}", split="train")
16
  tokenizer = EncodecTokenizer.from_pretrained()
17
 
18
 
 
12
  num_proc = int(os.getenv("NUM_PROC", 1))
13
  hf_org = os.getenv("HF_ORG", "asahi417")
14
  hf_dataset = f"seamless-align-{direction}"
15
+ dataset = load_dataset(f"{hf_org}/{hf_dataset}", f"subset_{dataset_id}", split="train")
16
  tokenizer = EncodecTokenizer.from_pretrained()
17
 
18