File size: 79,857 Bytes
6fa4bc9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
{
    "paper_id": "O12-1007",
    "header": {
        "generated_with": "S2ORC 1.0.0",
        "date_generated": "2023-01-19T08:02:50.955309Z"
    },
    "title": "Measuring Individual Differences in Word Recognition: The Role of Individual Lexical Behaviors",
    "authors": [
        {
            "first": "Hsin-Ni",
            "middle": [],
            "last": "Lin",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "Linguistics Division National Taiwan Normal University",
                "location": {}
            },
            "email": ""
        },
        {
            "first": "Shu-Kai",
            "middle": [],
            "last": "Hsieh",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "National Taiwan University",
                "location": {}
            },
            "email": "shukaihsieh@ntu.edu.tw"
        },
        {
            "first": "Shiao-Hui",
            "middle": [],
            "last": "Chan",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "Normal University",
                "location": {
                    "country": "Taiwan"
                }
            },
            "email": "shiaohui@ntnu.edu.tw"
        }
    ],
    "year": "",
    "venue": null,
    "identifiers": {},
    "abstract": "This study adopts a corpus-based computational linguistic approach to measure individual differences (IDs) in visual word recognition. Word recognition has been a cardinal issue in the field of psycholinguistics. Previous studies examined the IDs by resorting to test-based or questionnaire-based measures. Those measures, however, confined the research within the scope where they can evaluate. To extend the research to approximate to IDs in real life, the present study undertakes the issue from the observations of experiment participants' daily-life lexical behaviors. Based on participants' Facebook posts, two types of personal lexical behaviors are computed, including the frequency index of personal word usage and personal word frequency. It is investigated that to what extent each of them accounts for participants' variances in Chinese word recognition. The data analyses are carried out by mixed-effects models, which can precisely estimate by-subject differences. Results showed that the effects of personal word frequency reached significance; participants responded themselves more rapidly when encountering more frequently used words. People with lower frequency indices of personal word usage had a lower accuracy rates than others, which was contrary to our prediction. Comparison and discussion of the results also reveal methodology issues that can provide noteworthy suggestions for future research on measuring personal lexical behaviors.",
    "pdf_parse": {
        "paper_id": "O12-1007",
        "_pdf_hash": "",
        "abstract": [
            {
                "text": "This study adopts a corpus-based computational linguistic approach to measure individual differences (IDs) in visual word recognition. Word recognition has been a cardinal issue in the field of psycholinguistics. Previous studies examined the IDs by resorting to test-based or questionnaire-based measures. Those measures, however, confined the research within the scope where they can evaluate. To extend the research to approximate to IDs in real life, the present study undertakes the issue from the observations of experiment participants' daily-life lexical behaviors. Based on participants' Facebook posts, two types of personal lexical behaviors are computed, including the frequency index of personal word usage and personal word frequency. It is investigated that to what extent each of them accounts for participants' variances in Chinese word recognition. The data analyses are carried out by mixed-effects models, which can precisely estimate by-subject differences. Results showed that the effects of personal word frequency reached significance; participants responded themselves more rapidly when encountering more frequently used words. People with lower frequency indices of personal word usage had a lower accuracy rates than others, which was contrary to our prediction. Comparison and discussion of the results also reveal methodology issues that can provide noteworthy suggestions for future research on measuring personal lexical behaviors.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Abstract",
                "sec_num": null
            }
        ],
        "body_text": [
            {
                "text": "In the field of psycholinguistics, a major research interest is to investigate how people recognize written words or access the corresponding word representations stored in their mental lexicon. Psycholinguists usually undertake the investigation starting from isolated words since less factors are involved, compared to words within sentences. Therefore, research on the isolated word recognition is fundamental for understanding how lexical access takes places. In general, the term 'visual word recognition' is used to simply address the recognition of isolated written words.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1."
            },
            {
                "text": "Research of word recognition traditionally have concentrated on how characteristics of words per se (e.g. word length, word frequency, or neighborhood size) affected the procedure of recognition [1] [2] [3] [4] [5] , taking the discrepancies between participants' performance as merely statistical deviation. Recently, however, there has been a growing interest in the individual differences (IDs, henceforth) of experiment participants. Results of the ID studies showed that the issue was noteworthy because personal experiences and knowledge of words (e.g. print-exposure experience [6] [7] , reading skills [8] , or vocabulary knowledge [9] [10] [11] ) accounted for systematic variances between participants in word recognition. Even when participants were homogeneous in their educational level, their IDs sufficiently resulted in distinct performance in word recognition. Furthermore, [8] provided compelling evidence that conflicting results of regularity effects 1 in the literature were attributable to lacking control over participants' IDs of reading skills.",
                "cite_spans": [
                    {
                        "start": 195,
                        "end": 198,
                        "text": "[1]",
                        "ref_id": "BIBREF0"
                    },
                    {
                        "start": 203,
                        "end": 206,
                        "text": "[3]",
                        "ref_id": "BIBREF2"
                    },
                    {
                        "start": 211,
                        "end": 214,
                        "text": "[5]",
                        "ref_id": "BIBREF4"
                    },
                    {
                        "start": 585,
                        "end": 588,
                        "text": "[6]",
                        "ref_id": "BIBREF5"
                    },
                    {
                        "start": 589,
                        "end": 592,
                        "text": "[7]",
                        "ref_id": "BIBREF6"
                    },
                    {
                        "start": 610,
                        "end": 613,
                        "text": "[8]",
                        "ref_id": "BIBREF7"
                    },
                    {
                        "start": 640,
                        "end": 643,
                        "text": "[9]",
                        "ref_id": "BIBREF8"
                    },
                    {
                        "start": 649,
                        "end": 653,
                        "text": "[11]",
                        "ref_id": "BIBREF10"
                    },
                    {
                        "start": 891,
                        "end": 894,
                        "text": "[8]",
                        "ref_id": "BIBREF7"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1."
            },
            {
                "text": "To date, the studies of IDs, however, have focused on test-measured or self-rated ID variables. In such approaches, the observed IDs were confined in the boundary of a test or questionnaire design, and the uniqueness of each individual in real life was neglected. In an attempt to examine the approximate real-life IDs, this research measures and analyzes IDs based on each participant's own lexical behaviors. Lexical behaviors here refer to a person's word usage and preference in his/her daily life. Intuitively, language usage reveals one's vocabulary knowledge, such as the words the person knows and how to use those words within context. Vocabulary knowledge was proved relating to word recognition [9] [10] [11] ; hence, it is highly possible that IDs of lexical behaviors can explain the disparity of participants' performance in word recognition. The lexical behaviors mainly have two merits over the measure of vocabulary tests. First, people's lexical knowledge will be evaluated not by a small set of vocabularies in a given test, but by the words used by themselves. In this case, a variable's value assigned to a given participant is personalized and not confined to the scale or the total score of a test. The other merit resides in that the data of language usage can provide a deeper insight into a person's lexical knowledge, compared with a vocabulary test. If a person is able to use or produce a given word naturally (and frequently), it suggests that the word's representation has been firmly established in his/her mental lexicon.",
                "cite_spans": [
                    {
                        "start": 706,
                        "end": 709,
                        "text": "[9]",
                        "ref_id": "BIBREF8"
                    },
                    {
                        "start": 715,
                        "end": 719,
                        "text": "[11]",
                        "ref_id": "BIBREF10"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1."
            },
            {
                "text": "Besides, it is worth noting that the stance we take in measuring the 'individuality' is naturalistic rather than natural, in that the lexical behaviors we describe are assumedly anchored in the interaction as naturalistic situated interactions, rather than natural ones (like using camera to collect data). A pitfall of the natural ones is that when observers and/or cameras are present those interactions are not quite what they would be in our absence.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1."
            },
            {
                "text": "Therefore, the present study begins with a preliminary survey on the lexical behaviors of participants' naturalistic data on Facebook 2 Walls (Figure 1 ).",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 142,
                        "end": 151,
                        "text": "(Figure 1",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1."
            },
            {
                "text": "Our attention for lexical behaviors computed from participants' Facebook data is fastened upon the frequency index of personal word usage and the personal word frequency calculated from participants' language data. Whether the two variables are associated with participant's performance in a lexical decision task 3 will be explored respectively in two experiments. More important, as a pioneer study on lexical behaviors and word recognition, the other main objective of this research is to preliminarily explore its computational methodology.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Figure 1. A snapshot of Facebook Wall",
                "sec_num": null
            },
            {
                "text": "The rest of this paper is organized as follows: Section 2 presents the procedure of our data collection, including conducting a lexical decision experiment and extracting the experiment participants' language usage data from the Facebook. Section 3 demonstrates the methods and results of two experiments, each of which computed a lexical behavior variable and further examined the relationships between participants' IDs of lexical behaviors and lexical-decision responses. Section 4 concludes this study by giving a summary and contributions of the current study. Section 5 provides potential research directions for future work.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Figure 1. A snapshot of Facebook Wall",
                "sec_num": null
            },
            {
                "text": "Sixteen Chinese native speakers (10 females and 6 males; ages ranging from 21 to 29 years old) consented to participant in the task and were offered participant fees. For the purpose of augmenting the possibility of finding individual differences (IDs) of personal lexical behaviors, the participants were recruited from diverse backgrounds. They should be right-handed, which was examined via a self-report handedness inventory [12] .",
                "cite_spans": [
                    {
                        "start": 429,
                        "end": 433,
                        "text": "[12]",
                        "ref_id": "BIBREF11"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Participants",
                "sec_num": "2.1.1"
            },
            {
                "text": "Experiment materials included 456 Chinese words and 456 non-words. The word stimuli were nouns selected from the Chinese Lexicon Profile (CLP) 4 , comprising 152 high-frequency, 152 mid-frequency, and 152 low-frequency words. In addition to word frequency, the number of characters, the number of senses, and the neighborhood size of words were collected from the CLP and will be treated as covariates at the stage of statistical analysis because we intended to disentangle their impacts on the lexical-decision responses.",
                "cite_spans": [
                    {
                        "start": 143,
                        "end": 144,
                        "text": "4",
                        "ref_id": "BIBREF3"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Materials",
                "sec_num": "2.1.2"
            },
            {
                "text": "To equalize yes and no stimuli, 456 non-words were also subsumed into the stimuli. These non-words were randomly generated by using characters of existing nouns in Chinese. Take two-character non-words for example. The procedure of random generation is illustrated in Figure 2 . The first and second characters of existing nominal words were separately stored into two vectors. Next, the first and second characters of a non-word were randomly selected from the two vectors respectively and then combined altogether. If an automatically generated non-word sounded like an existing word, it would be removed from the non-word list.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 268,
                        "end": 276,
                        "text": "Figure 2",
                        "ref_id": "FIGREF0"
                    }
                ],
                "eq_spans": [],
                "section": "Materials",
                "sec_num": "2.1.2"
            },
            {
                "text": "The task is a within-subjects design; that is, a participant saw all of the 912 stimuli. The non-words, high-, mid-, and low-frequency words were evenly divided into four blocks. The order of four blocks was counterbalanced across 16 participants. Within a block, experimental stimuli were administered in a random order. Each participant was tested individually in a quiet room. The experiment was conducted and presented on a laptop via E-prime 2.0 professional. Participants were instructed to judge whether a visually presented stimulus was a meaningful word in Mandarin Chinese. They were required to respond as quickly as possible but without expense of accuracy, and their judgment were recorded as soon as they pressed the 'yes' or 'no' response button.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Materials",
                "sec_num": "2.1.2"
            },
            {
                "text": "The procedure of a trial was initiated with a fixation sign (+) appearing in the center of the monitor for 1000 ms. Next, a stimulus was presented. The presentation would be terminated immediately when a participant responded. If no response was detected in 4000 ms, the given stimulus would be removed from the monitor. After termination of the stimulus presentation, a feedback was provided on the monitor for 750 ms, along with the participant's accumulated accuracy rate in a block.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Materials",
                "sec_num": "2.1.2"
            },
            {
                "text": "The entire experiment included four blocks and lasted approximately one hour. Prior to the experiment, a practice session was given to familiarize participants with the experimental procedure. The session contained 4 words and 4 non-words, none of which appeared in the formal experiment.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Materials",
                "sec_num": "2.1.2"
            },
            {
                "text": "The Facebook module in i-Corpus 5 was employed to gathering participants' data of language usage and preferences. The procedure is presented beneath. For the module was in its rudimentary stage of development, it was still semi-autonomous; more specifically, the initial steps in the procedure were manually accomplished.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Facebook data",
                "sec_num": "2.2"
            },
            {
                "text": "[",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Facebook data",
                "sec_num": "2.2"
            },
            {
                "text": "Step one] Log in an APP to get a user's access token to Facebook Step five] Extract each message in categories of post, photo, comment, and other users' walls (One message was saved as a text.) In this study, the quantification of participants' lexical behaviors is based on only the category of posts given that other categories of messages have context which is not shown in themselves.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Facebook data",
                "sec_num": "2.2"
            },
            {
                "text": "[",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Facebook data",
                "sec_num": "2.2"
            },
            {
                "text": "Step six] Pre-process the 'post' messages by the CKIP Chinese Word Segmentation System 6 . After the segmentation, we obtained the token number in each participant' data of language usage (see Table 1 ). Results of the automatic segmentation were not further checked and corrected by human labor because the present study purports to explore and develop a methodology that is not labor-consuming and rather feasible for future research to compute and control the IDs of lexical behaviors. The segmented words from participants' Facebook posts were prepared for the computation of personal lexical behaviors proposed in the subsequent section.",
                "cite_spans": [
                    {
                        "start": 87,
                        "end": 88,
                        "text": "6",
                        "ref_id": "BIBREF5"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 193,
                        "end": 200,
                        "text": "Table 1",
                        "ref_id": "TABREF0"
                    }
                ],
                "eq_spans": [],
                "section": "Facebook data",
                "sec_num": "2.2"
            },
            {
                "text": "Word frequency in corpora was attested to have a high negative correlation with word difficulty [13] . In this experiment, the Academia Sinica Balanced Corpus 7 frequency of a word was analogously taken as the possibility that the word is generally acquired and used by native speakers, thus being referred to for computing the frequency index of personal word usage. A lower frequency index of word usage indicates that a person was apt to use low-frequency words, which was preliminarily assumed to imply a person's relatively broader vocabulary knowledge. It was concerned that whether IDs of the frequency indices across participants were capable of explaining their differences in response latencies and accuracies.",
                "cite_spans": [
                    {
                        "start": 96,
                        "end": 100,
                        "text": "[13]",
                        "ref_id": "BIBREF12"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Experiments on the individual differences of lexical behaviors 3.1 Experiment 1: The role of the frequency index of personal word usage in visual word recognition",
                "sec_num": "3."
            },
            {
                "text": "There were four steps to compute the frequency index per person, as shown in the following.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3.1.1"
            },
            {
                "text": "[",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3.1.1"
            },
            {
                "text": "Step one] Produce a list per participant which contained all of the words he/she used and the occurrence frequency of those words in his/her segmented Facebook data. Examples are shown in the first and second columns of Table 2 .",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 220,
                        "end": 227,
                        "text": "Table 2",
                        "ref_id": "TABREF1"
                    }
                ],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3.1.1"
            },
            {
                "text": "[",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3.1.1"
            },
            {
                "text": "Step two] Gather from the CLP the corresponding word frequency in Sinica Corpus of each word on the list, as exemplified in the third column of Table 2 . Note that a few words were assigned a missing value \"NA\" in the column since they did not appear in the Sinica Corpus. Those words, which possessed no Sinica frequencies, would be excluded from the calculation of participants' frequency indices. Given that some of them were a string that was erroneously grouped as a word by the automatic segmentation program (e.g. zai4 wuo3 nao3 ( ) 'in my brain'), the exclusion enabled this experiment to filter out the data noise procured by automatic segmentation, thus diminishing the impact of segmentation errors on the calculation of individual lexical behaviors.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 144,
                        "end": 151,
                        "text": "Table 2",
                        "ref_id": "TABREF1"
                    }
                ],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3.1.1"
            },
            {
                "text": "[",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3.1.1"
            },
            {
                "text": "Step three] Compute the frequency index of personal word usage of the participant j by (1) , where was the participant's personal frequency of the ith word, and was the word's frequency in the Sinica Corpus. In this equation, can be interpreted as the mean Sinica frequency of words used by the participant j on the Facebook. The lower the index was, the more rarely-seen words used by the participant were, which assumedly meant the person had broader word knowledge.",
                "cite_spans": [
                    {
                        "start": 87,
                        "end": 90,
                        "text": "(1)",
                        "ref_id": "BIBREF0"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3.1.1"
            },
            {
                "text": "[",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3.1.1"
            },
            {
                "text": "Step four] The index of each participant was put along with his/her response latencies and accuracies in the lexical decision task for analysis.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3.1.1"
            },
            {
                "text": "The steps of computation introduced above applied to the complete word list of each participant (called as \"the Intact word list\" hereafter). In addition to the list, this experiment also made the other word list for each participant to calculate another index. This word list (called as \"the NV word list\" hereafter) comprised only multi-character words tagged as nouns and verbs by CKIP Segmentation System and was preliminarily considered to be less affected by segmentation errors, compared with the Intact list. ",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3.1.1"
            },
            {
                "text": "The data analyses were conducted by mixed-effects models in the lme4 package of R 8 since the models can precisely estimate by-subject differences. In both the latency and accuracy analyses, experiment stimuli and participants were treated as random factors in the models. Procedure variables (i.e. block number and trial number) as well as word variables including types of word frequency, sense number, character number, and neighborhood size were taken as covariates. The inclusion of covariates was intended to disentangle their independent influences on the reaction latencies and accuracies. Provided that any covariate did not reach significance, it would be dropped out of the analysis; afterwards, the other variables would refit the models.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Results and Discussion",
                "sec_num": "3.1.2"
            },
            {
                "text": "Ahead of the analysis of response latencies, incorrect responses (2.57%) were discarded at first. Two frequency indices of personal word usage respectively fitted mixed-effects models together with the above-mentioned random factors and covariates. Besides, note that the response latencies put into statistical analyses were log-transformed so as to reduce skewed distribution of reaction time. Inspection of the residuals of the models found notable non-normality, as shown in the upper right panel of Figure 3 9 . To improve the goodness of fit, we removed outliers with standardized residuals outside the interval (-2.5, 2.5) [14, 15] , which were 2.54% of the correct-response data set in models of the Intact list and the NV list. After the removal, the models were refitted; the residuals of the refitted models are displayed in the lower right panel in the figure. As can be seen, the non-normality of the residuals was attenuated. In the final models, statistical results showed that the frequency indices from the Intact list (p = .3638) and NV list (p = .4926) both did not significantly vary with participants' response. Concerning the analysis of response accuracies, responses to all of the word stimuli in the task were taken into the analysis. Correct responses were coded as ones, and incorrect response as zeros. Seeing that the accuracy values were binomial, the analysis was carried out by the logistic mixed-effect models. Results suggested that the index computed from participants' NV lists was found to affect response accuracies (p < .001). Its effect on the accuracy, however, was opposite to our preliminary prediction that lower indices should suggest a person had broader lexical knowledge, thus relating to higher accuracy rates. Experimental results revealed that people with lower indices responded less accurately than those with higher indices. The counter-prediction may be ascribed to our methodology of computing the frequency index in two aspects.",
                "cite_spans": [
                    {
                        "start": 630,
                        "end": 634,
                        "text": "[14,",
                        "ref_id": "BIBREF13"
                    },
                    {
                        "start": 635,
                        "end": 638,
                        "text": "15]",
                        "ref_id": "BIBREF14"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 504,
                        "end": 514,
                        "text": "Figure 3 9",
                        "ref_id": "FIGREF2"
                    }
                ],
                "eq_spans": [],
                "section": "Results and Discussion",
                "sec_num": "3.1.2"
            },
            {
                "text": "The first aspect resides in that the personal indices were calculated by referring to an external lexical resource (i.e. the Academia Sinica Balance Corpus), where word frequency counts mainly came from written data rather than spoken data. When observing the calculation, we found that low-frequency words in the Sinica corpus encompassed not only rarely-used words but also words that were commonly used in daily-life conversation. Under the circumstances, a participant might receive a low frequency index from our computation because he/she utilized a number of 'low-frequency' words that are ubiquitous in spoken data, which are certainly not associated with broad lexical knowledge. This problem would become apparent when the frequency index was computed from the NV list of personal word usage. Unlike the NV list, the Intact list contained function words in addition to nouns and verbs. Function words, such as pronouns or conjunctions, are words that express grammatical relations between sentences and other words, so their occurrence in both written and spoken data must be high. With the involvement of function words, the Intact list could relieve the computation problem which was yielded by the huge discrepancy of word frequencies between written and spoken data. This is the possible reason why our results depending on the NV list showed that people with lower frequency indices had lower response accuracies but the results relying on the Intact list did not.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Results and Discussion",
                "sec_num": "3.1.2"
            },
            {
                "text": "The second aspect is that participants posted messages on their own Facebook Wall for diverse main purposes. Facebook is a social network designed for users to convey themselves and communicate with friends. Users can freely post any kind of messages they would like to share on their own Facebook Walls. Some users favored confiding their feelings at one moment; some preferred sharing anecdotes they experienced on a day; others often made serious comments on news and social events to evoke friends' or even the public's awareness. A skim over the Facebook data we collected could detect that the phenomena happened to users participating in this study. Accordingly, modes of the collected personal language data varied over a continuum illustrated in Figure 4 . For instance, participants who were used to casually express their feelings in the data would be closer to the \"informal\" and \"spoken\" end of the continuum. A concern is raised about those who tended to take the Facebook Wall as the space to share informal messages. Even if a person has broad vocabulary knowledge and would use rarely-seen words when writing formal messages or articles, the possibility that he/she uses those words in the informal/spoken mode might decrease. Furthermore, due to the inconsistent modes across participants' Facebook data, the seriousness of the problem caused by the Sinica Corpus word frequency might vary from person to person. As mentioned above, various commonly-used spoken or informal words were shown as low-frequency words in the Sinica corpus. Those spoken vocabularies were the sources from which our computed frequency indices were distorted. Consequently, if one's Facebook posts were generally close to the informal end of the mode continuum, his/her index would be largely affected by the problem originated from the Sinica word frequency. According to the two forgoing aspects, our counter-hypothesis findings were predominately accredited to the Sinica word frequencies. Thereupon, it is suggested that the computation of frequency indices in future research should take a spoken corpus as the reference of general word frequencies. With respect to the concern that people with broad lexical knowledge may use informal register and extensively-used vocabularies on the Facebook, it is a reflection we had when looking at the Facebook data. The extent to which it impacted on the index computation was unsure. A future research may probe into the extent by comparing the frequency indices calculated from people's Facebook posts with those form their compositions in an academic exam. The compositions in an exam would be scored. In that case, people must write in the formal mode to show their competence as they can as possible. Via a comparison with this formal data of language usage, the influence of the informal Facebook posts on the frequency index can be known.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 755,
                        "end": 763,
                        "text": "Figure 4",
                        "ref_id": "FIGREF3"
                    }
                ],
                "eq_spans": [],
                "section": "Results and Discussion",
                "sec_num": "3.1.2"
            },
            {
                "text": "This experiment investigates whether a subject's personal word frequency of a certain LDT 10 stimulus would influence his/her corresponding reaction latency. It was preliminarily hypothesized that if he/she used a word more frequently than other words, the response to the word would be more rapid. Besides, as shown in Table 1 , each participant's data differ in length; to render frequency counts across the data sets comparable, two kinds of normalization were conducted. A comparison on the effectiveness of the normalization methods is also provided in the discussion on experiment results.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 320,
                        "end": 327,
                        "text": "Table 1",
                        "ref_id": "TABREF0"
                    }
                ],
                "eq_spans": [],
                "section": "Experiment 2: The role of personal word frequency in visual word recognition",
                "sec_num": "3.2"
            },
            {
                "text": "The personal word frequency referred to the relative degrees to which a given LDT occurred in one's Facebook posts. Steps for its calculation are as follows:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3.2.1"
            },
            {
                "text": "[",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3.2.1"
            },
            {
                "text": "Step one] All of 16 participants' Facebook data were joined altogether into a file at first. If an LDT word stimulus appeared at least once in the file, it was chosen to be examined in this experiment. In total, there were 218 LDT stimuli conforming to the criterion, thus taken as the stimuli in this experiment.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3.2.1"
            },
            {
                "text": "[",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3.2.1"
            },
            {
                "text": "Step two] Personal word frequencies of the 218 stimuli were automatically counted.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3.2.1"
            },
            {
                "text": "[",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3.2.1"
            },
            {
                "text": "Step three] Two distinct methods were utilized to normalize the frequency counts. The first method was to divide the each subject's word frequencies by his/her own summed token numbers (see (2) ). In the equation, was the participant j's frequency count of the ith word; the i was limited between 1 to 218 since only 218 words were selected as stimuli in this experiment. However, note that the i in the denominator was not limited within the range, but by n instead. The n was the number of word types in a participant's Facebook data. In other words, the denominator added up word frequencies of all word types, thus representing the participant's total token number. Consequently, the output of the equation, , was the participant j's frequency ratio of the ith stimulus.",
                "cite_spans": [
                    {
                        "start": 190,
                        "end": 193,
                        "text": "(2)",
                        "ref_id": "BIBREF1"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3.2.1"
            },
            {
                "text": "A potential problem of (2) was that the normalized figures were affected by each participant's token number. The token number was calculated according to the results of automatic segmentation, so it certainly would be contaminated by segmentation errors. Therefore, the other approach (i.e. the z-score approach) was also adopted. Like the previous equation, in (3) was the participant j's frequency count of the ith word. was the mean of the participant's 218 word frequency counts, and was the standard deviation of those frequency counts.",
                "cite_spans": [
                    {
                        "start": 362,
                        "end": 365,
                        "text": "(3)",
                        "ref_id": "BIBREF2"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3.2.1"
            },
            {
                "text": "[",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3.2.1"
            },
            {
                "text": "Step four] The two types of personal word frequency were respectively put along with his/her response latencies in the lexical decision task for analysis. 11 ",
                "cite_spans": [
                    {
                        "start": 155,
                        "end": 157,
                        "text": "11",
                        "ref_id": "BIBREF10"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "3.2.1"
            },
            {
                "text": "Response errors in the lexical decision task (approximately 0.06% of the data set) were first screened. Two types of normalized personal word frequencies (i.e. ratio and z-score) were analyzed by mixed-effects models. Like the analysis in Experiment 1, in both models, two random factors and six covariates were also included. Random factors encompassed experiment stimuli and participants. Covariates were procedure variables (i.e. block number and trial number) and word variables (i.e. types of word frequency, sense number, character number, and neighborhood size). The covariates were subsumed in order to avoid mis-attributing the variances caused by procedure and word variables to the effect of personal word frequency. If there was any covariate not reaching significance, which meant it statistically did not affect the lexical-decision responses, it would be removed from the analysis and the other variables refitted the mixed models.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Results and Discussion",
                "sec_num": "3.2.2"
            },
            {
                "text": "The residuals of the two models, however, showed marked non-normality, especially at the end of long response latencies (see the upper right panel in Figure 5 ) 12 . To attenuate the unfitness, outliers with standardized residuals outside the interval (-2.5, 2.5) were removed. The removed data in both the ratio and z-score models were 2.48% of the data set. After trimming the outliers, we refitted the models. The residuals in the trimmed models were close to normality, as shown in the lower right panel of Figure 5 .",
                "cite_spans": [
                    {
                        "start": 161,
                        "end": 163,
                        "text": "12",
                        "ref_id": "BIBREF11"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 150,
                        "end": 158,
                        "text": "Figure 5",
                        "ref_id": "FIGREF4"
                    },
                    {
                        "start": 511,
                        "end": 519,
                        "text": "Figure 5",
                        "ref_id": "FIGREF4"
                    }
                ],
                "eq_spans": [],
                "section": "Results and Discussion",
                "sec_num": "3.2.2"
            },
            {
                "text": "Statistical results showed that personal word frequency significantly accounted for response latencies in both the analyses of frequency ratio (p < .001) and z-score (p < .05). The estimates of them were negative, which are visualized in Figure 6 . According to the figures, the negative estimates indicated that participants responded faster to stimuli with higher personal word frequencies. The experimental results revealed that IDs of frequencies of stimuli could explain individual variances between participants in lexical decision.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 238,
                        "end": 246,
                        "text": "Figure 6",
                        "ref_id": "FIGREF5"
                    }
                ],
                "eq_spans": [],
                "section": "Results and Discussion",
                "sec_num": "3.2.2"
            },
            {
                "text": "Words that frequency occurred in one's Facebook data revealed the things or issues he/she paid closer attention, the words he/she got accustomed to use but was unaware of, or his/her daily-life surroundings. Therefore, the effect of personal word frequencies in this experiment was considered to result from people's conscious or subconscious familiarity with words or concepts. The familiarity with word form and meaning facilitated the access to corresponding underlying lexical representations in the participants' mental lexicon.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Results and Discussion",
                "sec_num": "3.2.2"
            },
            {
                "text": "Another discussion brought up in this experiment is a methodological issue of computing personal lexical behaviors. Among two types of normalization of personal word frequency counts, the ratio method was assumed to be possibly problematic since segmentation errors were involved, and the z-score method was hypothesized to be a better one. Nevertheless, the analyses of word frequency ratio and z-score both reached significance. This indicated that normalizing frequency counts by the token number in each personal corpus is feasible even though there are segmentation errors and noise among the tokens. Evidence can be found when we compare each participant's total token number, which includes segmentation errors, with his token number summed from the 218 stimuli in Experiment 2, which includes no errors. The two categories of token numbers are highly correlated (r = .95). The correlation suggests that although segmentation errors make the total token numbers of Facebook data imprecise and inaccurate, the numbers still generally reflect the comparative differences between participants' genuine token numbers. ",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Results and Discussion",
                "sec_num": "3.2.2"
            },
            {
                "text": "By integrating the approach of computational linguistics into a psycholinguistic experiment, the current study sheds a new light on methods of capturing the nature of IDs in word recognition. The interdisciplinary effort testified that the quantified personal lexical Proceedings of the Twenty-Fourth Conference on Computational Linguistics and Speech Processing (ROCLING 2012) behaviors were associated with word recognition, thus uncovering a territory to be explored. One promising prospect of this study is that as the methodology of measuring lexical behaviors grows mature in the future, the readily available data of language usage, like Facebook posts, can function as convenient and valid resources for researchers to control the participant factors.",
                "cite_spans": [
                    {
                        "start": 363,
                        "end": 377,
                        "text": "(ROCLING 2012)",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Conclusion",
                "sec_num": "4."
            },
            {
                "text": "Furthermore, through the comparison of experimental results, the present study made a preliminary exploration on the methodology of measuring lexical behaviors and suggests the relatively appropriate methods. The counter-prediction finding in the frequency index experiment was possibly attributed to that the Sinica Corpus mainly consists of written data; therefore, it is suggested that similar experiments in future research resort to the frequency counts in a spoken corpus. Additionally, according to our examination, a person's total token number is feasible for normalizing his/her frequency counts even though word segmentation errors were contained within the tokens. Finally, when naturalistic data like the Facebook posts are utilized for the measurement, it is recommend basing the computation on personal preference or pattern of lexical usage (e.g. Experiment 2), instead of on every single word in one's language usage data (Experiment 1).",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Conclusion",
                "sec_num": "4."
            },
            {
                "text": "The present study examines word recognition by only concentrating on the lexical decision task. To obtain a clearer picture of the IDs in recognition, the future work can collect converging evidence from other types of extensively-used tasks, such as the naming task [16, 17] . Besides, this preliminary research recruited 16 participants. It is expected that when the number of participants increases in future research, it might give us other or deeper insight into the issue of individual differences (IDs). Moreover, in the Chinese Lexicon Profile (CLP) corpus mentioned in Section 2.1.2, there provides a great number of characteristics of words per se. Researchers may try to compute and explore individual lexical behaviors from the available characteristics, aside from the word frequency which is utilized in this study. In the respect of personal language usage data, we are constructing i-Corpus, which will comprise individualized corpora. A corpus per person will include various types of his/her language usage data, which can be looked into in the future so as to uncover multiple facets of personal language usage.",
                "cite_spans": [
                    {
                        "start": 267,
                        "end": 271,
                        "text": "[16,",
                        "ref_id": "BIBREF15"
                    },
                    {
                        "start": 272,
                        "end": 275,
                        "text": "17]",
                        "ref_id": "BIBREF16"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Future Work",
                "sec_num": "5."
            },
            {
                "text": "Proceedings of the Twenty-Fourth Conference on Computational Linguistics and Speech Processing(ROCLING 2012)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "Regularity denotes that the extent to which the spelling-to-sound correspondence in words are invariant. The effects of regularity are that a response is made slower to less 'regular' words (e.g. pint) than to 'regular' words (e.g. name).Proceedings of the Twenty-Fourth Conference on Computational Linguistics and Speech Processing(ROCLING 2012)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "http://www.facebook.com/3 The lexical decision task is an extensively-used experiment of visual word recognition.Proceedings of the Twenty-Fourth Conference on Computational Linguistics and Speech Processing (ROCLING 2012)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "The Chinese Lexicon Profile (CLP) is a research project launched at LOPE lab at National Taiwan University. The project purports to build up a large-scaled open lexical database platform for Chinese mono-syllabic to tri-syllabic words used in Taiwan. With its incorporation of behavioral and normative data in the long term, the CLP would allow researchers across various disciplines to explore different statistical models in search for the determinant variables that influence lexical processing tasks, as well as the training and verification of computational simulation studies. The number of Chinese words in CLP has been accumulated up to 204,922 so far.Proceedings of the Twenty-Fourth Conference on Computational Linguistics and Speech Processing(ROCLING 2012)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "i-Corpus is an on-going NSC-granted research project conducted at the LOPE lab, National Taiwan University. This project envisions an effort to construct i-corpora so as to obtain and analyze a wide spectrum of individual linguistic and extra-linguistic data. Considering the collected material is restricted by some copyright issues, a set of iCorpus toolkits is proposed which performs the tasks of autonomous corpus data collection and exploitation (by running an integrated software package) to extract, analyze huge volumes of individual language usage data, and automatically provide an idiolect sketch with quantitative information for the benefits of linguistic and above all, sociolinguistic studies.6 http://ckipsvr.iis.sinica.edu.tw/ Proceedings of the Twenty-Fourth Conference on Computational Linguistics and Speech Processing(ROCLING 2012)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "http://db1x.sinica.edu.tw/kiwi/mkiwi/ Proceedings of the Twenty-Fourth Conference on Computational Linguistics and Speech Processing (ROCLING 2012)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "http://www.r-project.org/ 9 Figure 3 displays the residuals of the model fitted by the values computed from the Intact word list. The plot of residuals in the NV list model is not demonstrated because it was the same as Figure 3.(1)Proceedings of the Twenty-Fourth Conference on Computational Linguistics and Speech Processing (ROCLING 2012)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "LDT refers to the lexical decision task in this paper.(2) (3)Proceedings of the Twenty-Fourth Conference on Computational Linguistics and Speech Processing (ROCLING 2012)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "Unlike Experiment 1, the response accuracies were not analyzed in this experiment. It was because the accuracy of the 218 stimuli here was extremely high (99.4%).12 Figure 5is the residuals of the model fitted by the personal word frequency ratios. The residuals of the z-score model are the same as those of the ratio model, so its residual plot is not given here.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            }
        ],
        "back_matter": [],
        "bib_entries": {
            "BIBREF0": {
                "ref_id": "b0",
                "title": "Frequency and neighborhood effects on lexical access: Activation or search?",
                "authors": [
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Andrews",
                        "suffix": ""
                    }
                ],
                "year": 1989,
                "venue": "Journal of Experimental Psychology: Learning, Memory, and Cognition",
                "volume": "15",
                "issue": "",
                "pages": "802--814",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "S. Andrews, \"Frequency and neighborhood effects on lexical access: Activation or search?,\" Journal of Experimental Psychology: Learning, Memory, and Cognition, vol. 15, pp. 802-814, 1989.",
                "links": null
            },
            "BIBREF1": {
                "ref_id": "b1",
                "title": "Lexical access and naming time",
                "authors": [
                    {
                        "first": "K",
                        "middle": [
                            "I"
                        ],
                        "last": "Forster",
                        "suffix": ""
                    },
                    {
                        "first": "S",
                        "middle": [
                            "M"
                        ],
                        "last": "Chambers",
                        "suffix": ""
                    }
                ],
                "year": 1973,
                "venue": "Journal of Verbal Learning and Verbal Behavior",
                "volume": "12",
                "issue": "",
                "pages": "627--635",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "K. I. Forster and S. M. Chambers, \"Lexical access and naming time,\" Journal of Verbal Learning and Verbal Behavior, vol. 12, pp. 627-635, 1973.",
                "links": null
            },
            "BIBREF2": {
                "ref_id": "b2",
                "title": "Word frequency and neighborhood frequency effects in lexical decision and naming",
                "authors": [
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Grainger",
                        "suffix": ""
                    }
                ],
                "year": 1990,
                "venue": "Journal of Memory and Language",
                "volume": "29",
                "issue": "",
                "pages": "228--244",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "J. Grainger, \"Word frequency and neighborhood frequency effects in lexical decision and naming,\" Journal of Memory and Language, vol. 29, pp. 228-244, 1990.",
                "links": null
            },
            "BIBREF3": {
                "ref_id": "b3",
                "title": "Reexamining word length effects in visual word recognition: New evidence from the English Lexicon Project",
                "authors": [
                    {
                        "first": "B",
                        "middle": [],
                        "last": "New",
                        "suffix": ""
                    }
                ],
                "year": 2006,
                "venue": "Psychonomic Bulletin & Review",
                "volume": "13",
                "issue": "",
                "pages": "45--52",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "B. New, et al., \"Reexamining word length effects in visual word recognition: New evidence from the English Lexicon Project,\" Psychonomic Bulletin & Review, vol. 13, pp. 45-52, 2006.",
                "links": null
            },
            "BIBREF4": {
                "ref_id": "b4",
                "title": "Word-nonword classification time",
                "authors": [
                    {
                        "first": "C",
                        "middle": [
                            "P"
                        ],
                        "last": "Whaley",
                        "suffix": ""
                    }
                ],
                "year": 1978,
                "venue": "Journal of Verbal Learning and Proceedings of the Twenty-Fourth Conference on Computational Linguistics and Speech Processing (ROCLING 2012) Verbal Behavior",
                "volume": "17",
                "issue": "",
                "pages": "143--154",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "C. P. Whaley, \"Word-nonword classification time,\" Journal of Verbal Learning and Proceedings of the Twenty-Fourth Conference on Computational Linguistics and Speech Processing (ROCLING 2012) Verbal Behavior, vol. 17, pp. 143-154, 1978.",
                "links": null
            },
            "BIBREF5": {
                "ref_id": "b5",
                "title": "Exposure to print and word recognition processes",
                "authors": [
                    {
                        "first": "D",
                        "middle": [],
                        "last": "Chateau",
                        "suffix": ""
                    },
                    {
                        "first": "D",
                        "middle": [],
                        "last": "Jared",
                        "suffix": ""
                    }
                ],
                "year": 2000,
                "venue": "Memory & Cognition",
                "volume": "28",
                "issue": "",
                "pages": "143--153",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "D. Chateau and D. Jared, \"Exposure to print and word recognition processes,\" Memory & Cognition, vol. 28, pp. 143-153, 2000.",
                "links": null
            },
            "BIBREF6": {
                "ref_id": "b6",
                "title": "Is there an effect of print exposure on the word frequency effect and the neighborhood size effect?",
                "authors": [
                    {
                        "first": "C",
                        "middle": [],
                        "last": "Sears",
                        "suffix": ""
                    }
                ],
                "year": 2008,
                "venue": "Journal of Psycholinguistic Research",
                "volume": "37",
                "issue": "",
                "pages": "269--291",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "C. Sears, et al., \"Is there an effect of print exposure on the word frequency effect and the neighborhood size effect?,\" Journal of Psycholinguistic Research, vol. 37, pp. 269-291, 2008.",
                "links": null
            },
            "BIBREF7": {
                "ref_id": "b7",
                "title": "The impact of reader skill on phonological processing in visual word recognition",
                "authors": [
                    {
                        "first": "S",
                        "middle": [
                            "J"
                        ],
                        "last": "Unsworth",
                        "suffix": ""
                    },
                    {
                        "first": "P",
                        "middle": [
                            "M"
                        ],
                        "last": "Pexman",
                        "suffix": ""
                    }
                ],
                "year": 2003,
                "venue": "Quarterly Journal of Experimental Psychology",
                "volume": "56",
                "issue": "",
                "pages": "63--81",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "S. J. Unsworth and P. M. Pexman, \"The impact of reader skill on phonological processing in visual word recognition,\" Quarterly Journal of Experimental Psychology, vol. 56A, pp. 63-81, 2003.",
                "links": null
            },
            "BIBREF8": {
                "ref_id": "b8",
                "title": "Lexical familiarity and processing efficiency: Individual differences in naming, lexical decision, and semantic categorization",
                "authors": [
                    {
                        "first": "M",
                        "middle": [
                            "J"
                        ],
                        "last": "Lewellen",
                        "suffix": ""
                    }
                ],
                "year": 1993,
                "venue": "Journal of Experimental Psychology: General",
                "volume": "122",
                "issue": "",
                "pages": "316--330",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "M. J. Lewellen, et al., \"Lexical familiarity and processing efficiency: Individual differences in naming, lexical decision, and semantic categorization,\" Journal of Experimental Psychology: General, vol. 122, pp. 316-330, 1993.",
                "links": null
            },
            "BIBREF9": {
                "ref_id": "b9",
                "title": "What lexical decision and naming tell us about reading",
                "authors": [
                    {
                        "first": "L",
                        "middle": [],
                        "last": "Katz",
                        "suffix": ""
                    }
                ],
                "year": null,
                "venue": "Reading and Writing",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "L. Katz, et al., \"What lexical decision and naming tell us about reading,\" Reading and Writing, in press.",
                "links": null
            },
            "BIBREF10": {
                "ref_id": "b10",
                "title": "Individual differences in visual word recognition: Insights from the English Lexicon Project",
                "authors": [
                    {
                        "first": "M",
                        "middle": [
                            "J"
                        ],
                        "last": "Yap",
                        "suffix": ""
                    }
                ],
                "year": 2012,
                "venue": "Journal of Experimental Psychology: Human Perception and Performance",
                "volume": "38",
                "issue": "",
                "pages": "53--79",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "M. J. Yap, et al., \"Individual differences in visual word recognition: Insights from the English Lexicon Project,\" Journal of Experimental Psychology: Human Perception and Performance, vol. 38, pp. 53-79, 2012.",
                "links": null
            },
            "BIBREF11": {
                "ref_id": "b11",
                "title": "The assessment and analysis of handedness: The Edinburgh inventory",
                "authors": [
                    {
                        "first": "R",
                        "middle": [
                            "C"
                        ],
                        "last": "Oldfield",
                        "suffix": ""
                    }
                ],
                "year": 1971,
                "venue": "Neuropsychologia",
                "volume": "9",
                "issue": "",
                "pages": "97--113",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "R. C. Oldfield, \"The assessment and analysis of handedness: The Edinburgh inventory,\" Neuropsychologia, vol. 9, pp. 97-113, 1971.",
                "links": null
            },
            "BIBREF12": {
                "ref_id": "b12",
                "title": "Word frequency and word difficulty: A comparison of counts in four corpora",
                "authors": [
                    {
                        "first": "H",
                        "middle": [
                            "M"
                        ],
                        "last": "Breland",
                        "suffix": ""
                    }
                ],
                "year": 1996,
                "venue": "Psychological Science",
                "volume": "7",
                "issue": "",
                "pages": "96--99",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "H. M. Breland, \"Word frequency and word difficulty: A comparison of counts in four corpora,\" Psychological Science, vol. 7, pp. 96-99, 1996.",
                "links": null
            },
            "BIBREF13": {
                "ref_id": "b13",
                "title": "Statistical computing: An introdution to data analysis using S-plus",
                "authors": [
                    {
                        "first": "M",
                        "middle": [
                            "J"
                        ],
                        "last": "Crawley",
                        "suffix": ""
                    }
                ],
                "year": 2002,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "M. J. Crawley, Statistical computing: An introdution to data analysis using S-plus. Chichester: Wiley, 2002.",
                "links": null
            },
            "BIBREF14": {
                "ref_id": "b14",
                "title": "Analyzing linguistic data : A practical introduction to statistics using R",
                "authors": [
                    {
                        "first": "R",
                        "middle": [
                            "H"
                        ],
                        "last": "Baayen",
                        "suffix": ""
                    }
                ],
                "year": 2008,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "R. H. Baayen, Analyzing linguistic data : A practical introduction to statistics using R. Cambridge: Cambridge University Press, 2008.",
                "links": null
            },
            "BIBREF15": {
                "ref_id": "b15",
                "title": "Are lexical decisions a good measure of lexical access? The role of word frequency in the neglected decision stage",
                "authors": [
                    {
                        "first": "D",
                        "middle": [
                            "A"
                        ],
                        "last": "Balota",
                        "suffix": ""
                    },
                    {
                        "first": "J",
                        "middle": [
                            "I"
                        ],
                        "last": "Chumbley",
                        "suffix": ""
                    }
                ],
                "year": 1984,
                "venue": "Journal of Experimental Psychology: Human Perception and Performance",
                "volume": "10",
                "issue": "",
                "pages": "340--357",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "D. A. Balota and J. I. Chumbley, \"Are lexical decisions a good measure of lexical access? The role of word frequency in the neglected decision stage,\" Journal of Experimental Psychology: Human Perception and Performance, vol. 10, pp. 340-357, 1984.",
                "links": null
            },
            "BIBREF16": {
                "ref_id": "b16",
                "title": "The locus of word frequency effects in the pronunciation task: Lexical access and/or production?",
                "authors": [
                    {
                        "first": "D",
                        "middle": [
                            "A"
                        ],
                        "last": "Balota",
                        "suffix": ""
                    },
                    {
                        "first": "J",
                        "middle": [
                            "I"
                        ],
                        "last": "Chumbley",
                        "suffix": ""
                    }
                ],
                "year": 1985,
                "venue": "Journal of Memory and Language",
                "volume": "24",
                "issue": "",
                "pages": "89--106",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "D. A. Balota and J. I. Chumbley, \"The locus of word frequency effects in the pronunciation task: Lexical access and/or production?,\" Journal of Memory and Language, vol. 24, pp. 89-106, 1985.",
                "links": null
            }
        },
        "ref_entries": {
            "FIGREF0": {
                "uris": null,
                "num": null,
                "type_str": "figure",
                "text": "The procedure for random generation of two-character non-word stimuli in the visual lexical decision task 2.1.3 Procedure"
            },
            "FIGREF1": {
                "uris": null,
                "num": null,
                "type_str": "figure",
                "text": "Step two] Paste the access token in the i-Corpus program[Step three] Type in a participant's Facebook ID [Step four] Save the data on the participant's Facebook Wall (JSON format) ["
            },
            "FIGREF2": {
                "uris": null,
                "num": null,
                "type_str": "figure",
                "text": "Residual diagnostics for the models of the Intact list before (upper panels) and after (lower panels) removal of outliers"
            },
            "FIGREF3": {
                "uris": null,
                "num": null,
                "type_str": "figure",
                "text": "Continuum of modes in Facebook posts"
            },
            "FIGREF4": {
                "uris": null,
                "num": null,
                "type_str": "figure",
                "text": "Residual diagnostics for the model of personal word frequency ratios before (upper panels) and after (lower panels) removal of outliers"
            },
            "FIGREF5": {
                "uris": null,
                "num": null,
                "type_str": "figure",
                "text": "Partial effects of personal word frequency (ratio and z-score) in the analysis of Experiment 2"
            },
            "TABREF0": {
                "type_str": "table",
                "html": null,
                "content": "<table><tr><td>Subject01</td><td>12506 Subject09</td><td>7487</td></tr><tr><td>Subject02</td><td>2765 Subject10</td><td>7690</td></tr><tr><td>Subject03</td><td>2144 Subject11</td><td>4727</td></tr><tr><td>Subject04</td><td>3590 Subject12</td><td>4389</td></tr><tr><td>Subject05</td><td>8251 Subject13</td><td>5908</td></tr><tr><td>Subject06</td><td>3442 Subject14</td><td>18636</td></tr><tr><td>Subject07</td><td>4293 Subject15</td><td>985</td></tr><tr><td>Subject08</td><td>2960 Subject16</td><td>2260</td></tr></table>",
                "text": "The token numbers in participants' Facebook posts Subject Chinese Token Number Subject Chinese Token Number",
                "num": null
            },
            "TABREF1": {
                "type_str": "table",
                "html": null,
                "content": "<table><tr><td>Word</td><td>Personal word</td><td>Sinica word</td></tr><tr><td/><td>frequency</td><td>frequency</td></tr><tr><td/><td>12</td><td>48749</td></tr><tr><td/><td>4</td><td>7582</td></tr><tr><td/><td>2</td><td>3280</td></tr><tr><td/><td>1</td><td>NA</td></tr><tr><td/><td>1</td><td>NA</td></tr><tr><td/><td>1</td><td>NA</td></tr></table>",
                "text": "An example of a portion of one participant's word list",
                "num": null
            }
        }
    }
}