File size: 91,265 Bytes
6fa4bc9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
{
    "paper_id": "2020",
    "header": {
        "generated_with": "S2ORC 1.0.0",
        "date_generated": "2023-01-19T14:57:56.114043Z"
    },
    "title": "Combining Expert Knowledge with Frequency Information to Infer CEFR Levels for Words",
    "authors": [
        {
            "first": "Alice",
            "middle": [],
            "last": "Pintard",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "Le Mans Universit\u00e9",
                "location": {
                    "settlement": "Le Mans",
                    "country": "France"
                }
            },
            "email": "alice.pintard.etu@univ-lemans.fr"
        },
        {
            "first": "Thomas",
            "middle": [],
            "last": "Fran\u00e7ois",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "University of Louvain",
                "location": {
                    "settlement": "Louvain-la-Neuve",
                    "country": "Belgique"
                }
            },
            "email": "thomas.francois@uclouvain.be"
        }
    ],
    "year": "",
    "venue": null,
    "identifiers": {},
    "abstract": "Traditional approaches to set goals in second language (L2) vocabulary acquisition relied either on word lists that were obtained from large L1 corpora or on collective knowledge and experience of L2 experts, teachers, and examiners. Both approaches are known to offer some advantages, but also to have some limitations. In this paper, we try to combine both sources of information, namely the official reference level description for French language and the FLElex lexical database. Our aim is to train a statistical model on the French RLD that would be able to turn the distributional information from FLElex into one of the six levels of the Common European Framework of Reference for languages (CEFR). We show that such approach yields a gain of 29% in accuracy compared to the method currently used in the CEFRLex project. Besides, our experiments also offer deeper insights into the advantages and shortcomings of the two traditional sources of information (frequency vs. expert knowledge).",
    "pdf_parse": {
        "paper_id": "2020",
        "_pdf_hash": "",
        "abstract": [
            {
                "text": "Traditional approaches to set goals in second language (L2) vocabulary acquisition relied either on word lists that were obtained from large L1 corpora or on collective knowledge and experience of L2 experts, teachers, and examiners. Both approaches are known to offer some advantages, but also to have some limitations. In this paper, we try to combine both sources of information, namely the official reference level description for French language and the FLElex lexical database. Our aim is to train a statistical model on the French RLD that would be able to turn the distributional information from FLElex into one of the six levels of the Common European Framework of Reference for languages (CEFR). We show that such approach yields a gain of 29% in accuracy compared to the method currently used in the CEFRLex project. Besides, our experiments also offer deeper insights into the advantages and shortcomings of the two traditional sources of information (frequency vs. expert knowledge).",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Abstract",
                "sec_num": null
            }
        ],
        "body_text": [
            {
                "text": "Second language acquisition (SLA) research established a strong relationship between the development of reading abilities and the knowledge of vocabulary (Laufer, 1992) . For Grabe (2014, 13) : \"The real goal for more advanced L2 reading is an L2 recognition vocabulary level anywhere above 10,000 [words]\". It is no surprise that vocabulary resources used by designers of L2 curricula, publishers of educational materials, and teachers to set vocabulary learning goals are close to such size. For French language, the popular \"Fran\u00e7ais Fondamental\" (Gougenheim et al., 1964) , which was built from a corpus of authentic documents and influenced a whole generation of French teachers and SLA researchers, includes about 8800 words. Similarly, the currently most popular lexical resources are the Reference Level Descriptions (RLDs), based on the Common European Framework of Reference for languages (CEFR), and available in various languages. The French version, designed by a team of experts, also amounts to about 9,000 words and expressions. However, both type of lists -either built from language data or from the expertise of language and teaching experts -are faced with the issue of identifying the most important words to teach at each stage of the learning process.",
                "cite_spans": [
                    {
                        "start": 154,
                        "end": 168,
                        "text": "(Laufer, 1992)",
                        "ref_id": "BIBREF13"
                    },
                    {
                        "start": 175,
                        "end": 191,
                        "text": "Grabe (2014, 13)",
                        "ref_id": null
                    },
                    {
                        "start": 550,
                        "end": 575,
                        "text": "(Gougenheim et al., 1964)",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1."
            },
            {
                "text": "The most common answers to that challenge have been (1) to use frequency lists obtained from a large corpus of texts intended for native readers and split the list into N frequency bands, each of which is related to one of the stage of the learning process; or (2) to rely on expert knowledge, such as teacher expertise or linguists' recommendations, to assign each word to a given level of reading proficiency. This classification of words in developmental stages is a delicate process whose reliability has hardly been assessed in a systematic manner on L2 learners. Besides, the two main sources of information to build vocabulary lists -word frequency in massive corpora or the knowledge of L2 teach-ing experts -have hardly been exploited together 1 . Recently, an alternative research avenue was investigated within the framework of the CEFRLex project. It offers receptive vocabulary lists for 5 languages: English (D\u00fcrlich and Fran\u00e7ois, 2018) , French (Fran\u00e7ois et al., 2014) , Swedish (Fran\u00e7ois et al., 2016) , Dutch (Tack et al., 2018) , and Spanish. Its innovative side resides in the fact that it does not provide a single frequency for each word, but rather a frequency distribution across the six levels of the CEFR. Moreover, frequencies have been estimated on documents intended for L2 learners, i.e. textbooks and simplified readers, instead of L1 texts. As a result, the resource provides further insights about the way a given word is used across the various development stages of the L2 curriculum. It is also possible to compare word frequency at a given level (e.g. A2) in order to define priorities in terms of vocabulary learning goals. Unfortunately, when it comes to assigning a CEFR level at which a given word should be learned, it is not obvious how the frequency distributions should be transformed in a single CEFR level. In this paper, we aim to investigate two main issues. First, we will test whether we can leverage the knowledge from the French RLD to train a mathematical function, based on machine learning algorithms, able to transform any CEFR-Lex distribution into a CEFR level. Second, we will take advantage of these experiments to further characterize the linguistic and pedagogical differences between these two approaches -building a frequency list from a corpus vs. assigning words to proficiency levels based on expert knowledge -to set vocabulary learning goals. The paper is organized as follows: Section 2. provides more details about the two approaches we will compare (frequency lists and RLD) and reports previous attempts to transform CEFR frequency distributions into a unique CEFR level. Section 3. introduces all methodological details related to our experiments: the three lexical resources used in the study (French RLD, Lexique3, and FLELex) and the process by which these resources were prepared for training machine learning algorithms. This section ends up with the description of the experiments carried out. In Section 4., we report the results of the various experiments before taking advantage of a manual error analysis to discuss the differences between expert knowledge and frequency-based lists at Section 5..",
                "cite_spans": [
                    {
                        "start": 922,
                        "end": 950,
                        "text": "(D\u00fcrlich and Fran\u00e7ois, 2018)",
                        "ref_id": "BIBREF6"
                    },
                    {
                        "start": 960,
                        "end": 983,
                        "text": "(Fran\u00e7ois et al., 2014)",
                        "ref_id": "BIBREF7"
                    },
                    {
                        "start": 994,
                        "end": 1017,
                        "text": "(Fran\u00e7ois et al., 2016)",
                        "ref_id": "BIBREF8"
                    },
                    {
                        "start": 1026,
                        "end": 1045,
                        "text": "(Tack et al., 2018)",
                        "ref_id": "BIBREF19"
                    },
                    {
                        "start": 2768,
                        "end": 2780,
                        "text": "(French RLD,",
                        "ref_id": null
                    },
                    {
                        "start": 2781,
                        "end": 2790,
                        "text": "Lexique3,",
                        "ref_id": null
                    },
                    {
                        "start": 2791,
                        "end": 2802,
                        "text": "and FLELex)",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1."
            },
            {
                "text": "The process of setting vocabulary goals for L2 learners generally relies on graded lexical resources in which words are assigned to one proficiency level. Such resources are usually built based on the two approaches we have previously outlined: leveraging word frequency information estimated on a corpus or using L2 teaching experts knowledge. Frequency lists, built from a corpus, have been used since the seminal work of Thorndike (1921) who laboriously built the first significant vocabulary for English, including 20,000 words, without the help of any computer. The first computational list was obtained by Ku\u010dera and Francis (1967) from the Brown corpus and has a large influence in education and psychology. At the same time, Gougenheim et al. (1964) published the Fran\u00e7ais Fondamental that would impact a generation of L2 French teachers. More recently, other lists have been developed from larger corpora, such as the CELEX database (Baayen et al., 1993) , the list based on the British National Corpus (Leech et al., 2001) , or SUBTLEX (Brysbaert and New, 2009) . The main shortcomings of such lists for L2 education are that (1) they represent the native distribution of words, which is not fully compatible with the distribution of words in books and textbooks intended for L2 learners; (2) they do not specify at which proficiency level a given word is supposed to be learned. As regards expert knowledge, the most recent and influential resource is connected to the CEFR framework. Since 2001, this framework has been widely adopted within Europe to help standardizing L2 curricula, which involves defining a proficiency scale ranging from A1 (beginners) to C2 (mastery). However, textbook designers, assessment experts and language teachers have agreed that it lacks precision when it comes to describing the linguistic forms that should be learned at a given proficiency level. In a number of countries, efforts have been made to interpret the CEFR guidelines in the form of reference level descriptions 2 . These books describe the language competences expected from an L2 learner in each of the CEFR levels, including lists of words, syntactic structures, and expressions associated with specific communicative functions or themes. Finally, a few papers specifically investigated methods to transform CEFRLex word distribution into a CEFR level",
                "cite_spans": [
                    {
                        "start": 424,
                        "end": 440,
                        "text": "Thorndike (1921)",
                        "ref_id": "BIBREF20"
                    },
                    {
                        "start": 612,
                        "end": 637,
                        "text": "Ku\u010dera and Francis (1967)",
                        "ref_id": "BIBREF12"
                    },
                    {
                        "start": 733,
                        "end": 757,
                        "text": "Gougenheim et al. (1964)",
                        "ref_id": null
                    },
                    {
                        "start": 942,
                        "end": 963,
                        "text": "(Baayen et al., 1993)",
                        "ref_id": "BIBREF1"
                    },
                    {
                        "start": 1012,
                        "end": 1032,
                        "text": "(Leech et al., 2001)",
                        "ref_id": null
                    },
                    {
                        "start": 1046,
                        "end": 1071,
                        "text": "(Brysbaert and New, 2009)",
                        "ref_id": "BIBREF4"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Previous work",
                "sec_num": "2."
            },
            {
                "text": "int/t/dg4/linguistic/dnr_EN.asp?",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Previous work",
                "sec_num": "2."
            },
            {
                "text": "coherent from a pedagogical perspective. (Gala et al., 2013) suggested two approaches. The first one, to which we will refer as First occ, assigns to a given word the level of the textbook it was first observed in. In other words, the level of a word corresponds to the first CEFR level for which FLELex reports a non null frequency. Although simplistic, this rule appeared to be the most effective to predict unknown words reported by four Dutch learners of FFL (Tack et al., 2016) and was consequently used in the CEFRLex interface. The second approach was a variant of the First occ that yields continuous scores and prove to be inferior to the first one. More recently, Alfter et al. (2016) introduced the concept of significant onset of use, that consists in selecting the first level having a sufficiently large enough delta compared to its previous level. All of these studies used mathematical rules to transform distribution into CEFR levels and later use those level as gold-standard for further process. So far, no experiments were reported that tried to cross-validate such mathematical rules, for instance using learners data.",
                "cite_spans": [
                    {
                        "start": 41,
                        "end": 60,
                        "text": "(Gala et al., 2013)",
                        "ref_id": "BIBREF9"
                    },
                    {
                        "start": 463,
                        "end": 482,
                        "text": "(Tack et al., 2016)",
                        "ref_id": "BIBREF18"
                    },
                    {
                        "start": 674,
                        "end": 694,
                        "text": "Alfter et al. (2016)",
                        "ref_id": "BIBREF0"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Previous work",
                "sec_num": "2."
            },
            {
                "text": "Our approach consists in considering the French RLD as a gold standard regarding the assignment of words to a given CEFR level. We then infer, from this pedagogical information, a statistical model able to transform the word frequency distribution from FLELex into a CEFR level. To carry out this experiment, the following steps had to be realized. The acquisition and digitization of the French RLD word list is described at Section 3.1., which also briefly reminds the reader of the main characteristics of the two other lexical resources used in our study, namely Lexique3 (New et al., 2007) and FLELex (Fran\u00e7ois et al., 2014) . In the next section (Section 3.2.), we describe a preliminary step prior to the statistical modelling, which consists in delineating the intersection between the three resources. This stage aims at ensuring that missing words would not lead to results biased towards one of the resources. We also took advantage of this step to investigate the coverage discrepancies between the French RLD and FLELex as a first way to characterize the differences between the expert knowledge approach and the frequency-based one. Section 3.3. describes the design of two datasets used for our experiments, whereas Section 3.4. presents the different baselines and models tested.",
                "cite_spans": [
                    {
                        "start": 576,
                        "end": 594,
                        "text": "(New et al., 2007)",
                        "ref_id": "BIBREF15"
                    },
                    {
                        "start": 606,
                        "end": 629,
                        "text": "(Fran\u00e7ois et al., 2014)",
                        "ref_id": "BIBREF7"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Methodology",
                "sec_num": "3."
            },
            {
                "text": "3.1.1. The French RLD word list The RLD for French language was created by Beacco and his collaborators between 2004 and 2016 (Beacco et al., 2008; Riba, 2016) . Each level -corresponding to a distinct book -is split into 10 chapters representing various dimensions of the linguistic knowledge (e.g. vocabulary, syntactic structures, phonemes, graphemes, fonctional skills, etc.), except for C1 and C2 levels which share the same volume and have a different structure 3 . The classification of linguistic forms to a given level was performed based on crite- ria selected by the authors for their relevance and objectivity: essentially the official descriptors from the CEFR, collective knowledge and experience of experts, teachers and examiners, and examples of learner productions deemed to be at a particular level (Beacco et al., 2008) . To our knowledge, the French RLD, also refered to as \"Beacco\" in this study, has not been used so far in any NLP approaches as it was published in paper format only and is not available in a digitized version. As a consequence, we had to digitize the two chapters relative to lexicon, namely chapter 4, focusing on general notions (e.g. quantity, space), and chapter 6 that focuses on specific notions (e.g. human body, feelings, sports). Those chapters share the same architecture across all levels, organizing words within semantic categories, then specifying the part-of-speech (POS) categories and sometimes providing a context. Polysemous words can therefore have up to 8 entries across the four levels (e.g. \"\u00eatre\", to be). However, as FLELex and Lexique3 do not provide fine-grained semantic distinctions for forms (all meanings are gathered under the same orthographic form), we decided to drop the information on semantic category from the French RLD. When a form had several CEFR levels associated to it, we kept the lowest one, which is in line with the way polysemy is handled in FLELex. This process led us to drop about 2,968 entries, going from 8,486 to 5,518 entries. The number of entries per CEFR level is described in Table 1 (#Beacco).",
                "cite_spans": [
                    {
                        "start": 126,
                        "end": 147,
                        "text": "(Beacco et al., 2008;",
                        "ref_id": null
                    },
                    {
                        "start": 148,
                        "end": 159,
                        "text": "Riba, 2016)",
                        "ref_id": "BIBREF16"
                    },
                    {
                        "start": 818,
                        "end": 839,
                        "text": "(Beacco et al., 2008)",
                        "ref_id": null
                    }
                ],
                "ref_spans": [
                    {
                        "start": 2079,
                        "end": 2086,
                        "text": "Table 1",
                        "ref_id": "TABREF0"
                    }
                ],
                "eq_spans": [],
                "section": "Source Word Lists",
                "sec_num": "3.1."
            },
            {
                "text": "As previous approaches relying on word frequencies to assign proficiency levels to words relied on a L1 corpus, we decided to compare the performance obtained with FLELex with a word list whose frequencies were estimated on a large L1 corpus. We used Lexique3 (New et al., 2007) for this purpose, as it is a rather modern database. The lexicon includes about 50,000 lemmas and 125,000 inflected forms whose frequencies were obtained from movie subtitles.",
                "cite_spans": [
                    {
                        "start": 260,
                        "end": 278,
                        "text": "(New et al., 2007)",
                        "ref_id": "BIBREF15"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Lexique3 word list",
                "sec_num": "3.1.2."
            },
            {
                "text": "FLELex (Fran\u00e7ois et al., 2014) is one of the resources being part of the CEFRLex project described above. Similarly to the other languages, it offers frequency distributions for French words across the six CEFR levels. There are discourse in terms of rhetorical effectiveness, natural sequencing or adherence to collaborative principles.",
                "cite_spans": [
                    {
                        "start": 7,
                        "end": 30,
                        "text": "(Fran\u00e7ois et al., 2014)",
                        "ref_id": "BIBREF7"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "FLELex",
                "sec_num": "3.1.3."
            },
            {
                "text": "two versions of FLELex: one is based on the TreeTagger (FLELex-TT) and includes 14,236 entries, but no multiword expressions as they cannot be detected by the Tree-Tagger; the second one is based on a conditional random field (CRF) tagger and amounts to 17,871 entries, including 2,037 multi-word expressions. However, the second version has not yet been manually checked and includes various problematic forms. This is why we decided to carry out our experiments based on the FLELex-TT version. Table 1 summarizes the total number of entries having a non null frequency per level (#FLELex), along with the number of new entries per level, currently used in the CEFRLex project to assign a unique level to a given word (#First occ).",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "FLELex",
                "sec_num": "3.1.3."
            },
            {
                "text": "As explained above, in order to ensure a comparability of results for each of the three word lists, we delineated their intersection. A prerequisite step was to arrange the POS tagsets compatibility. The main differences regarding those tagsets are that Beacco divides conjunctions in two categories (coordination and subordination), whereas FLELex and Lexique3 split determiners and prepositions (DET:ART vs. DET:POS and PRP vs. PRP:det). We merged all split categories, keeping the TreeTagger labels (Schmid, 1994) . After standardization, nine POS remained: ADJ, ADV, KON, DET, INT, NOM, PRP, PRO, and VER. Second, we identified words in common between Beacco and FLELex: their intersection contains 4,020 entries. This leaves 1,498 words from Beacco that do not appear in FLELex and 10,216 FLELex words absent from Beacco. Such figures were expected as the coverage of FLELex is larger due to its building principles. However, we were concerned by the fact that so many words from Beacco were not found in FLELex and carried out a manual investigation of these. Most missing words can be related to the following causes:",
                "cite_spans": [
                    {
                        "start": 502,
                        "end": 516,
                        "text": "(Schmid, 1994)",
                        "ref_id": "BIBREF17"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Mapping the RLD to FLELex and Lexique3",
                "sec_num": "3.2."
            },
            {
                "text": "\u2022 Beacco includes 113 past participle forms of verbs that have not been lemmatized, whereas it is the case in FLELex (e.g. \"assis\" sat, \"\u00e9pic\u00e9\" seasoned);",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Mapping the RLD to FLELex and Lexique3",
                "sec_num": "3.2."
            },
            {
                "text": "\u2022 Similarly, Beacco also includes 103 feminine or plural forms which are lemmatized in FLELex (e.g. \"vacances\" holiday, \"lunettes\" glasses, \"serveuse\" waitress, etc.);",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Mapping the RLD to FLELex and Lexique3",
                "sec_num": "3.2."
            },
            {
                "text": "\u2022 Words were sometimes shared by both resources, but were assigned a different POS-tag, preventing automatic matching (e.g. \"bonjour\" hi ! or \"vite\" be quick are interjections in Beacco, but are tagged as nouns or adverbs in FLELex);",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Mapping the RLD to FLELex and Lexique3",
                "sec_num": "3.2."
            },
            {
                "text": "\u2022 61 entries were kept with capital letters in Beacco as a way to provide information about the word in use (e.g. \"Attention\" Look up !, \"Courage\" Cheer up !);",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Mapping the RLD to FLELex and Lexique3",
                "sec_num": "3.2."
            },
            {
                "text": "\u2022 Unlike Beacco, FLELex does not include acronyms (e.g.: \"CD\", \"DVD\", \"CV\", etc.);",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Mapping the RLD to FLELex and Lexique3",
                "sec_num": "3.2."
            },
            {
                "text": "\u2022 Some words were not captured in FLELex despite their presence in FFL textbooks, because they appear in the instructions, grammatical boxes, maps, or calendars rather than in texts related to comprehension tasks (e.g. \"fois\" time, \"adjectif\" adjective, \"virgule\" comma, \"Asie\" Asia, etc.);",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Mapping the RLD to FLELex and Lexique3",
                "sec_num": "3.2."
            },
            {
                "text": "\u2022 Other words refer to very salient objects in the real world that are poorly represented in corpora. Since Mich\u00e9a (1953) , they are known as available words and, as was expected, some of them were not found in the corpus used to build FLELex (e.g. \"cuisini\u00e8re\" cooker, \"s\u00e8che-cheveux\" hair-dryer, etc.);",
                "cite_spans": [
                    {
                        "start": 108,
                        "end": 121,
                        "text": "Mich\u00e9a (1953)",
                        "ref_id": "BIBREF14"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Mapping the RLD to FLELex and Lexique3",
                "sec_num": "3.2."
            },
            {
                "text": "\u2022 Finally, a few words in Beacco were extremely specific (e.g. \"humagne\", a type of grape or \"escrimeur\" fencer).",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Mapping the RLD to FLELex and Lexique3",
                "sec_num": "3.2."
            },
            {
                "text": "This manual investigation was, to some extent, reassuring, as a fair amount of missing words from Beacco were due to discrepancies in the lemmatization process between a systematic tool and a human. Lexical availability was also an issue, but a predictable one as it concerns all frequencybased approaches. Finally, it appears that restricting the selection of textbook materials to texts related to receptive tasks might help to better model receptive knowledge of L2 learners, but also comes at a cost as regards coverage. We manually solved some of these issues by lemmatizing the entries; converting the POS of all interjections that were nouns or adverbs in FLELex, and replacing capital letters by lowercase letters. In this process, we lost precious information from the RLD about the function of some linguistic forms, but were able to reintroduce 314 words that were not considered as shared by both lexicons before. As a result, the intersection between both resources amounts to 4,334 words. Finally, in order to compare FLELex with Lexique3, we computed the intersection between all three lexicons. Lexique3 having the larger coverage (51,100), there were only 38 words missing from it. The final intersection therefore includes 4,296 entries.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Mapping the RLD to FLELex and Lexique3",
                "sec_num": "3.2."
            },
            {
                "text": "Based on this intersection between the three resources, we defined two datasets that will be used for our experiments.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Preparing datasets for experiments",
                "sec_num": "3.3."
            },
            {
                "text": "This first dataset corresponds to the intersection between FLELex, Lexique3 and Beacco as defined at Section 3.2.. It contains 4,296 entries, shared by the three lexicons, and classified from A1 to B2 according to Beacco. In this dataset, each entry (word + POS-tag) is related to its CEFR reference level from Beacco and is described with 8 frequency variables, as shown in Table 2 . The frequency variables includes the 7 frequencies provided by FLELex along with the frequency from Lexique3. The latter will however be used only for the computation of the Lexique3 baseline (see Section 4.).",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 375,
                        "end": 382,
                        "text": "Table 2",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "BeaccoFLELexAtoB",
                "sec_num": "3.3.1."
            },
            {
                "text": "The main application of this study is to develop a more accurate mathematical model to transform FLELex frequencies into a single CEFR level, with the purpose of integrating this model within the web interface of the CEFR-Lex project instead of the First occ heuristic currently used. Therefore, training our model on the intersection described above has a main shortcoming: it is not able to classify any entries beyond B2 level, since it would not have seen any word from the C levels. In the FLELex interface, we nevertheless want to be able to classify words at those levels, as FLELex likely contains more difficult words than Beacco.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "BeaccoFLELexC",
                "sec_num": "3.3.2."
            },
            {
                "text": "To create this second dataset (BeaccoFLELexC), we first assumed that the 9,903 FLELex entries missing from Beacco can be considered as C level. However, before adding these entries to the 4,296 word intersection, we manually investigated them and noticed that about 2% present high frequencies in A levels textbooks, which is not expected for C words. We thus considered these cases as anomalies. Some causes of these anomalies were already discussed previously, but new issues also arose:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "BeaccoFLELexC",
                "sec_num": "3.3.2."
            },
            {
                "text": "\u2022 Function words appearing in Beacco's chapter 5, i.e. the grammar section, were not digitized, but they were logically captured in FLELex. They include personal pronouns (\"je\", \"tu\", \"toi\"), interrogative pronouns (\"combien\", \"o\u00f9\", \"comment\", \"quand\"), determiners (\"la\"), prepositions (\"en\", \"sur\"), conjunctions (\"apr\u00e8s\", \"pour\", \"que\"), modals (\"devoir\", \"pouvoir\"), and negative particles (\"non\", \"ne\", \"pas\");",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "BeaccoFLELexC",
                "sec_num": "3.3.2."
            },
            {
                "text": "\u2022 We also identified a few words appearing in chapter 3, linked to particular communicative functions, that were also excluded from our digitizing process (e.g. \"cher\" dear, \"bise\" kiss, \"peut-\u00eatre\" maybe, \"d'accord\" all right, etc.);",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "BeaccoFLELexC",
                "sec_num": "3.3.2."
            },
            {
                "text": "\u2022 Other words are very likely part of the A levels even if they are not included in Beacco's chapters we digitized (e.g. \"joli\" pretty, \"dormir\" to sleep, \"anglais\" English, or \"espagnol\" Spanish);",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "BeaccoFLELexC",
                "sec_num": "3.3.2."
            },
            {
                "text": "\u2022 Finally, we identified a few remaining tagging problems in FLELex that escaped the manual cleaning process (e.g. \"\u00e9tudiant\" student, \"ami\" friend were found as adjectives in FLELex instead of nouns).",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "BeaccoFLELexC",
                "sec_num": "3.3.2."
            },
            {
                "text": "To resolve some of these issues, we manually corrected tagging problems in FLELex and added the missing words appearing in chapters 3 and 5, assigning them their correct Beacco level. In total, 87 words were thus corrected, but some problems remain for a few entries. The last step in the preparation of this dataset Beac-coFLELexC consisted in creating a balanced dataset. Adding 9,903 C entries obviously produced a classimbalanced issue within the data, which we rebalanced using undersampling of overrepresented categories (C and B2). We used a random undersampling technique based on the number of entries in B1, reducing the size of this dataset from 14,236 to 4,878 words.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "BeaccoFLELexC",
                "sec_num": "3.3.2."
            },
            {
                "text": "For our experiments, we decided to use three standard machine learning algorithms, namely tree classification, boosting, and support vector machine (SVM). Neural networks were not considered due to the limited amount of data. We also defined four baselines to compare with, that are described below. All experiments were conducted following the same methodology. We first split each dataset into a training Table 2 : Examples of entries for \"plier\" to fold, \"chanteur\" singer, \"humide\" humid and \"entre\" between from the first dataset, illustrating the variables used in our experiments.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 407,
                        "end": 414,
                        "text": "Table 2",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Experiments",
                "sec_num": "3.4."
            },
            {
                "text": "(and validation) set including 80% of the entries and a test set including 20% of the entries. We then applied a grid search on the training set using a stratified 10-fold cross-validation setup to estimate the performance of each set of meta-parameters tested. Once the best set of metaparameters was chosen, we estimated the classification accuracy of the model on the test set. This procedure is more reliable than a standard 10-fold cross-validation setup as the meta-parameters and the parameters are not optimized on the same data.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Experiments",
                "sec_num": "3.4."
            },
            {
                "text": "Four baselines were used in this study. The first one (Maj class) assigns to all words the level of the majority class. It is a common baseline for all classification tasks. The second baseline (First occ) assigns to a given word the level of the textbook it was first observed in. The third baseline (Most freq), used for instance in Todirascu et al. (Todirascu et al., 2019) , assigns to each word the level with the highest frequency. For the fourth baseline, we trained three models (SVM, tree, and boosting) based only on Lexique3 frequencies, as a way to assess whether the L2-specific and more fine-grained frequency information from FLELex would lead to some improvements on the task.",
                "cite_spans": [
                    {
                        "start": 335,
                        "end": 376,
                        "text": "Todirascu et al. (Todirascu et al., 2019)",
                        "ref_id": "BIBREF21"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Baselines",
                "sec_num": "3.4.1."
            },
            {
                "text": "We applied the three above-mentioned algorithms to both our datasets: BeaccoFLELexAtoB and BeaccoFLELexC.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The models",
                "sec_num": "3.4.2."
            },
            {
                "text": "\u2022 On the former, the optimal meta-parameters found by the grid search for Lexique 3 were: Tree (max depth = 4, min sample leaf = 40, and min sample split = 50); SVM (RBF kernel with C = 0.01 and \u03b3 = 0.001); Boosting with 5 iterations.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The models",
                "sec_num": "3.4.2."
            },
            {
                "text": "\u2022 The meta-parameters found for FLELex frequencies were: Tree (max depth = 3, min sample leaf = 20, and min sample split = 50); SVM (RBF kernel with C = 1 and \u03b3 = 0.0001); Boosting with 5 iterations.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The models",
                "sec_num": "3.4.2."
            },
            {
                "text": "\u2022 On the latter, BeaccoFLELexC, the optimal metaparameters found using the grid search were: Tree (max depth = 3, min sample leaf = 20, and min sample split = 50); SVM (RBF kernel with C = 1 and \u03b3 = 0.001); Boosting with 5 iterations.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The models",
                "sec_num": "3.4.2."
            },
            {
                "text": "In this study, we aim to predict L2 expert knowledge based on word frequencies and thus obtain a machine learning algorithm able to transform a given word's frequency into a unique CEFR level. First, our systematic evaluation on the BeaccoFLELexAtoB dataset, whose results are reported in Table 3 , reveals that the First occ rule, currently used in the CEFRLex interface, yields poor performance. Its accuracy is as low as 25%, which is actually lower than the accuracy reached by a majority class classifier (40%) and its mean absolute error is 1.25, which means that this classification rule can miss targeted levels by even more than one level on average. Similarly, the Most freq rule, sometimes used as a simple and intuitive solution by some researchers, appears to be quite disappointing: its accuracy of 18% reveals that it is actually biased towards wrong answers. Using a machine learning algorithm to train a non-linear and more complex mathematical rule to transform FLELex distributions into CEFR levels seems to be a better path. We were able to reach 54% for the Boosting classifier and a mean absolute error of 0.66. The SVM model is more than twice as good as the First occ rule, and it outperforms the majority class classifier by 13%. On the second dataset, that corresponds better to the pragmatic problem we want to solve, it is interesting to notice that First occ outperforms the dummy baseline using majority class by 5%. The Most freq rule remains the worst option, whereas machine learning remains the best with the boosting algorithm reaching 48% of accuracy and a MAE of 0.75. Performance are slightly behind for the second dataset, but this is generally the case when one increases the number of classes.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 289,
                        "end": 296,
                        "text": "Table 3",
                        "ref_id": "TABREF3"
                    }
                ],
                "eq_spans": [],
                "section": "Results",
                "sec_num": "4."
            },
            {
                "text": "We also performed an ablation study on both datasets in order to find which frequency level contributed the most to the predictions. Results are presented in Table 4 and clearly shows that the frequency from the A1 to B1 levels are the more informative, especially the A1 level. Furthermore, one can notice that the total frequency (computed over all six levels) is also a good source of information.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 158,
                        "end": 165,
                        "text": "Table 4",
                        "ref_id": "TABREF5"
                    }
                ],
                "eq_spans": [],
                "section": "Results",
                "sec_num": "4."
            },
            {
                "text": "In our experiments, we wanted to know whether the L2specific and fine-grained frequency information provided in the CEFRLex resources would be better able to predict expert knowledge than a L1 frequency list. Table 3 shows that the models trained with FLELex slightly outperform (+5% in accuracy) the ones trained with Lexique3. However, this comparison is unfair, as the models leveraging FLELex information include more variables than the Lex-ique3 ones (7 vs. 1). Looking at the ablation study table, we can see performance when only the total frequency variable of FLELex is used. In such configuration, FLELex still outperforms Lexique3 by 1% accuracy, which seems to mean that L2 frequencies -even estimated on a much smaller corpus -might be used instead of L1 frequencies. This is, per se, a very interesting result, as the second language acquisition literature tends to believe the opposite and L2 intended word list created from a L1 corpus still remains the standard. In any case, those similar results can also be explained by the high correlation between the frequencies of these two lists, as was already reported in Fran\u00e7ois et al. (2014) . If we consider the performance of the full model (54%) compared to that of the model based only on the total frequency, the 4% improvement could be interpreted as a confirmation of the greater informational richness provided by a frequency distribution over proficiency levels compared to a unique word frequency. ",
                "cite_spans": [
                    {
                        "start": 1132,
                        "end": 1154,
                        "text": "Fran\u00e7ois et al. (2014)",
                        "ref_id": "BIBREF7"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 209,
                        "end": 216,
                        "text": "Table 3",
                        "ref_id": "TABREF3"
                    }
                ],
                "eq_spans": [],
                "section": "FLELex vs. Lexique 3",
                "sec_num": "4.1."
            },
            {
                "text": "The analysis of the precision, recall, and F1 values for each level reveals that models predictions are affected by one level in particular, level A2, which is already underrepresented in Beacco. Hence, the BeaccoFLELexAtoB dataset only includes 573 words at this level, whereas A1, B1 and B2 levels contain respectively 788, 1158 and 1777 words. Table 5 shows that extreme levels score much better than the middle ones, a recurrent outcome in readability classification tasks. It also reveals that, despite its lower accuracy score compare to the boosting model, the classification Tree model takes less drastic choices when assigning words to a class, which makes it a better option if we want a system that assigns words to all levels. We also noticed that, besides their under-representation in the RLD, A2 words are difficult to predict due to a high correlation between word frequencies in A1, A2 and B1 levels. Another problematic level is the C level, specially from a reading comprehension perspective. According to the CEFR descriptors, a C1 user \"can understand a wide range of demanding, longer texts, and recognise implicit meaning\", while a C2 user \"can understand with ease virtually everything heard or read\". Trying to translate these descriptors into actual words is difficult, as testified by the fact that Riba, who wrote the RLD opus for C levels, expressed some reserves concerning those descriptors, mainly because of the notion of perfection which emanate from them (Riba, 2016) , and the fact that C users depicted by the CEFR are only highly educated individuals, outperforming most of the native speakers. Consequently, we had to use a simple decision to define our C words in the gold-standard: considering everything minus words from levels A1 to B2. A final issue regarding C words is the fact that textbooks for those levels are less numerous than for the lower ones, providing FLELex with fewer words to observe and count.",
                "cite_spans": [
                    {
                        "start": 1490,
                        "end": 1502,
                        "text": "(Riba, 2016)",
                        "ref_id": "BIBREF16"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 347,
                        "end": 354,
                        "text": "Table 5",
                        "ref_id": "TABREF7"
                    }
                ],
                "eq_spans": [],
                "section": "Problematic levels",
                "sec_num": "4.2."
            },
            {
                "text": "In this section, we carried out a manual error analysis of some misclassification errors as a way to bring up to light some strengths and weaknesses of both approaches that can be used for selecting appropriate vocabulary in L2 learning: frequency vs. expert knowledge.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Discussion",
                "sec_num": "5."
            },
            {
                "text": "One characteristic of the RLDs that is worth remembering is the fact that lexical chapters are organised semantically, as the authors agreed that giving a list of words ranked alphabetically is of little use when it comes to design syllabus or build a teaching sequence (Beacco et al., 2008) . Hence, words evolving around the same notional scope come together, along with their synonyms, antonyms and as a matter of fact words belonging to the same family as well (e.g. \"heureux/malheureux\" happy/unhappy, \"maigre, gros/ maigrir, grossir\" skinny, fat / to loose weight, to put on weight). This conveys the idea that they should be taught together -in other words, at the same CEFR level -since building strong lexical networks is critical for vocabulary retention. Conversely, FLELex does not have such structure and is likely to estimate different frequency distributions for the various terms from a given semantic field. When we transform those distributions using either the First occ rule or even a machine learning algorithm, they are prone to end up at different levels (e.g. of predictions using the SVM: \"gros\"A2 / \"grossir\"B2, \"heureux\"A1 / \"malheureux\"A2). In this regard, using frequency lists to establish a vocabulary progression through several learning stages is limited because words are seen as isolated.",
                "cite_spans": [
                    {
                        "start": 270,
                        "end": 291,
                        "text": "(Beacco et al., 2008)",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Lexical approach",
                "sec_num": "5.1."
            },
            {
                "text": "Beacco's semantic organisation also enables it to better capture the effect of situation frequency, usually referred to as lexical availability (Mich\u00e9a, 1953) . The B2 level was the first RLD written, and it consists of extensive lists of words relating to specific centers of interest. The lower levels were compiled later, gradually picking up words from the B2 RLD according to the learning stage they could be taught at. As a result of this semantically-driven procedure, Beacco includes more available words than FLELex (e.g. of missing available words are \"soutien-gorge\" bra, \"cuisini\u00e8re\" cooker, \"s\u00e8che-cheveux\" hair-dryer, etc.).",
                "cite_spans": [
                    {
                        "start": 144,
                        "end": 158,
                        "text": "(Mich\u00e9a, 1953)",
                        "ref_id": "BIBREF14"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Lexical approach",
                "sec_num": "5.1."
            },
            {
                "text": "The question of topics covered on the L2 learning path is very relevant for this study, because it highlights the limits of both methods. FLELex computational approach aims to report words frequencies in the CEFR levels in an objective and descriptive way, but by using L2 material, it is compelled to favour certain topics and exclude others. Compared to Beacco, we found that textbooks mainly avoid potentially polemic themes such as religion or death, or subjects in which students could have not much to say such as DIY, and topics commonly considered as complicated, for instance economics or sciences. In contrast, the topics found in FLELex are highly influenced by the choice of texts comprised in the corpus and can sometimes be overrepresented. A clear materialization of this shortcoming appeared when we looked at FLELex frequent words absent from Beacco and discovered that words related to fairytales were abundant (e.g. \"ch\u00e2teau\" castle, \"reine\" queen, \"prince\" prince, \"chevalier\" knigth, \"dragon\" dragon, \"magique\" magic, and even \"\u00e9pouser\" to marry or \"r\u00eaver\" dream). This can be explained by the inclusion of a simplified reader dedicated to King Arthur legend in the corpus. On the other hand, the RLD's semantic structure has a downside since it may lead to loose sight of the CEFR descriptors, specially in topics where finding a progression between the items in terms of language acts is arduous. The most iconic theme we came across is food and drinks, with 150 words absent from FLELex, but geography and the human body also share the same characteristics at a lesser degree. We distinguished those topics from the others because they are mostly composed of nouns with closely related meaning (e.g. in B2, \"pain fran\u00e7ais\", \"baguette\", \"boule\", \"b\u00e2tard\", \"pain de campagne\", \"carr\u00e9\", \"pain int\u00e9gral\", \"pain complet\", \"pistolet\", \"petit pain\", \"sandwich\", \"petit pain au lait\", all being different types of bread). The large number of words in these topics is a reflection of reality usually bypassed in textbooks, since these nouns don't offer a wide variety of communicative situations.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Topics",
                "sec_num": "5.2."
            },
            {
                "text": "FLELex is a descriptive tool built from texts related to reading comprehension tasks in FFL materials, illustrating therefore the contents of written reception activities. The RLD also presents its contents as to be mastered in comprehension tasks, leaving the decision to teachers and curriculum designers regarding what learners should be able to produce (Beacco et al., 2008) . However, we identified four POS in which the ability to produce words seems to be the selection criteria for Beacco: determiners, conjunctions (e.g. \"comme\" in B1), pronouns (e.g. \"le\" in B2), and prepositions. We detected them because the frequencies of those POS are among the highest of the corpus while their levels nevertheless vary from A1 to B2 in Beacco. Even though words belonging to these POS are probably understood at early stages due to repeated exposure, the RLD proposes a gradation in the different learning stages they should be taught at, which is likely motivated either by the CEFR descriptors regarding production and interaction or by intrinsic characteristics of the word. We therefore found that the two approaches are not compatible for those specific POS, as the prescriptive aspect of the RLD implies to take into account learners objectives and abilities in production tasks as well, while FLELex only illustrates the language used in reception tasks.",
                "cite_spans": [
                    {
                        "start": 357,
                        "end": 378,
                        "text": "(Beacco et al., 2008)",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Reception or production",
                "sec_num": "5.3."
            },
            {
                "text": "Beacco's intent is to propose a reference calibration for CEFR levels, but not a list of words that would be mandatory and identical, in all places and at all times. In the introduction, the authors minimize the inherent normative aspect of their lists, presenting them as only a starting point to insure compatibility between syllabus and exams of different educational systems. Therefore, they display vocabulary in three possible ways:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Normative and adaptable",
                "sec_num": "5.4."
            },
            {
                "text": "\u2022 closed lists, e.g. \"b\u00e9b\u00e9, enfant, lait\"",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Normative and adaptable",
                "sec_num": "5.4."
            },
            {
                "text": "\u2022 open lists, e.g. \"[...] agr\u00e9able, b\u00eate, calme, content\"",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Normative and adaptable",
                "sec_num": "5.4."
            },
            {
                "text": "\u2022 list descriptors, e.g. \"[...] noms de nationalit\u00e9s\" Such behavior, intended to human readers, however raises some issues for an automatic approach. Facing list descriptors, we generally ignored them in the digitizing process, which explains why words such as \"anglais\" English and \"espagnol\" Spanish -which are nationalities -were not found in our version of Beacco, although present in FLELex. For our study, open lists and list descriptors are very problematic in the sense that the absence of a word from a level cannot be considered as 100% certain. From a teacher's perspective though, those open lists and item descriptions are coherent with the authors goal to provide content adaptable to all contexts, and indications that the items are to be chosen according to the geographic, cultural and educational situation (e.g. for the nationalities, \"japonais\", \"cor\u00e9en\" and \"vietnamien\" are likely to be taught in A1 to Asian learners, whereas they might not be needed from A1 in a South American classroom).",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Normative and adaptable",
                "sec_num": "5.4."
            },
            {
                "text": "In this research, we aimed to infer CEFR levels from CE-FRLex word frequency distribution using expert knowledge from the French RLD as a gold-standard. Such approach enabled us to apply, for the first time, machine learning algorithms to such task whereas previous work used simple mathematical rules. After standardisation of the data, we trained three machine learning models on two sets, reaching an accuracy score of 0.54 for the dataset BeaccoFLELexA-toB and of 0.48 for the BeaccoFLELexC dataset. These results clearly outperforms results reached by the First occ rule currently used in the CEFRLex interface. Our work has direct repercussions on this project, as our best classifier has been integrated in the interface 4 , offering now the choice between Beacco or First occ to classify words. Our experiments also yield other interesting results. First, comparing our results with those of a L1 frequency word list revealed that the distributional information contained in FLELex indeed seems richer and finer-grained than the one of a standard L1 list. Second, we carried out an analysis on the most important classification errors as a way to sharpen our understanding of the differences existing between the two approaches we compared: frequency and expert knowledge. This analysis stressed the importance of lexical networks in L2 learning to ensure a better representation of available words and of words connected to topics generally avoided in textbooks. We also noticed that although CEFRLex resources only represent receptive skills, Beacco might have sometimes classified words based on criteria relative to both receptive and productive skills. Finally, the presence of list descriptors in RLD is a serious issue for their automatic exploitation, as they contain some implicit knowledge. We believe that all these discrepancies partially explain why our statistical model is not able to better predict Beacco's level. In other words, although a better option than the First occ rule, using expert knowledge also has shortcomings. In the future, we plan to investigate the use of L2 learners data as an alternative source of information to transform CEFRLex distribution into levels.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Conclusion",
                "sec_num": "6."
            },
            {
                "text": "In the case of the English Vocabulary Profile (EVP), the designers have indeed combined lexicographical and pedagogical knowledge with word frequency information(Capel, 2010). However, the frequencies were estimated from a learner corpus and therefore are representative of productive skills rather than receptive ones.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "See the list of concerned languages at http://www.coe.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "The RLD book for the C levels(Riba, 2016) was not used in this study, as it doesn't provide lists of lexical items, but rather describe more conceptual abilities, like managing and structuring",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            }
        ],
        "back_matter": [],
        "bib_entries": {
            "BIBREF0": {
                "ref_id": "b0",
                "title": "From distributions to labels: A lexical proficiency analysis using learner corpora",
                "authors": [
                    {
                        "first": "D",
                        "middle": [],
                        "last": "Alfter",
                        "suffix": ""
                    },
                    {
                        "first": "Y",
                        "middle": [],
                        "last": "Bizzoni",
                        "suffix": ""
                    },
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Agebj\u00f3rn",
                        "suffix": ""
                    },
                    {
                        "first": "E",
                        "middle": [],
                        "last": "Volodina",
                        "suffix": ""
                    },
                    {
                        "first": "I",
                        "middle": [],
                        "last": "Pil\u00e1n",
                        "suffix": ""
                    }
                ],
                "year": 2016,
                "venue": "Proceedings of the joint workshop on NLP4CALL and NLP for Language Acquisition at SLTC, number 130",
                "volume": "",
                "issue": "",
                "pages": "1--7",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Alfter, D., Bizzoni, Y., Agebj\u00f3rn, A., Volodina, E., and Pil\u00e1n, I. (2016). From distributions to labels: A lexical proficiency analysis using learner corpora. In Proceed- ings of the joint workshop on NLP4CALL and NLP for Language Acquisition at SLTC, number 130, pages 1-7. Link\u00f6ping University Electronic Press.",
                "links": null
            },
            "BIBREF1": {
                "ref_id": "b1",
                "title": "The {CELEX} lexical data base on {CD-ROM}. Linguistic Data Consortium",
                "authors": [
                    {
                        "first": "R",
                        "middle": [],
                        "last": "Baayen",
                        "suffix": ""
                    },
                    {
                        "first": "R",
                        "middle": [],
                        "last": "Piepenbrock",
                        "suffix": ""
                    },
                    {
                        "first": "H",
                        "middle": [],
                        "last": "Van Rijn",
                        "suffix": ""
                    }
                ],
                "year": 1993,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Baayen, R., Piepenbrock, R., and van Rijn, H. (1993). The {CELEX} lexical data base on {CD-ROM}. Linguistic Data Consortium, Philadelphia: Univ. of Pennsylvania.",
                "links": null
            },
            "BIBREF3": {
                "ref_id": "b3",
                "title": "Niveau A2 pour le fran\u00e7ais: Un r\u00e9f\u00e9rentiel. Didier",
                "authors": [],
                "year": null,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Niveau A2 pour le fran\u00e7ais: Un r\u00e9f\u00e9rentiel. Didier.",
                "links": null
            },
            "BIBREF4": {
                "ref_id": "b4",
                "title": "Moving beyond ku\u010dera and francis: A critical evaluation of current word frequency norms and the introduction of a new and improved word frequency measure for american english",
                "authors": [
                    {
                        "first": "M",
                        "middle": [],
                        "last": "Brysbaert",
                        "suffix": ""
                    },
                    {
                        "first": "B",
                        "middle": [],
                        "last": "New",
                        "suffix": ""
                    }
                ],
                "year": 2009,
                "venue": "Behavior research methods",
                "volume": "41",
                "issue": "4",
                "pages": "977--990",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Brysbaert, M. and New, B. (2009). Moving beyond ku\u010dera and francis: A critical evaluation of current word fre- quency norms and the introduction of a new and im- proved word frequency measure for american english. Behavior research methods, 41(4):977-990.",
                "links": null
            },
            "BIBREF5": {
                "ref_id": "b5",
                "title": "A1-b2 vocabulary: Insights and issues arising from the english profile wordlists project",
                "authors": [
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Capel",
                        "suffix": ""
                    }
                ],
                "year": 2010,
                "venue": "English Profile Journal",
                "volume": "1",
                "issue": "1",
                "pages": "1--11",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Capel, A. (2010). A1-b2 vocabulary: Insights and issues arising from the english profile wordlists project. En- glish Profile Journal, 1(1):1-11.",
                "links": null
            },
            "BIBREF6": {
                "ref_id": "b6",
                "title": "EFLLex: A Graded Lexical Resource for Learners of English as a Foreign Language",
                "authors": [
                    {
                        "first": "L",
                        "middle": [],
                        "last": "D\u00fcrlich",
                        "suffix": ""
                    },
                    {
                        "first": "T",
                        "middle": [],
                        "last": "Fran\u00e7ois",
                        "suffix": ""
                    }
                ],
                "year": 2018,
                "venue": "Proceedings of LREC 2018",
                "volume": "",
                "issue": "",
                "pages": "873--879",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "D\u00fcrlich, L. and Fran\u00e7ois, T. (2018). EFLLex: A Graded Lexical Resource for Learners of English as a Foreign Language. In Proceedings of LREC 2018, pages 873- 879.",
                "links": null
            },
            "BIBREF7": {
                "ref_id": "b7",
                "title": "FLELex: a graded lexical resource for French foreign learners",
                "authors": [
                    {
                        "first": "T",
                        "middle": [],
                        "last": "Fran\u00e7ois",
                        "suffix": ""
                    },
                    {
                        "first": "N",
                        "middle": [],
                        "last": "Gala",
                        "suffix": ""
                    },
                    {
                        "first": "P",
                        "middle": [],
                        "last": "Watrin",
                        "suffix": ""
                    },
                    {
                        "first": "C",
                        "middle": [],
                        "last": "Fairon",
                        "suffix": ""
                    }
                ],
                "year": 2014,
                "venue": "Proceedings of LREC 2014",
                "volume": "",
                "issue": "",
                "pages": "3766--3773",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Fran\u00e7ois, T., Gala, N., Watrin, P., and Fairon, C. (2014). FLELex: a graded lexical resource for French foreign learners. In Proceedings of LREC 2014, pages 3766- 3773.",
                "links": null
            },
            "BIBREF8": {
                "ref_id": "b8",
                "title": "SVALex: a CEFR-graded lexical resource for Swedish foreign and second language learners",
                "authors": [
                    {
                        "first": "T",
                        "middle": [],
                        "last": "Fran\u00e7ois",
                        "suffix": ""
                    },
                    {
                        "first": "E",
                        "middle": [],
                        "last": "Volodina",
                        "suffix": ""
                    },
                    {
                        "first": "P",
                        "middle": [],
                        "last": "Ildik\u00f3",
                        "suffix": ""
                    },
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Tack",
                        "suffix": ""
                    }
                ],
                "year": 2016,
                "venue": "Proceedings of LREC 2016",
                "volume": "",
                "issue": "",
                "pages": "213--219",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Fran\u00e7ois, T., Volodina, E., Ildik\u00f3, P., and Tack, A. (2016). SVALex: a CEFR-graded lexical resource for Swedish foreign and second language learners. In Proceedings of LREC 2016, pages 213-219.",
                "links": null
            },
            "BIBREF9": {
                "ref_id": "b9",
                "title": "Towards a french lexicon with difficulty measures: Nlp helping to bridge the gap between traditional dictionaries and specialized lexicons",
                "authors": [
                    {
                        "first": "N",
                        "middle": [],
                        "last": "Gala",
                        "suffix": ""
                    },
                    {
                        "first": "T",
                        "middle": [],
                        "last": "Fran\u00e7ois",
                        "suffix": ""
                    },
                    {
                        "first": "C",
                        "middle": [],
                        "last": "Fairon",
                        "suffix": ""
                    }
                ],
                "year": 2013,
                "venue": "Proceedings of eLex2013",
                "volume": "",
                "issue": "",
                "pages": "132--151",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Gala, N., Fran\u00e7ois, T., and Fairon, C. (2013). Towards a french lexicon with difficulty measures: Nlp helping to bridge the gap between traditional dictionaries and spe- cialized lexicons. In Proceedings of eLex2013, pages 132-151.",
                "links": null
            },
            "BIBREF11": {
                "ref_id": "b11",
                "title": "Key issues in l2 reading development",
                "authors": [
                    {
                        "first": "W",
                        "middle": [],
                        "last": "Grabe",
                        "suffix": ""
                    }
                ],
                "year": 2014,
                "venue": "CELC Symposium Bridging Research and Pedagogy",
                "volume": "",
                "issue": "",
                "pages": "8--18",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Grabe, W. (2014). Key issues in l2 reading development. In CELC Symposium Bridging Research and Pedagogy, pages 8-18.",
                "links": null
            },
            "BIBREF12": {
                "ref_id": "b12",
                "title": "Computational analysis of present-day American English",
                "authors": [
                    {
                        "first": "H",
                        "middle": [],
                        "last": "Ku\u010dera",
                        "suffix": ""
                    },
                    {
                        "first": "W",
                        "middle": [],
                        "last": "Francis",
                        "suffix": ""
                    }
                ],
                "year": 1967,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Ku\u010dera, H. and Francis, W. (1967). Computational analy- sis of present-day American English. Brown University Press, Providence.",
                "links": null
            },
            "BIBREF13": {
                "ref_id": "b13",
                "title": "How much lexis is necessary for reading comprehension?",
                "authors": [
                    {
                        "first": "B",
                        "middle": [
                            "; G"
                        ],
                        "last": "Laufer",
                        "suffix": ""
                    },
                    {
                        "first": "P",
                        "middle": [],
                        "last": "Rayson",
                        "suffix": ""
                    },
                    {
                        "first": "Wilson",
                        "middle": [],
                        "last": "",
                        "suffix": ""
                    },
                    {
                        "first": "A",
                        "middle": [],
                        "last": "",
                        "suffix": ""
                    }
                ],
                "year": 1992,
                "venue": "Word frequencies in written and spoken english: based on the british national corpus",
                "volume": "",
                "issue": "",
                "pages": "126--132",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Laufer, B. (1992). How much lexis is necessary for reading comprehension? In Vocabulary and applied linguistics, pages 126-132. Springer. Leech, G., Rayson, P., and Wilson, A. (2001). Word fre- quencies in written and spoken english: based on the british national corpus.",
                "links": null
            },
            "BIBREF14": {
                "ref_id": "b14",
                "title": "Mots fr\u00e9quents et mots disponibles. un aspect nouveau de la statistique du langage. Les langues modernes",
                "authors": [
                    {
                        "first": "R",
                        "middle": [],
                        "last": "Mich\u00e9a",
                        "suffix": ""
                    }
                ],
                "year": 1953,
                "venue": "",
                "volume": "47",
                "issue": "",
                "pages": "338--344",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Mich\u00e9a, R. (1953). Mots fr\u00e9quents et mots disponibles. un aspect nouveau de la statistique du langage. Les langues modernes, 47(4):338-344.",
                "links": null
            },
            "BIBREF15": {
                "ref_id": "b15",
                "title": "The use of film subtitles to estimate word frequencies",
                "authors": [
                    {
                        "first": "B",
                        "middle": [],
                        "last": "New",
                        "suffix": ""
                    },
                    {
                        "first": "M",
                        "middle": [],
                        "last": "Brysbaert",
                        "suffix": ""
                    },
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Veronis",
                        "suffix": ""
                    },
                    {
                        "first": "C",
                        "middle": [],
                        "last": "Pallier",
                        "suffix": ""
                    }
                ],
                "year": 2007,
                "venue": "Applied Psycholinguistics",
                "volume": "28",
                "issue": "04",
                "pages": "661--677",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "New, B., Brysbaert, M., Veronis, J., and Pallier, C. (2007). The use of film subtitles to estimate word frequencies. Applied Psycholinguistics, 28(04):661-677.",
                "links": null
            },
            "BIBREF16": {
                "ref_id": "b16",
                "title": "Niveaux C1 / C2 pour le fran\u00e7ais: el\u00e9ments pour un r\u00e9f\u00e9rentiel",
                "authors": [
                    {
                        "first": "P",
                        "middle": [],
                        "last": "Riba",
                        "suffix": ""
                    }
                ],
                "year": 2016,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Riba, P. (2016). Niveaux C1 / C2 pour le fran\u00e7ais: el\u00e9ments pour un r\u00e9f\u00e9rentiel. Didier.",
                "links": null
            },
            "BIBREF17": {
                "ref_id": "b17",
                "title": "Probabilistic part-of-speech tagging using decision trees",
                "authors": [
                    {
                        "first": "H",
                        "middle": [],
                        "last": "Schmid",
                        "suffix": ""
                    }
                ],
                "year": 1994,
                "venue": "Proceedings of International Conference on New Methods in Language Processing",
                "volume": "12",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Schmid, H. (1994). Probabilistic part-of-speech tagging using decision trees. In Proceedings of International Conference on New Methods in Language Processing, volume 12. Manchester, UK.",
                "links": null
            },
            "BIBREF18": {
                "ref_id": "b18",
                "title": "Evaluating lexical simplification and vocabulary knowledge for learners of french: possibilities of using the flelex resource",
                "authors": [
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Tack",
                        "suffix": ""
                    },
                    {
                        "first": "T",
                        "middle": [],
                        "last": "Fran\u00e7ois",
                        "suffix": ""
                    },
                    {
                        "first": "A.-L",
                        "middle": [],
                        "last": "Ligozat",
                        "suffix": ""
                    },
                    {
                        "first": "C",
                        "middle": [],
                        "last": "Fairon",
                        "suffix": ""
                    }
                ],
                "year": 2016,
                "venue": "Proceedings of LREC2016",
                "volume": "",
                "issue": "",
                "pages": "230--236",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Tack, A., Fran\u00e7ois, T., Ligozat, A.-L., and Fairon, C. (2016). Evaluating lexical simplification and vocabulary knowledge for learners of french: possibilities of using the flelex resource. In Proceedings of LREC2016, pages 230-236.",
                "links": null
            },
            "BIBREF19": {
                "ref_id": "b19",
                "title": "NT2Lex: A CEFR-Graded Lexical Resource for Dutch as a Foreign Language Linked to Open Dutch WordNet",
                "authors": [
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Tack",
                        "suffix": ""
                    },
                    {
                        "first": "T",
                        "middle": [],
                        "last": "Fran\u00e7ois",
                        "suffix": ""
                    },
                    {
                        "first": "P",
                        "middle": [],
                        "last": "Desmet",
                        "suffix": ""
                    },
                    {
                        "first": "C",
                        "middle": [],
                        "last": "Fairon",
                        "suffix": ""
                    }
                ],
                "year": 2018,
                "venue": "Proceedings of",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Tack, A., Fran\u00e7ois, T., Desmet, P., and Fairon, C. (2018). NT2Lex: A CEFR-Graded Lexical Resource for Dutch as a Foreign Language Linked to Open Dutch WordNet. In Proceedings of BEA 2018.",
                "links": null
            },
            "BIBREF20": {
                "ref_id": "b20",
                "title": "Word knowledge in the elementary school",
                "authors": [
                    {
                        "first": "E",
                        "middle": [],
                        "last": "Thorndike",
                        "suffix": ""
                    }
                ],
                "year": 1921,
                "venue": "The Teachers College Record",
                "volume": "22",
                "issue": "4",
                "pages": "334--370",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Thorndike, E. (1921). Word knowledge in the elementary school. The Teachers College Record, 22(4):334-370.",
                "links": null
            },
            "BIBREF21": {
                "ref_id": "b21",
                "title": "Polylexfle: une base de donn\u00e9es d'expressions polylexicales pour le fle",
                "authors": [
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Todirascu",
                        "suffix": ""
                    },
                    {
                        "first": "M",
                        "middle": [],
                        "last": "Cargill",
                        "suffix": ""
                    },
                    {
                        "first": "T",
                        "middle": [],
                        "last": "Fran\u00e7ois",
                        "suffix": ""
                    }
                ],
                "year": 2019,
                "venue": "Actes de la conf\u00e9rence TALN 2019",
                "volume": "",
                "issue": "",
                "pages": "143--156",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Todirascu, A., Cargill, M., and Fran\u00e7ois, T. (2019). Polylexfle: une base de donn\u00e9es d'expressions polylexi- cales pour le fle. In Actes de la conf\u00e9rence TALN 2019, pages 143-156.",
                "links": null
            }
        },
        "ref_entries": {
            "TABREF0": {
                "type_str": "table",
                "content": "<table><tr><td colspan=\"4\">Level # FLELex # First occ # Beacco</td></tr><tr><td>A1</td><td>4,097</td><td>4,097</td><td>827</td></tr><tr><td>A2</td><td>5,768</td><td>2,699</td><td>615</td></tr><tr><td>B1</td><td>9,074</td><td>3,980</td><td>1334</td></tr><tr><td>B2</td><td>6,309</td><td>1,299</td><td>2742</td></tr><tr><td>C1</td><td>7,267</td><td>1,665</td><td>x</td></tr><tr><td>C2</td><td>3,932</td><td>496</td><td>x</td></tr></table>",
                "num": null,
                "text": "Distribution of entries per CEFR level, including the total number of items per level in FLELex, the number of items per level calculated with First occ, and the number of words per level in Beacco.",
                "html": null
            },
            "TABREF3": {
                "type_str": "table",
                "content": "<table/>",
                "num": null,
                "text": "Test results on both datasets.",
                "html": null
            },
            "TABREF5": {
                "type_str": "table",
                "content": "<table/>",
                "num": null,
                "text": "Variable ablation study on both datasets, using the boosting model.",
                "html": null
            },
            "TABREF7": {
                "type_str": "table",
                "content": "<table/>",
                "num": null,
                "text": "F1 scores per level for the three models, on the BeaccoFLELexAtoB dataset.",
                "html": null
            }
        }
    }
}