| Literature DB >> 35535319 |
Mi-Young Kim1,2, Juliano Rabelo2, Kingsley Okeke1, Randy Goebel2.
Abstract
We describe the techniques applied by the University of Alberta (UA) team in the most recent Competition on Legal Information Extraction and Entailment (COLIEE 2021). We participated in retrieval and entailment tasks for both case law and statute law; we applied a transformer-based approach for the case law entailment task, an information retrieval technique based on BM25 for legal information retrieval, and a natural language inference mechanism using semantic knowledge applied to statute law texts. This competition included 25 teams from 14 countries; our case law entailment approach was ranked no. 4 in Task 2, the BM25 technique for legal information retrieval was ranked no. 3 in Task 3, and the natural language inference technique incorporating semantic information was ranked no. 4 in Task 4. The combination of the latter two techniques on Task 5 was ranked no. 2. We also performed error analysis of our system in Task 4, which provides some insight into current state-of-the-art and research priorities for future directions.Entities:
Keywords: BM25; Legal Information Entailment; Legal Information Extraction; Transformers
Year: 2022 PMID: 35535319 PMCID: PMC9075849 DOI: 10.1007/s12626-022-00103-1
Source DB: PubMed Journal: Rev Socionetwork Strateg ISSN: 1867-3236
Fig. 1Architectures of Tasks 3, 4 and 5
Task 2 official results
| Team | File | F1 |
|---|---|---|
| NM | Run_task2_DebertaT5.txt | 0.6912 |
| NM | Run_task2_monoT5.txt | 0.6610 |
| NM | Run_task2_Deberta.txt | 0.6339 |
| UA | UA_reg_pp.txt | 0.6274 |
| JNLP | JNLP.task2.BM25Sup._Den..txt | 0.6116 |
| JNLP | JNLP.task2.BM25Sup._Den._F..txt | 0.6091 |
| UA | UA_def_pp.txt | 0.5875 |
| JNLP | JNLP.task2.NFSP_BM25.txt | 0.5868 |
| siat | siatCLS_result-task2.txt | 0.5860 |
| DSSIR | run_test_bm25.txt | 0.5806 |
| siat | siatFGM_result-task2.txt | 0.5670 |
| UA | UA_loose_pp.txt | 0.5603 |
| TR | task2_TR.txt | 0.5438 |
| DSSIR | run_test_bm25_dpr.txt | 0.5161 |
| DSSIR | run_test_dpr.txt | 0.5161 |
| MAN01 | [MAN01] task2 run1.txt | 0.5069 |
| MAN01 | [MAN01] task2 run0.txt | 0.2500 |
IR (Task3) results on test run data in COLIEE 2021
| Team | F2 | P | R | MAP | R_5 | R_10 | R_30 |
|---|---|---|---|---|---|---|---|
| OvGU_run1 | 0.73 | 0.67 | 0.77 | 0.74 | 0.75 | 0.81 | 0.85 |
| JNLP.CLMLT | 0.72 | 0.60 | 0.80 | 0.79 | 0.78 | 0.89 | 0.95 |
| BM25.UA | 0.70 | 0.75 | 0.70 | 0.75 | 0.71 | 0.73 | 0.81 |
| JNLP.CLBJP | 0.70 | 0.62 | 0.77 | 0.77 | 0.82 | 0.84 | 0.90 |
| R3.LLNTU | 0.70 | 0.66 | 0.74 | 0.78 | 0.79 | 0.83 | 0.91 |
| R2.LLNTU | 0.70 | 0.67 | 0.73 | 0.78 | 0.78 | 0.84 | 0.91 |
| R1.LLNTU | 0.68 | 0.63 | 0.73 | 0.78 | 0.78 | 0.84 | 0.91 |
| JNLP.CLBJ | 0.68 | 0.55 | 0.77 | 0.77 | 0.81 | 0.84 | 0.91 |
| OvGU_run2 | 0.67 | 0.48 | 0.80 | 0.75 | 0.75 | 0.81 | 0.90 |
| TFIDF.UA | 0.65 | 0.67 | 0.65 | 0.73 | 0.72 | 0.74 | 0.81 |
| LM.UA | 0.54 | 0.56 | 0.54 | 0.64 | 0.64 | 0.68 | 0.81 |
| TR_HB | 0.52 | 0.33 | 0.61 | 0.66 | 0.71 | 0.74 | 0.84 |
| HUKB-3 | 0.52 | 0.29 | 0.69 | 0.61 | 0.68 | 0.74 | 0.87 |
| HUKB-1 | 0.47 | 0.23 | 0.65 | 0.61 | 0.66 | 0.75 | 0.87 |
| TR_AV1 | 0.35 | 0.26 | 0.51 | 0.46 | 0.43 | 0.47 | 0.56 |
| TR_AV2 | 0.33 | 0.14 | 0.55 | 0.43 | 0.39 | 0.44 | 0.49 |
| HUKB-2 | 0.32 | 0.32 | 0.32 | 0.41 | 0.46 | 0.54 | 0.61 |
| OvGU_run3 | 0.30 | 0.15 | 0.70 | 0.55 | 0.57 | 0.61 | 0.70 |
Fig. 2Kadokawa Thesaurus Hierarchy [39]
Fig. 3Architecture of Task 4
NLI (Task 4) results on test data
| Team | Sid | Correct | Accuracy |
|---|---|---|---|
| BaseLine | 43/All 81 | 0.5309 | |
| HUKB | HUKB-2 | 57 | 0.7037 |
| HUKB | HUKB-1 | 55 | 0.6790 |
| HUKB | HUKB-3 | 55 | 0.6790 |
| UA | UA_parser | 54 | 0.6667 |
| JNLP | JNLP.EC | 51 | 0.6296 |
| JNLP | JNLP.ECS | 51 | 0.6296 |
| JNLP | JNLP.EB | 51 | 0.6296 |
| OVGU | OVGU_run3 | 48 | 0.5926 |
| TR | TR-Ensemble | 48 | 0.5926 |
| TR | TR-MTE | 48 | 0.5926 |
| OVGU | OVGU_run2 | 45 | 0.5556 |
| KIS | KIS1 | 44 | 0.5432 |
| KIS | KIS3 | 44 | 0.5432 |
| UA | UA_elmo | 44 | 0.5432 |
| KIS | KIS2 | 43 | 0.5309 |
| UA | UA_dl | 43 | 0.5309 |
| TR | TR_Electra | 41 | 0.5062 |
| OVGU | OVGU_run1 | 36 | 0.4444 |
Task 5 (IR+NLI) results on test data in COLIEE 2021
| Team | Sid | Correct | Accuracy |
|---|---|---|---|
| BaseLine | 43/All 81 | 0.5309 | |
| JNLP | JNLP.NFSP | 49 | 0.6049 |
| UA | UA_parser | 46 | 0.5679 |
| JNLP | JNLP.NMSP | 45 | 0.5556 |
| UA | UA_dl | 45 | 0.5556 |
| TR | TRDistillRoberta | 44 | 0.5432 |
| KIS | KIS_2 | 41 | 0.5062 |
| KIS | KIS_3 | 41 | 0.5062 |
| UA | UA_elmo | 40 | 0.4938 |
| JNLP | JNLP.task5.B_M | 38 | 0.4691 |
| KIS | KIS_1 | 35 | 0.4321 |
| TR | TRGPT3Ada | 35 | 0.4321 |
| TR | TRGPT3Davinci | 35 | 0.4321 |
Task 4 Error types
| Error type | UA_parser | UA_dl | UA_elmo |
|---|---|---|---|
| Wrong analysis of condition | 7 | 9 | 6 |
| Wrong analysis of conclusion | 1 | 1 | 1 |
| Negation detection error | 1 | 5 | 9 |
| Paraphrase detection error | 15 | 15 | 16 |
| Reference resolution error | 1 | 3 | 3 |
| Wrong analysis of conjunction | 1 | 2 | 1 |
| etc. | 1 | 3 | 1 |