Commit
·
3224233
1
Parent(s):
bf77c15
Model save
Browse files- README.md +333 -0
- all_results.json +407 -0
- final_checkpoint/README.md +333 -0
- final_checkpoint/added_tokens.json +4 -0
- final_checkpoint/config.json +228 -0
- final_checkpoint/merges.txt +0 -0
- final_checkpoint/model.safetensors +3 -0
- final_checkpoint/special_tokens_map.json +51 -0
- final_checkpoint/tokenizer.json +0 -0
- final_checkpoint/tokenizer_config.json +75 -0
- final_checkpoint/training_args.bin +3 -0
- final_checkpoint/vocab.json +0 -0
- model.safetensors +1 -1
- runs/Nov27_07-50-08_trinity/events.out.tfevents.1701071419.trinity.224163.0 +2 -2
- runs/Nov27_07-50-08_trinity/events.out.tfevents.1701076911.trinity.224163.1 +3 -0
- test_results.json +407 -0
README.md
ADDED
|
@@ -0,0 +1,333 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
library_name: span-marker
|
| 3 |
+
tags:
|
| 4 |
+
- span-marker
|
| 5 |
+
- token-classification
|
| 6 |
+
- ner
|
| 7 |
+
- named-entity-recognition
|
| 8 |
+
- generated_from_span_marker_trainer
|
| 9 |
+
datasets:
|
| 10 |
+
- DFKI-SLT/few-nerd
|
| 11 |
+
metrics:
|
| 12 |
+
- precision
|
| 13 |
+
- recall
|
| 14 |
+
- f1
|
| 15 |
+
widget:
|
| 16 |
+
- text: In response, in May or June 1125, a 3,000-strong Crusader coalition commanded
|
| 17 |
+
by King Baldwin II of Jerusalem confronted and defeated the 15,000-strong Muslim
|
| 18 |
+
coalition at the Battle of Azaz, raising the siege of the town.
|
| 19 |
+
- text: Cardenal made several visits to Jesuit universities in the United States,
|
| 20 |
+
including the University of Detroit Mercy in 2013, and the John Carroll University
|
| 21 |
+
in 2014.
|
| 22 |
+
- text: Other super-spreaders, defined as those that transmit SARS to at least eight
|
| 23 |
+
other people, included the incidents at the Hotel Metropole in Hong Kong, the
|
| 24 |
+
Amoy Gardens apartment complex in Hong Kong and one in an acute care hospital
|
| 25 |
+
in Toronto, Ontario, Canada.
|
| 26 |
+
- text: The District Court for the Northern District of California rejected 321 Studios'
|
| 27 |
+
claims for declaratory relief, holding that both DVD Copy Plus and DVD-X Copy
|
| 28 |
+
violated the DMCA and that the DMCA was not unconstitutional.
|
| 29 |
+
- text: The Sunday Edition is a television programme broadcast on the ITV Network
|
| 30 |
+
in the United Kingdom focusing on political interview and discussion, produced
|
| 31 |
+
by ITV Productions.
|
| 32 |
+
pipeline_tag: token-classification
|
| 33 |
+
model-index:
|
| 34 |
+
- name: SpanMarker
|
| 35 |
+
results:
|
| 36 |
+
- task:
|
| 37 |
+
type: token-classification
|
| 38 |
+
name: Named Entity Recognition
|
| 39 |
+
dataset:
|
| 40 |
+
name: Unknown
|
| 41 |
+
type: DFKI-SLT/few-nerd
|
| 42 |
+
split: test
|
| 43 |
+
metrics:
|
| 44 |
+
- type: f1
|
| 45 |
+
value: 0.703084859534267
|
| 46 |
+
name: F1
|
| 47 |
+
- type: precision
|
| 48 |
+
value: 0.7034273336857051
|
| 49 |
+
name: Precision
|
| 50 |
+
- type: recall
|
| 51 |
+
value: 0.7027427186979075
|
| 52 |
+
name: Recall
|
| 53 |
+
---
|
| 54 |
+
|
| 55 |
+
# SpanMarker
|
| 56 |
+
|
| 57 |
+
This is a [SpanMarker](https://github.com/tomaarsen/SpanMarkerNER) model trained on the [DFKI-SLT/few-nerd](https://huggingface.co/datasets/DFKI-SLT/few-nerd) dataset that can be used for Named Entity Recognition.
|
| 58 |
+
|
| 59 |
+
## Model Details
|
| 60 |
+
|
| 61 |
+
### Model Description
|
| 62 |
+
- **Model Type:** SpanMarker
|
| 63 |
+
<!-- - **Encoder:** [Unknown](https://huggingface.co/unknown) -->
|
| 64 |
+
- **Maximum Sequence Length:** 256 tokens
|
| 65 |
+
- **Maximum Entity Length:** 8 words
|
| 66 |
+
- **Training Dataset:** [DFKI-SLT/few-nerd](https://huggingface.co/datasets/DFKI-SLT/few-nerd)
|
| 67 |
+
<!-- - **Language:** Unknown -->
|
| 68 |
+
<!-- - **License:** Unknown -->
|
| 69 |
+
|
| 70 |
+
### Model Sources
|
| 71 |
+
|
| 72 |
+
- **Repository:** [SpanMarker on GitHub](https://github.com/tomaarsen/SpanMarkerNER)
|
| 73 |
+
- **Thesis:** [SpanMarker For Named Entity Recognition](https://raw.githubusercontent.com/tomaarsen/SpanMarkerNER/main/thesis.pdf)
|
| 74 |
+
|
| 75 |
+
### Model Labels
|
| 76 |
+
| Label | Examples |
|
| 77 |
+
|:-----------------------------------------|:---------------------------------------------------------------------------------------------------------|
|
| 78 |
+
| art-broadcastprogram | "Street Cents", "Corazones", "The Gale Storm Show : Oh , Susanna" |
|
| 79 |
+
| art-film | "L'Atlantide", "Shawshank Redemption", "Bosch" |
|
| 80 |
+
| art-music | "Champion Lover", "Atkinson , Danko and Ford ( with Brockie and Hilton )", "Hollywood Studio Symphony" |
|
| 81 |
+
| art-other | "Aphrodite of Milos", "The Today Show", "Venus de Milo" |
|
| 82 |
+
| art-painting | "Production/Reproduction", "Cofiwch Dryweryn", "Touit" |
|
| 83 |
+
| art-writtenart | "Time", "Imelda de ' Lambertazzi", "The Seven Year Itch" |
|
| 84 |
+
| building-airport | "Sheremetyevo International Airport", "Luton Airport", "Newark Liberty International Airport" |
|
| 85 |
+
| building-hospital | "Yeungnam University Hospital", "Memorial Sloan-Kettering Cancer Center", "Hokkaido University Hospital" |
|
| 86 |
+
| building-hotel | "Radisson Blu Sea Plaza Hotel", "Flamingo Hotel", "The Standard Hotel" |
|
| 87 |
+
| building-library | "British Library", "Berlin State Library", "Bayerische Staatsbibliothek" |
|
| 88 |
+
| building-other | "Communiplex", "Henry Ford Museum", "Alpha Recording Studios" |
|
| 89 |
+
| building-restaurant | "Carnegie Deli", "Trumbull", "Fatburger" |
|
| 90 |
+
| building-sportsfacility | "Sports Center", "Boston Garden", "Glenn Warner Soccer Facility" |
|
| 91 |
+
| building-theater | "Sanders Theatre", "Pittsburgh Civic Light Opera", "National Paris Opera" |
|
| 92 |
+
| event-attack/battle/war/militaryconflict | "Vietnam War", "Jurist", "Easter Offensive" |
|
| 93 |
+
| event-disaster | "1990s North Korean famine", "the 1912 North Mount Lyell Disaster", "1693 Sicily earthquake" |
|
| 94 |
+
| event-election | "1982 Mitcham and Morden by-election", "Elections to the European Parliament", "March 1898 elections" |
|
| 95 |
+
| event-other | "Eastwood Scoring Stage", "Union for a Popular Movement", "Masaryk Democratic Movement" |
|
| 96 |
+
| event-protest | "French Revolution", "Iranian Constitutional Revolution", "Russian Revolution" |
|
| 97 |
+
| event-sportsevent | "World Cup", "National Champions", "Stanley Cup" |
|
| 98 |
+
| location-GPE | "Mediterranean Basin", "the Republic of Croatia", "Croatian" |
|
| 99 |
+
| location-bodiesofwater | "Arthur Kill", "Atatürk Dam Lake", "Norfolk coast" |
|
| 100 |
+
| location-island | "Staten Island", "new Samsat district", "Laccadives" |
|
| 101 |
+
| location-mountain | "Miteirya Ridge", "Ruweisat Ridge", "Salamander Glacier" |
|
| 102 |
+
| location-other | "Northern City Line", "Victoria line", "Cartuther" |
|
| 103 |
+
| location-park | "Painted Desert Community Complex Historic District", "Gramercy Park", "Shenandoah National Park" |
|
| 104 |
+
| location-road/railway/highway/transit | "NJT", "Newark-Elizabeth Rail Link", "Friern Barnet Road" |
|
| 105 |
+
| organization-company | "Church 's Chicken", "Texas Chicken", "Dixy Chicken" |
|
| 106 |
+
| organization-education | "Barnard College", "MIT", "Belfast Royal Academy and the Ulster College of Physical Education" |
|
| 107 |
+
| organization-government/governmentagency | "Diet", "Supreme Court", "Congregazione dei Nobili" |
|
| 108 |
+
| organization-media/newspaper | "Al Jazeera", "Clash", "TimeOut Melbourne" |
|
| 109 |
+
| organization-other | "Defence Sector C", "4th Army", "IAEA" |
|
| 110 |
+
| organization-politicalparty | "Al Wafa ' Islamic", "Shimpotō", "Kenseitō" |
|
| 111 |
+
| organization-religion | "Jewish", "UPCUSA", "Christian" |
|
| 112 |
+
| organization-showorganization | "Mr. Mister", "Lizzy", "Bochumer Symphoniker" |
|
| 113 |
+
| organization-sportsleague | "NHL", "First Division", "China League One" |
|
| 114 |
+
| organization-sportsteam | "Arsenal", "Luc Alphand Aventures", "Tottenham" |
|
| 115 |
+
| other-astronomything | "Algol", "Zodiac", "`` Caput Larvae ''" |
|
| 116 |
+
| other-award | "Order of the Republic of Guinea and Nigeria", "GCON", "Grand Commander of the Order of the Niger" |
|
| 117 |
+
| other-biologything | "Amphiphysin", "BAR", "N-terminal lipid" |
|
| 118 |
+
| other-chemicalthing | "sulfur", "uranium", "carbon dioxide" |
|
| 119 |
+
| other-currency | "$", "Travancore Rupee", "lac crore" |
|
| 120 |
+
| other-disease | "hypothyroidism", "bladder cancer", "French Dysentery Epidemic of 1779" |
|
| 121 |
+
| other-educationaldegree | "BSc ( Hons ) in physics", "Master", "Bachelor" |
|
| 122 |
+
| other-god | "El", "Raijin", "Fujin" |
|
| 123 |
+
| other-language | "Latin", "English", "Breton-speaking" |
|
| 124 |
+
| other-law | "United States Freedom Support Act", "Thirty Years ' Peace", "Leahy–Smith America Invents Act ( AIA" |
|
| 125 |
+
| other-livingthing | "insects", "monkeys", "patchouli" |
|
| 126 |
+
| other-medical | "pediatrician", "Pediatrics", "amitriptyline" |
|
| 127 |
+
| person-actor | "Edmund Payne", "Tchéky Karyo", "Ellaline Terriss" |
|
| 128 |
+
| person-artist/author | "Gaetano Donizett", "George Axelrod", "Hicks" |
|
| 129 |
+
| person-athlete | "Tozawa", "Jaguar", "Neville" |
|
| 130 |
+
| person-director | "Bob Swaim", "Frank Darabont", "Richard Quine" |
|
| 131 |
+
| person-other | "Holden", "Richard Benson", "Campbell" |
|
| 132 |
+
| person-politician | "Rivière", "Emeric", "William" |
|
| 133 |
+
| person-scholar | "Stalmine", "Wurdack", "Stedman" |
|
| 134 |
+
| person-soldier | "Krukenberg", "Joachim Ziegler", "Helmuth Weidling" |
|
| 135 |
+
| product-airplane | "EC135T2 CPDS", "Spey-equipped FGR.2s", "Luton" |
|
| 136 |
+
| product-car | "100EX", "Corvettes - GT1 C6R", "Phantom" |
|
| 137 |
+
| product-food | "yakiniku", "V. labrusca", "red grape" |
|
| 138 |
+
| product-game | "Airforce Delta", "Splinter Cell", "Hardcore RPG" |
|
| 139 |
+
| product-other | "X11", "Fairbottom Bobs", "PDP-1" |
|
| 140 |
+
| product-ship | "Essex", "HMS `` Chinkara ''", "Congress" |
|
| 141 |
+
| product-software | "Wikipedia", "Apdf", "AmiPDF" |
|
| 142 |
+
| product-train | "High Speed Trains", "Royal Scots Grey", "55022" |
|
| 143 |
+
| product-weapon | "ZU-23-2M Wróbel", "AR-15 's", "ZU-23-2MR Wróbel II" |
|
| 144 |
+
|
| 145 |
+
## Evaluation
|
| 146 |
+
|
| 147 |
+
### Metrics
|
| 148 |
+
| Label | Precision | Recall | F1 |
|
| 149 |
+
|:-----------------------------------------|:----------|:-------|:-------|
|
| 150 |
+
| **all** | 0.7034 | 0.7027 | 0.7031 |
|
| 151 |
+
| art-broadcastprogram | 0.6024 | 0.5904 | 0.5963 |
|
| 152 |
+
| art-film | 0.7761 | 0.7533 | 0.7645 |
|
| 153 |
+
| art-music | 0.7825 | 0.7551 | 0.7685 |
|
| 154 |
+
| art-other | 0.4193 | 0.3327 | 0.3710 |
|
| 155 |
+
| art-painting | 0.5882 | 0.5263 | 0.5556 |
|
| 156 |
+
| art-writtenart | 0.6819 | 0.6488 | 0.6649 |
|
| 157 |
+
| building-airport | 0.8064 | 0.8352 | 0.8205 |
|
| 158 |
+
| building-hospital | 0.7282 | 0.8022 | 0.7634 |
|
| 159 |
+
| building-hotel | 0.7033 | 0.7245 | 0.7138 |
|
| 160 |
+
| building-library | 0.7550 | 0.7380 | 0.7464 |
|
| 161 |
+
| building-other | 0.5867 | 0.5840 | 0.5853 |
|
| 162 |
+
| building-restaurant | 0.6205 | 0.5216 | 0.5667 |
|
| 163 |
+
| building-sportsfacility | 0.6113 | 0.7976 | 0.6921 |
|
| 164 |
+
| building-theater | 0.7060 | 0.7495 | 0.7271 |
|
| 165 |
+
| event-attack/battle/war/militaryconflict | 0.7945 | 0.7395 | 0.7660 |
|
| 166 |
+
| event-disaster | 0.5604 | 0.5604 | 0.5604 |
|
| 167 |
+
| event-election | 0.4286 | 0.1484 | 0.2204 |
|
| 168 |
+
| event-other | 0.4885 | 0.4400 | 0.4629 |
|
| 169 |
+
| event-protest | 0.3798 | 0.4759 | 0.4225 |
|
| 170 |
+
| event-sportsevent | 0.6198 | 0.6162 | 0.6180 |
|
| 171 |
+
| location-GPE | 0.8157 | 0.8552 | 0.8350 |
|
| 172 |
+
| location-bodiesofwater | 0.7268 | 0.7690 | 0.7473 |
|
| 173 |
+
| location-island | 0.7504 | 0.6842 | 0.7158 |
|
| 174 |
+
| location-mountain | 0.7352 | 0.7298 | 0.7325 |
|
| 175 |
+
| location-other | 0.4427 | 0.3104 | 0.3649 |
|
| 176 |
+
| location-park | 0.7153 | 0.6856 | 0.7001 |
|
| 177 |
+
| location-road/railway/highway/transit | 0.7090 | 0.7324 | 0.7205 |
|
| 178 |
+
| organization-company | 0.6963 | 0.7061 | 0.7012 |
|
| 179 |
+
| organization-education | 0.7994 | 0.7986 | 0.7990 |
|
| 180 |
+
| organization-government/governmentagency | 0.5524 | 0.4533 | 0.4980 |
|
| 181 |
+
| organization-media/newspaper | 0.6513 | 0.6656 | 0.6584 |
|
| 182 |
+
| organization-other | 0.5978 | 0.5375 | 0.5661 |
|
| 183 |
+
| organization-politicalparty | 0.6793 | 0.7315 | 0.7044 |
|
| 184 |
+
| organization-religion | 0.5575 | 0.6131 | 0.5840 |
|
| 185 |
+
| organization-showorganization | 0.6035 | 0.5839 | 0.5935 |
|
| 186 |
+
| organization-sportsleague | 0.6393 | 0.6610 | 0.6499 |
|
| 187 |
+
| organization-sportsteam | 0.7259 | 0.7796 | 0.7518 |
|
| 188 |
+
| other-astronomything | 0.7794 | 0.8024 | 0.7907 |
|
| 189 |
+
| other-award | 0.7180 | 0.6649 | 0.6904 |
|
| 190 |
+
| other-biologything | 0.6864 | 0.6238 | 0.6536 |
|
| 191 |
+
| other-chemicalthing | 0.5688 | 0.6036 | 0.5856 |
|
| 192 |
+
| other-currency | 0.6996 | 0.8423 | 0.7643 |
|
| 193 |
+
| other-disease | 0.6591 | 0.7410 | 0.6977 |
|
| 194 |
+
| other-educationaldegree | 0.6114 | 0.6198 | 0.6156 |
|
| 195 |
+
| other-god | 0.6486 | 0.7181 | 0.6816 |
|
| 196 |
+
| other-language | 0.6507 | 0.8313 | 0.7300 |
|
| 197 |
+
| other-law | 0.6934 | 0.7331 | 0.7127 |
|
| 198 |
+
| other-livingthing | 0.6019 | 0.6605 | 0.6298 |
|
| 199 |
+
| other-medical | 0.5124 | 0.5214 | 0.5169 |
|
| 200 |
+
| person-actor | 0.8384 | 0.8051 | 0.8214 |
|
| 201 |
+
| person-artist/author | 0.7122 | 0.7531 | 0.7321 |
|
| 202 |
+
| person-athlete | 0.8318 | 0.8422 | 0.8370 |
|
| 203 |
+
| person-director | 0.7083 | 0.7365 | 0.7221 |
|
| 204 |
+
| person-other | 0.6833 | 0.6737 | 0.6785 |
|
| 205 |
+
| person-politician | 0.6807 | 0.6836 | 0.6822 |
|
| 206 |
+
| person-scholar | 0.5397 | 0.5209 | 0.5301 |
|
| 207 |
+
| person-soldier | 0.5053 | 0.5920 | 0.5452 |
|
| 208 |
+
| product-airplane | 0.6617 | 0.6692 | 0.6654 |
|
| 209 |
+
| product-car | 0.7313 | 0.7132 | 0.7222 |
|
| 210 |
+
| product-food | 0.5787 | 0.5787 | 0.5787 |
|
| 211 |
+
| product-game | 0.7364 | 0.7140 | 0.7250 |
|
| 212 |
+
| product-other | 0.5567 | 0.4210 | 0.4795 |
|
| 213 |
+
| product-ship | 0.6842 | 0.6842 | 0.6842 |
|
| 214 |
+
| product-software | 0.6495 | 0.6648 | 0.6570 |
|
| 215 |
+
| product-train | 0.5942 | 0.5924 | 0.5933 |
|
| 216 |
+
| product-weapon | 0.6435 | 0.5353 | 0.5844 |
|
| 217 |
+
|
| 218 |
+
## Uses
|
| 219 |
+
|
| 220 |
+
### Direct Use for Inference
|
| 221 |
+
|
| 222 |
+
```python
|
| 223 |
+
from span_marker import SpanMarkerModel
|
| 224 |
+
|
| 225 |
+
# Download from the 🤗 Hub
|
| 226 |
+
model = SpanMarkerModel.from_pretrained("supreethrao/instructNER_fewnerd_xl")
|
| 227 |
+
# Run inference
|
| 228 |
+
entities = model.predict("The Sunday Edition is a television programme broadcast on the ITV Network in the United Kingdom focusing on political interview and discussion, produced by ITV Productions.")
|
| 229 |
+
```
|
| 230 |
+
|
| 231 |
+
### Downstream Use
|
| 232 |
+
You can finetune this model on your own dataset.
|
| 233 |
+
|
| 234 |
+
<details><summary>Click to expand</summary>
|
| 235 |
+
|
| 236 |
+
```python
|
| 237 |
+
from span_marker import SpanMarkerModel, Trainer
|
| 238 |
+
|
| 239 |
+
# Download from the 🤗 Hub
|
| 240 |
+
model = SpanMarkerModel.from_pretrained("supreethrao/instructNER_fewnerd_xl")
|
| 241 |
+
|
| 242 |
+
# Specify a Dataset with "tokens" and "ner_tag" columns
|
| 243 |
+
dataset = load_dataset("conll2003") # For example CoNLL2003
|
| 244 |
+
|
| 245 |
+
# Initialize a Trainer using the pretrained model & dataset
|
| 246 |
+
trainer = Trainer(
|
| 247 |
+
model=model,
|
| 248 |
+
train_dataset=dataset["train"],
|
| 249 |
+
eval_dataset=dataset["validation"],
|
| 250 |
+
)
|
| 251 |
+
trainer.train()
|
| 252 |
+
trainer.save_model("supreethrao/instructNER_fewnerd_xl-finetuned")
|
| 253 |
+
```
|
| 254 |
+
</details>
|
| 255 |
+
|
| 256 |
+
<!--
|
| 257 |
+
### Out-of-Scope Use
|
| 258 |
+
|
| 259 |
+
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
| 260 |
+
-->
|
| 261 |
+
|
| 262 |
+
<!--
|
| 263 |
+
## Bias, Risks and Limitations
|
| 264 |
+
|
| 265 |
+
*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
|
| 266 |
+
-->
|
| 267 |
+
|
| 268 |
+
<!--
|
| 269 |
+
### Recommendations
|
| 270 |
+
|
| 271 |
+
*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
|
| 272 |
+
-->
|
| 273 |
+
|
| 274 |
+
## Training Details
|
| 275 |
+
|
| 276 |
+
### Training Set Metrics
|
| 277 |
+
| Training set | Min | Median | Max |
|
| 278 |
+
|:----------------------|:----|:--------|:----|
|
| 279 |
+
| Sentence length | 1 | 24.4945 | 267 |
|
| 280 |
+
| Entities per sentence | 0 | 2.5832 | 88 |
|
| 281 |
+
|
| 282 |
+
### Training Hyperparameters
|
| 283 |
+
- learning_rate: 5e-05
|
| 284 |
+
- train_batch_size: 16
|
| 285 |
+
- eval_batch_size: 16
|
| 286 |
+
- seed: 42
|
| 287 |
+
- distributed_type: multi-GPU
|
| 288 |
+
- num_devices: 2
|
| 289 |
+
- total_train_batch_size: 32
|
| 290 |
+
- total_eval_batch_size: 32
|
| 291 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
| 292 |
+
- lr_scheduler_type: linear
|
| 293 |
+
- lr_scheduler_warmup_ratio: 0.1
|
| 294 |
+
- num_epochs: 3
|
| 295 |
+
- mixed_precision_training: Native AMP
|
| 296 |
+
|
| 297 |
+
### Framework Versions
|
| 298 |
+
- Python: 3.10.13
|
| 299 |
+
- SpanMarker: 1.5.0
|
| 300 |
+
- Transformers: 4.35.2
|
| 301 |
+
- PyTorch: 2.1.1
|
| 302 |
+
- Datasets: 2.15.0
|
| 303 |
+
- Tokenizers: 0.15.0
|
| 304 |
+
|
| 305 |
+
## Citation
|
| 306 |
+
|
| 307 |
+
### BibTeX
|
| 308 |
+
```
|
| 309 |
+
@software{Aarsen_SpanMarker,
|
| 310 |
+
author = {Aarsen, Tom},
|
| 311 |
+
license = {Apache-2.0},
|
| 312 |
+
title = {{SpanMarker for Named Entity Recognition}},
|
| 313 |
+
url = {https://github.com/tomaarsen/SpanMarkerNER}
|
| 314 |
+
}
|
| 315 |
+
```
|
| 316 |
+
|
| 317 |
+
<!--
|
| 318 |
+
## Glossary
|
| 319 |
+
|
| 320 |
+
*Clearly define terms in order to be accessible across audiences.*
|
| 321 |
+
-->
|
| 322 |
+
|
| 323 |
+
<!--
|
| 324 |
+
## Model Card Authors
|
| 325 |
+
|
| 326 |
+
*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
|
| 327 |
+
-->
|
| 328 |
+
|
| 329 |
+
<!--
|
| 330 |
+
## Model Card Contact
|
| 331 |
+
|
| 332 |
+
*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
|
| 333 |
+
-->
|
all_results.json
ADDED
|
@@ -0,0 +1,407 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"epoch": 3.0,
|
| 3 |
+
"test_art-broadcastprogram": {
|
| 4 |
+
"f1": 0.5963149078726968,
|
| 5 |
+
"number": 603,
|
| 6 |
+
"precision": 0.6023688663282571,
|
| 7 |
+
"recall": 0.5903814262023217
|
| 8 |
+
},
|
| 9 |
+
"test_art-film": {
|
| 10 |
+
"f1": 0.7645466847090664,
|
| 11 |
+
"number": 750,
|
| 12 |
+
"precision": 0.7760989010989011,
|
| 13 |
+
"recall": 0.7533333333333333
|
| 14 |
+
},
|
| 15 |
+
"test_art-music": {
|
| 16 |
+
"f1": 0.7685459940652819,
|
| 17 |
+
"number": 1029,
|
| 18 |
+
"precision": 0.7824773413897281,
|
| 19 |
+
"recall": 0.7551020408163265
|
| 20 |
+
},
|
| 21 |
+
"test_art-other": {
|
| 22 |
+
"f1": 0.37103174603174605,
|
| 23 |
+
"number": 562,
|
| 24 |
+
"precision": 0.4192825112107623,
|
| 25 |
+
"recall": 0.33274021352313166
|
| 26 |
+
},
|
| 27 |
+
"test_art-painting": {
|
| 28 |
+
"f1": 0.5555555555555555,
|
| 29 |
+
"number": 57,
|
| 30 |
+
"precision": 0.5882352941176471,
|
| 31 |
+
"recall": 0.5263157894736842
|
| 32 |
+
},
|
| 33 |
+
"test_art-writtenart": {
|
| 34 |
+
"f1": 0.6649020645844362,
|
| 35 |
+
"number": 968,
|
| 36 |
+
"precision": 0.6818675352877307,
|
| 37 |
+
"recall": 0.6487603305785123
|
| 38 |
+
},
|
| 39 |
+
"test_building-airport": {
|
| 40 |
+
"f1": 0.8205128205128206,
|
| 41 |
+
"number": 364,
|
| 42 |
+
"precision": 0.8063660477453581,
|
| 43 |
+
"recall": 0.8351648351648352
|
| 44 |
+
},
|
| 45 |
+
"test_building-hospital": {
|
| 46 |
+
"f1": 0.7633986928104576,
|
| 47 |
+
"number": 364,
|
| 48 |
+
"precision": 0.7281795511221946,
|
| 49 |
+
"recall": 0.8021978021978022
|
| 50 |
+
},
|
| 51 |
+
"test_building-hotel": {
|
| 52 |
+
"f1": 0.7137546468401488,
|
| 53 |
+
"number": 265,
|
| 54 |
+
"precision": 0.7032967032967034,
|
| 55 |
+
"recall": 0.7245283018867924
|
| 56 |
+
},
|
| 57 |
+
"test_building-library": {
|
| 58 |
+
"f1": 0.7464387464387464,
|
| 59 |
+
"number": 355,
|
| 60 |
+
"precision": 0.7550432276657061,
|
| 61 |
+
"recall": 0.7380281690140845
|
| 62 |
+
},
|
| 63 |
+
"test_building-other": {
|
| 64 |
+
"f1": 0.5853370122191565,
|
| 65 |
+
"number": 2543,
|
| 66 |
+
"precision": 0.5867246147767681,
|
| 67 |
+
"recall": 0.5839559575304758
|
| 68 |
+
},
|
| 69 |
+
"test_building-restaurant": {
|
| 70 |
+
"f1": 0.5667447306791569,
|
| 71 |
+
"number": 232,
|
| 72 |
+
"precision": 0.6205128205128205,
|
| 73 |
+
"recall": 0.521551724137931
|
| 74 |
+
},
|
| 75 |
+
"test_building-sportsfacility": {
|
| 76 |
+
"f1": 0.6921487603305786,
|
| 77 |
+
"number": 420,
|
| 78 |
+
"precision": 0.6113138686131386,
|
| 79 |
+
"recall": 0.7976190476190477
|
| 80 |
+
},
|
| 81 |
+
"test_building-theater": {
|
| 82 |
+
"f1": 0.7270788912579957,
|
| 83 |
+
"number": 455,
|
| 84 |
+
"precision": 0.7060041407867494,
|
| 85 |
+
"recall": 0.7494505494505495
|
| 86 |
+
},
|
| 87 |
+
"test_event-attack/battle/war/militaryconflict": {
|
| 88 |
+
"f1": 0.7660377358490565,
|
| 89 |
+
"number": 1098,
|
| 90 |
+
"precision": 0.7945205479452054,
|
| 91 |
+
"recall": 0.7395264116575592
|
| 92 |
+
},
|
| 93 |
+
"test_event-disaster": {
|
| 94 |
+
"f1": 0.5603864734299517,
|
| 95 |
+
"number": 207,
|
| 96 |
+
"precision": 0.5603864734299517,
|
| 97 |
+
"recall": 0.5603864734299517
|
| 98 |
+
},
|
| 99 |
+
"test_event-election": {
|
| 100 |
+
"f1": 0.22040816326530616,
|
| 101 |
+
"number": 182,
|
| 102 |
+
"precision": 0.42857142857142855,
|
| 103 |
+
"recall": 0.14835164835164835
|
| 104 |
+
},
|
| 105 |
+
"test_event-other": {
|
| 106 |
+
"f1": 0.4629404617253949,
|
| 107 |
+
"number": 866,
|
| 108 |
+
"precision": 0.48846153846153845,
|
| 109 |
+
"recall": 0.4399538106235566
|
| 110 |
+
},
|
| 111 |
+
"test_event-protest": {
|
| 112 |
+
"f1": 0.42245989304812837,
|
| 113 |
+
"number": 166,
|
| 114 |
+
"precision": 0.3798076923076923,
|
| 115 |
+
"recall": 0.4759036144578313
|
| 116 |
+
},
|
| 117 |
+
"test_event-sportsevent": {
|
| 118 |
+
"f1": 0.6179955171309639,
|
| 119 |
+
"number": 1566,
|
| 120 |
+
"precision": 0.619781631342325,
|
| 121 |
+
"recall": 0.6162196679438059
|
| 122 |
+
},
|
| 123 |
+
"test_location-GPE": {
|
| 124 |
+
"f1": 0.8349881570447639,
|
| 125 |
+
"number": 20405,
|
| 126 |
+
"precision": 0.8157255048616305,
|
| 127 |
+
"recall": 0.8551825532957609
|
| 128 |
+
},
|
| 129 |
+
"test_location-bodiesofwater": {
|
| 130 |
+
"f1": 0.7472984206151289,
|
| 131 |
+
"number": 1169,
|
| 132 |
+
"precision": 0.7267582861762328,
|
| 133 |
+
"recall": 0.7690333618477331
|
| 134 |
+
},
|
| 135 |
+
"test_location-island": {
|
| 136 |
+
"f1": 0.7157894736842105,
|
| 137 |
+
"number": 646,
|
| 138 |
+
"precision": 0.7504244482173175,
|
| 139 |
+
"recall": 0.6842105263157895
|
| 140 |
+
},
|
| 141 |
+
"test_location-mountain": {
|
| 142 |
+
"f1": 0.7324981577008107,
|
| 143 |
+
"number": 681,
|
| 144 |
+
"precision": 0.735207100591716,
|
| 145 |
+
"recall": 0.7298091042584435
|
| 146 |
+
},
|
| 147 |
+
"test_location-other": {
|
| 148 |
+
"f1": 0.36490474912798493,
|
| 149 |
+
"number": 2191,
|
| 150 |
+
"precision": 0.4427083333333333,
|
| 151 |
+
"recall": 0.31036056595162026
|
| 152 |
+
},
|
| 153 |
+
"test_location-park": {
|
| 154 |
+
"f1": 0.7001114827201783,
|
| 155 |
+
"number": 458,
|
| 156 |
+
"precision": 0.715261958997722,
|
| 157 |
+
"recall": 0.6855895196506551
|
| 158 |
+
},
|
| 159 |
+
"test_location-road/railway/highway/transit": {
|
| 160 |
+
"f1": 0.7204861111111112,
|
| 161 |
+
"number": 1700,
|
| 162 |
+
"precision": 0.708997722095672,
|
| 163 |
+
"recall": 0.7323529411764705
|
| 164 |
+
},
|
| 165 |
+
"test_loss": 0.022335968911647797,
|
| 166 |
+
"test_organization-company": {
|
| 167 |
+
"f1": 0.7011596788581624,
|
| 168 |
+
"number": 3896,
|
| 169 |
+
"precision": 0.6962794229309036,
|
| 170 |
+
"recall": 0.7061088295687885
|
| 171 |
+
},
|
| 172 |
+
"test_organization-education": {
|
| 173 |
+
"f1": 0.7990314769975787,
|
| 174 |
+
"number": 2066,
|
| 175 |
+
"precision": 0.7994186046511628,
|
| 176 |
+
"recall": 0.7986447241045499
|
| 177 |
+
},
|
| 178 |
+
"test_organization-government/governmentagency": {
|
| 179 |
+
"f1": 0.49800072700836057,
|
| 180 |
+
"number": 1511,
|
| 181 |
+
"precision": 0.5524193548387096,
|
| 182 |
+
"recall": 0.45334215751158174
|
| 183 |
+
},
|
| 184 |
+
"test_organization-media/newspaper": {
|
| 185 |
+
"f1": 0.6583701324769169,
|
| 186 |
+
"number": 1232,
|
| 187 |
+
"precision": 0.6513105639396346,
|
| 188 |
+
"recall": 0.6655844155844156
|
| 189 |
+
},
|
| 190 |
+
"test_organization-other": {
|
| 191 |
+
"f1": 0.5660735468564649,
|
| 192 |
+
"number": 4439,
|
| 193 |
+
"precision": 0.59784515159108,
|
| 194 |
+
"recall": 0.5375084478486145
|
| 195 |
+
},
|
| 196 |
+
"test_organization-politicalparty": {
|
| 197 |
+
"f1": 0.704431247144815,
|
| 198 |
+
"number": 1054,
|
| 199 |
+
"precision": 0.6792951541850221,
|
| 200 |
+
"recall": 0.7314990512333965
|
| 201 |
+
},
|
| 202 |
+
"test_organization-religion": {
|
| 203 |
+
"f1": 0.583982990786676,
|
| 204 |
+
"number": 672,
|
| 205 |
+
"precision": 0.557510148849797,
|
| 206 |
+
"recall": 0.6130952380952381
|
| 207 |
+
},
|
| 208 |
+
"test_organization-showorganization": {
|
| 209 |
+
"f1": 0.5935228023793787,
|
| 210 |
+
"number": 769,
|
| 211 |
+
"precision": 0.603494623655914,
|
| 212 |
+
"recall": 0.5838751625487646
|
| 213 |
+
},
|
| 214 |
+
"test_organization-sportsleague": {
|
| 215 |
+
"f1": 0.6499442586399109,
|
| 216 |
+
"number": 882,
|
| 217 |
+
"precision": 0.6392543859649122,
|
| 218 |
+
"recall": 0.6609977324263039
|
| 219 |
+
},
|
| 220 |
+
"test_organization-sportsteam": {
|
| 221 |
+
"f1": 0.7518034704620783,
|
| 222 |
+
"number": 2473,
|
| 223 |
+
"precision": 0.7259036144578314,
|
| 224 |
+
"recall": 0.779619894864537
|
| 225 |
+
},
|
| 226 |
+
"test_other-astronomything": {
|
| 227 |
+
"f1": 0.7906976744186047,
|
| 228 |
+
"number": 678,
|
| 229 |
+
"precision": 0.7793696275071633,
|
| 230 |
+
"recall": 0.8023598820058997
|
| 231 |
+
},
|
| 232 |
+
"test_other-award": {
|
| 233 |
+
"f1": 0.6903954802259886,
|
| 234 |
+
"number": 919,
|
| 235 |
+
"precision": 0.717978848413631,
|
| 236 |
+
"recall": 0.6648531011969532
|
| 237 |
+
},
|
| 238 |
+
"test_other-biologything": {
|
| 239 |
+
"f1": 0.6536203522504893,
|
| 240 |
+
"number": 1874,
|
| 241 |
+
"precision": 0.6864357017028773,
|
| 242 |
+
"recall": 0.6237993596584845
|
| 243 |
+
},
|
| 244 |
+
"test_other-chemicalthing": {
|
| 245 |
+
"f1": 0.5856459330143541,
|
| 246 |
+
"number": 1014,
|
| 247 |
+
"precision": 0.5687732342007435,
|
| 248 |
+
"recall": 0.6035502958579881
|
| 249 |
+
},
|
| 250 |
+
"test_other-currency": {
|
| 251 |
+
"f1": 0.7643384440658716,
|
| 252 |
+
"number": 799,
|
| 253 |
+
"precision": 0.6995841995841996,
|
| 254 |
+
"recall": 0.8423028785982478
|
| 255 |
+
},
|
| 256 |
+
"test_other-disease": {
|
| 257 |
+
"f1": 0.6976744186046512,
|
| 258 |
+
"number": 749,
|
| 259 |
+
"precision": 0.6591448931116389,
|
| 260 |
+
"recall": 0.7409879839786382
|
| 261 |
+
},
|
| 262 |
+
"test_other-educationaldegree": {
|
| 263 |
+
"f1": 0.615595075239398,
|
| 264 |
+
"number": 363,
|
| 265 |
+
"precision": 0.6114130434782609,
|
| 266 |
+
"recall": 0.6198347107438017
|
| 267 |
+
},
|
| 268 |
+
"test_other-god": {
|
| 269 |
+
"f1": 0.6816143497757848,
|
| 270 |
+
"number": 635,
|
| 271 |
+
"precision": 0.6486486486486487,
|
| 272 |
+
"recall": 0.7181102362204724
|
| 273 |
+
},
|
| 274 |
+
"test_other-language": {
|
| 275 |
+
"f1": 0.7300291545189505,
|
| 276 |
+
"number": 753,
|
| 277 |
+
"precision": 0.6507276507276507,
|
| 278 |
+
"recall": 0.8313413014608234
|
| 279 |
+
},
|
| 280 |
+
"test_other-law": {
|
| 281 |
+
"f1": 0.7126673532440783,
|
| 282 |
+
"number": 472,
|
| 283 |
+
"precision": 0.6933867735470942,
|
| 284 |
+
"recall": 0.7330508474576272
|
| 285 |
+
},
|
| 286 |
+
"test_other-livingthing": {
|
| 287 |
+
"f1": 0.6298342541436465,
|
| 288 |
+
"number": 863,
|
| 289 |
+
"precision": 0.6019007391763463,
|
| 290 |
+
"recall": 0.660486674391657
|
| 291 |
+
},
|
| 292 |
+
"test_other-medical": {
|
| 293 |
+
"f1": 0.5168539325842697,
|
| 294 |
+
"number": 397,
|
| 295 |
+
"precision": 0.5123762376237624,
|
| 296 |
+
"recall": 0.5214105793450882
|
| 297 |
+
},
|
| 298 |
+
"test_overall_accuracy": 0.9256893595441806,
|
| 299 |
+
"test_overall_f1": 0.703084859534267,
|
| 300 |
+
"test_overall_precision": 0.7034273336857051,
|
| 301 |
+
"test_overall_recall": 0.7027427186979075,
|
| 302 |
+
"test_person-actor": {
|
| 303 |
+
"f1": 0.8214397008413836,
|
| 304 |
+
"number": 1637,
|
| 305 |
+
"precision": 0.8384223918575063,
|
| 306 |
+
"recall": 0.8051313378130727
|
| 307 |
+
},
|
| 308 |
+
"test_person-artist/author": {
|
| 309 |
+
"f1": 0.7320701754385964,
|
| 310 |
+
"number": 3463,
|
| 311 |
+
"precision": 0.7121791370835608,
|
| 312 |
+
"recall": 0.7531042448743863
|
| 313 |
+
},
|
| 314 |
+
"test_person-athlete": {
|
| 315 |
+
"f1": 0.8370089593383873,
|
| 316 |
+
"number": 2884,
|
| 317 |
+
"precision": 0.8318493150684931,
|
| 318 |
+
"recall": 0.8422330097087378
|
| 319 |
+
},
|
| 320 |
+
"test_person-director": {
|
| 321 |
+
"f1": 0.7221238938053098,
|
| 322 |
+
"number": 554,
|
| 323 |
+
"precision": 0.7083333333333334,
|
| 324 |
+
"recall": 0.7364620938628159
|
| 325 |
+
},
|
| 326 |
+
"test_person-other": {
|
| 327 |
+
"f1": 0.6784606547960942,
|
| 328 |
+
"number": 8767,
|
| 329 |
+
"precision": 0.6833275483049867,
|
| 330 |
+
"recall": 0.6736625983802897
|
| 331 |
+
},
|
| 332 |
+
"test_person-politician": {
|
| 333 |
+
"f1": 0.6821515892420537,
|
| 334 |
+
"number": 2857,
|
| 335 |
+
"precision": 0.6807249912861624,
|
| 336 |
+
"recall": 0.6835841792089604
|
| 337 |
+
},
|
| 338 |
+
"test_person-scholar": {
|
| 339 |
+
"f1": 0.5301369863013699,
|
| 340 |
+
"number": 743,
|
| 341 |
+
"precision": 0.5397489539748954,
|
| 342 |
+
"recall": 0.5208613728129206
|
| 343 |
+
},
|
| 344 |
+
"test_person-soldier": {
|
| 345 |
+
"f1": 0.5451957295373664,
|
| 346 |
+
"number": 647,
|
| 347 |
+
"precision": 0.5052770448548812,
|
| 348 |
+
"recall": 0.5919629057187017
|
| 349 |
+
},
|
| 350 |
+
"test_product-airplane": {
|
| 351 |
+
"f1": 0.6654111738857501,
|
| 352 |
+
"number": 792,
|
| 353 |
+
"precision": 0.66167290886392,
|
| 354 |
+
"recall": 0.6691919191919192
|
| 355 |
+
},
|
| 356 |
+
"test_product-car": {
|
| 357 |
+
"f1": 0.7221812822402359,
|
| 358 |
+
"number": 687,
|
| 359 |
+
"precision": 0.7313432835820896,
|
| 360 |
+
"recall": 0.7132459970887919
|
| 361 |
+
},
|
| 362 |
+
"test_product-food": {
|
| 363 |
+
"f1": 0.5787037037037037,
|
| 364 |
+
"number": 432,
|
| 365 |
+
"precision": 0.5787037037037037,
|
| 366 |
+
"recall": 0.5787037037037037
|
| 367 |
+
},
|
| 368 |
+
"test_product-game": {
|
| 369 |
+
"f1": 0.7250257466529352,
|
| 370 |
+
"number": 493,
|
| 371 |
+
"precision": 0.7364016736401674,
|
| 372 |
+
"recall": 0.7139959432048681
|
| 373 |
+
},
|
| 374 |
+
"test_product-other": {
|
| 375 |
+
"f1": 0.4794617563739376,
|
| 376 |
+
"number": 1608,
|
| 377 |
+
"precision": 0.5567434210526315,
|
| 378 |
+
"recall": 0.4210199004975124
|
| 379 |
+
},
|
| 380 |
+
"test_product-ship": {
|
| 381 |
+
"f1": 0.6842105263157895,
|
| 382 |
+
"number": 380,
|
| 383 |
+
"precision": 0.6842105263157895,
|
| 384 |
+
"recall": 0.6842105263157895
|
| 385 |
+
},
|
| 386 |
+
"test_product-software": {
|
| 387 |
+
"f1": 0.6570316842690384,
|
| 388 |
+
"number": 889,
|
| 389 |
+
"precision": 0.6494505494505495,
|
| 390 |
+
"recall": 0.6647919010123734
|
| 391 |
+
},
|
| 392 |
+
"test_product-train": {
|
| 393 |
+
"f1": 0.5933014354066984,
|
| 394 |
+
"number": 314,
|
| 395 |
+
"precision": 0.5942492012779552,
|
| 396 |
+
"recall": 0.5923566878980892
|
| 397 |
+
},
|
| 398 |
+
"test_product-weapon": {
|
| 399 |
+
"f1": 0.584426946631671,
|
| 400 |
+
"number": 624,
|
| 401 |
+
"precision": 0.6435452793834296,
|
| 402 |
+
"recall": 0.5352564102564102
|
| 403 |
+
},
|
| 404 |
+
"test_runtime": 242.3784,
|
| 405 |
+
"test_samples_per_second": 189.732,
|
| 406 |
+
"test_steps_per_second": 5.933
|
| 407 |
+
}
|
final_checkpoint/README.md
ADDED
|
@@ -0,0 +1,333 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
library_name: span-marker
|
| 3 |
+
tags:
|
| 4 |
+
- span-marker
|
| 5 |
+
- token-classification
|
| 6 |
+
- ner
|
| 7 |
+
- named-entity-recognition
|
| 8 |
+
- generated_from_span_marker_trainer
|
| 9 |
+
datasets:
|
| 10 |
+
- DFKI-SLT/few-nerd
|
| 11 |
+
metrics:
|
| 12 |
+
- precision
|
| 13 |
+
- recall
|
| 14 |
+
- f1
|
| 15 |
+
widget:
|
| 16 |
+
- text: In response, in May or June 1125, a 3,000-strong Crusader coalition commanded
|
| 17 |
+
by King Baldwin II of Jerusalem confronted and defeated the 15,000-strong Muslim
|
| 18 |
+
coalition at the Battle of Azaz, raising the siege of the town.
|
| 19 |
+
- text: Cardenal made several visits to Jesuit universities in the United States,
|
| 20 |
+
including the University of Detroit Mercy in 2013, and the John Carroll University
|
| 21 |
+
in 2014.
|
| 22 |
+
- text: Other super-spreaders, defined as those that transmit SARS to at least eight
|
| 23 |
+
other people, included the incidents at the Hotel Metropole in Hong Kong, the
|
| 24 |
+
Amoy Gardens apartment complex in Hong Kong and one in an acute care hospital
|
| 25 |
+
in Toronto, Ontario, Canada.
|
| 26 |
+
- text: The District Court for the Northern District of California rejected 321 Studios'
|
| 27 |
+
claims for declaratory relief, holding that both DVD Copy Plus and DVD-X Copy
|
| 28 |
+
violated the DMCA and that the DMCA was not unconstitutional.
|
| 29 |
+
- text: The Sunday Edition is a television programme broadcast on the ITV Network
|
| 30 |
+
in the United Kingdom focusing on political interview and discussion, produced
|
| 31 |
+
by ITV Productions.
|
| 32 |
+
pipeline_tag: token-classification
|
| 33 |
+
model-index:
|
| 34 |
+
- name: SpanMarker
|
| 35 |
+
results:
|
| 36 |
+
- task:
|
| 37 |
+
type: token-classification
|
| 38 |
+
name: Named Entity Recognition
|
| 39 |
+
dataset:
|
| 40 |
+
name: Unknown
|
| 41 |
+
type: DFKI-SLT/few-nerd
|
| 42 |
+
split: test
|
| 43 |
+
metrics:
|
| 44 |
+
- type: f1
|
| 45 |
+
value: 0.703084859534267
|
| 46 |
+
name: F1
|
| 47 |
+
- type: precision
|
| 48 |
+
value: 0.7034273336857051
|
| 49 |
+
name: Precision
|
| 50 |
+
- type: recall
|
| 51 |
+
value: 0.7027427186979075
|
| 52 |
+
name: Recall
|
| 53 |
+
---
|
| 54 |
+
|
| 55 |
+
# SpanMarker
|
| 56 |
+
|
| 57 |
+
This is a [SpanMarker](https://github.com/tomaarsen/SpanMarkerNER) model trained on the [DFKI-SLT/few-nerd](https://huggingface.co/datasets/DFKI-SLT/few-nerd) dataset that can be used for Named Entity Recognition.
|
| 58 |
+
|
| 59 |
+
## Model Details
|
| 60 |
+
|
| 61 |
+
### Model Description
|
| 62 |
+
- **Model Type:** SpanMarker
|
| 63 |
+
<!-- - **Encoder:** [Unknown](https://huggingface.co/unknown) -->
|
| 64 |
+
- **Maximum Sequence Length:** 256 tokens
|
| 65 |
+
- **Maximum Entity Length:** 8 words
|
| 66 |
+
- **Training Dataset:** [DFKI-SLT/few-nerd](https://huggingface.co/datasets/DFKI-SLT/few-nerd)
|
| 67 |
+
<!-- - **Language:** Unknown -->
|
| 68 |
+
<!-- - **License:** Unknown -->
|
| 69 |
+
|
| 70 |
+
### Model Sources
|
| 71 |
+
|
| 72 |
+
- **Repository:** [SpanMarker on GitHub](https://github.com/tomaarsen/SpanMarkerNER)
|
| 73 |
+
- **Thesis:** [SpanMarker For Named Entity Recognition](https://raw.githubusercontent.com/tomaarsen/SpanMarkerNER/main/thesis.pdf)
|
| 74 |
+
|
| 75 |
+
### Model Labels
|
| 76 |
+
| Label | Examples |
|
| 77 |
+
|:-----------------------------------------|:---------------------------------------------------------------------------------------------------------|
|
| 78 |
+
| art-broadcastprogram | "Street Cents", "Corazones", "The Gale Storm Show : Oh , Susanna" |
|
| 79 |
+
| art-film | "L'Atlantide", "Shawshank Redemption", "Bosch" |
|
| 80 |
+
| art-music | "Champion Lover", "Atkinson , Danko and Ford ( with Brockie and Hilton )", "Hollywood Studio Symphony" |
|
| 81 |
+
| art-other | "Aphrodite of Milos", "The Today Show", "Venus de Milo" |
|
| 82 |
+
| art-painting | "Production/Reproduction", "Cofiwch Dryweryn", "Touit" |
|
| 83 |
+
| art-writtenart | "Time", "Imelda de ' Lambertazzi", "The Seven Year Itch" |
|
| 84 |
+
| building-airport | "Sheremetyevo International Airport", "Luton Airport", "Newark Liberty International Airport" |
|
| 85 |
+
| building-hospital | "Yeungnam University Hospital", "Memorial Sloan-Kettering Cancer Center", "Hokkaido University Hospital" |
|
| 86 |
+
| building-hotel | "Radisson Blu Sea Plaza Hotel", "Flamingo Hotel", "The Standard Hotel" |
|
| 87 |
+
| building-library | "British Library", "Berlin State Library", "Bayerische Staatsbibliothek" |
|
| 88 |
+
| building-other | "Communiplex", "Henry Ford Museum", "Alpha Recording Studios" |
|
| 89 |
+
| building-restaurant | "Carnegie Deli", "Trumbull", "Fatburger" |
|
| 90 |
+
| building-sportsfacility | "Sports Center", "Boston Garden", "Glenn Warner Soccer Facility" |
|
| 91 |
+
| building-theater | "Sanders Theatre", "Pittsburgh Civic Light Opera", "National Paris Opera" |
|
| 92 |
+
| event-attack/battle/war/militaryconflict | "Vietnam War", "Jurist", "Easter Offensive" |
|
| 93 |
+
| event-disaster | "1990s North Korean famine", "the 1912 North Mount Lyell Disaster", "1693 Sicily earthquake" |
|
| 94 |
+
| event-election | "1982 Mitcham and Morden by-election", "Elections to the European Parliament", "March 1898 elections" |
|
| 95 |
+
| event-other | "Eastwood Scoring Stage", "Union for a Popular Movement", "Masaryk Democratic Movement" |
|
| 96 |
+
| event-protest | "French Revolution", "Iranian Constitutional Revolution", "Russian Revolution" |
|
| 97 |
+
| event-sportsevent | "World Cup", "National Champions", "Stanley Cup" |
|
| 98 |
+
| location-GPE | "Mediterranean Basin", "the Republic of Croatia", "Croatian" |
|
| 99 |
+
| location-bodiesofwater | "Arthur Kill", "Atatürk Dam Lake", "Norfolk coast" |
|
| 100 |
+
| location-island | "Staten Island", "new Samsat district", "Laccadives" |
|
| 101 |
+
| location-mountain | "Miteirya Ridge", "Ruweisat Ridge", "Salamander Glacier" |
|
| 102 |
+
| location-other | "Northern City Line", "Victoria line", "Cartuther" |
|
| 103 |
+
| location-park | "Painted Desert Community Complex Historic District", "Gramercy Park", "Shenandoah National Park" |
|
| 104 |
+
| location-road/railway/highway/transit | "NJT", "Newark-Elizabeth Rail Link", "Friern Barnet Road" |
|
| 105 |
+
| organization-company | "Church 's Chicken", "Texas Chicken", "Dixy Chicken" |
|
| 106 |
+
| organization-education | "Barnard College", "MIT", "Belfast Royal Academy and the Ulster College of Physical Education" |
|
| 107 |
+
| organization-government/governmentagency | "Diet", "Supreme Court", "Congregazione dei Nobili" |
|
| 108 |
+
| organization-media/newspaper | "Al Jazeera", "Clash", "TimeOut Melbourne" |
|
| 109 |
+
| organization-other | "Defence Sector C", "4th Army", "IAEA" |
|
| 110 |
+
| organization-politicalparty | "Al Wafa ' Islamic", "Shimpotō", "Kenseitō" |
|
| 111 |
+
| organization-religion | "Jewish", "UPCUSA", "Christian" |
|
| 112 |
+
| organization-showorganization | "Mr. Mister", "Lizzy", "Bochumer Symphoniker" |
|
| 113 |
+
| organization-sportsleague | "NHL", "First Division", "China League One" |
|
| 114 |
+
| organization-sportsteam | "Arsenal", "Luc Alphand Aventures", "Tottenham" |
|
| 115 |
+
| other-astronomything | "Algol", "Zodiac", "`` Caput Larvae ''" |
|
| 116 |
+
| other-award | "Order of the Republic of Guinea and Nigeria", "GCON", "Grand Commander of the Order of the Niger" |
|
| 117 |
+
| other-biologything | "Amphiphysin", "BAR", "N-terminal lipid" |
|
| 118 |
+
| other-chemicalthing | "sulfur", "uranium", "carbon dioxide" |
|
| 119 |
+
| other-currency | "$", "Travancore Rupee", "lac crore" |
|
| 120 |
+
| other-disease | "hypothyroidism", "bladder cancer", "French Dysentery Epidemic of 1779" |
|
| 121 |
+
| other-educationaldegree | "BSc ( Hons ) in physics", "Master", "Bachelor" |
|
| 122 |
+
| other-god | "El", "Raijin", "Fujin" |
|
| 123 |
+
| other-language | "Latin", "English", "Breton-speaking" |
|
| 124 |
+
| other-law | "United States Freedom Support Act", "Thirty Years ' Peace", "Leahy–Smith America Invents Act ( AIA" |
|
| 125 |
+
| other-livingthing | "insects", "monkeys", "patchouli" |
|
| 126 |
+
| other-medical | "pediatrician", "Pediatrics", "amitriptyline" |
|
| 127 |
+
| person-actor | "Edmund Payne", "Tchéky Karyo", "Ellaline Terriss" |
|
| 128 |
+
| person-artist/author | "Gaetano Donizett", "George Axelrod", "Hicks" |
|
| 129 |
+
| person-athlete | "Tozawa", "Jaguar", "Neville" |
|
| 130 |
+
| person-director | "Bob Swaim", "Frank Darabont", "Richard Quine" |
|
| 131 |
+
| person-other | "Holden", "Richard Benson", "Campbell" |
|
| 132 |
+
| person-politician | "Rivière", "Emeric", "William" |
|
| 133 |
+
| person-scholar | "Stalmine", "Wurdack", "Stedman" |
|
| 134 |
+
| person-soldier | "Krukenberg", "Joachim Ziegler", "Helmuth Weidling" |
|
| 135 |
+
| product-airplane | "EC135T2 CPDS", "Spey-equipped FGR.2s", "Luton" |
|
| 136 |
+
| product-car | "100EX", "Corvettes - GT1 C6R", "Phantom" |
|
| 137 |
+
| product-food | "yakiniku", "V. labrusca", "red grape" |
|
| 138 |
+
| product-game | "Airforce Delta", "Splinter Cell", "Hardcore RPG" |
|
| 139 |
+
| product-other | "X11", "Fairbottom Bobs", "PDP-1" |
|
| 140 |
+
| product-ship | "Essex", "HMS `` Chinkara ''", "Congress" |
|
| 141 |
+
| product-software | "Wikipedia", "Apdf", "AmiPDF" |
|
| 142 |
+
| product-train | "High Speed Trains", "Royal Scots Grey", "55022" |
|
| 143 |
+
| product-weapon | "ZU-23-2M Wróbel", "AR-15 's", "ZU-23-2MR Wróbel II" |
|
| 144 |
+
|
| 145 |
+
## Evaluation
|
| 146 |
+
|
| 147 |
+
### Metrics
|
| 148 |
+
| Label | Precision | Recall | F1 |
|
| 149 |
+
|:-----------------------------------------|:----------|:-------|:-------|
|
| 150 |
+
| **all** | 0.7034 | 0.7027 | 0.7031 |
|
| 151 |
+
| art-broadcastprogram | 0.6024 | 0.5904 | 0.5963 |
|
| 152 |
+
| art-film | 0.7761 | 0.7533 | 0.7645 |
|
| 153 |
+
| art-music | 0.7825 | 0.7551 | 0.7685 |
|
| 154 |
+
| art-other | 0.4193 | 0.3327 | 0.3710 |
|
| 155 |
+
| art-painting | 0.5882 | 0.5263 | 0.5556 |
|
| 156 |
+
| art-writtenart | 0.6819 | 0.6488 | 0.6649 |
|
| 157 |
+
| building-airport | 0.8064 | 0.8352 | 0.8205 |
|
| 158 |
+
| building-hospital | 0.7282 | 0.8022 | 0.7634 |
|
| 159 |
+
| building-hotel | 0.7033 | 0.7245 | 0.7138 |
|
| 160 |
+
| building-library | 0.7550 | 0.7380 | 0.7464 |
|
| 161 |
+
| building-other | 0.5867 | 0.5840 | 0.5853 |
|
| 162 |
+
| building-restaurant | 0.6205 | 0.5216 | 0.5667 |
|
| 163 |
+
| building-sportsfacility | 0.6113 | 0.7976 | 0.6921 |
|
| 164 |
+
| building-theater | 0.7060 | 0.7495 | 0.7271 |
|
| 165 |
+
| event-attack/battle/war/militaryconflict | 0.7945 | 0.7395 | 0.7660 |
|
| 166 |
+
| event-disaster | 0.5604 | 0.5604 | 0.5604 |
|
| 167 |
+
| event-election | 0.4286 | 0.1484 | 0.2204 |
|
| 168 |
+
| event-other | 0.4885 | 0.4400 | 0.4629 |
|
| 169 |
+
| event-protest | 0.3798 | 0.4759 | 0.4225 |
|
| 170 |
+
| event-sportsevent | 0.6198 | 0.6162 | 0.6180 |
|
| 171 |
+
| location-GPE | 0.8157 | 0.8552 | 0.8350 |
|
| 172 |
+
| location-bodiesofwater | 0.7268 | 0.7690 | 0.7473 |
|
| 173 |
+
| location-island | 0.7504 | 0.6842 | 0.7158 |
|
| 174 |
+
| location-mountain | 0.7352 | 0.7298 | 0.7325 |
|
| 175 |
+
| location-other | 0.4427 | 0.3104 | 0.3649 |
|
| 176 |
+
| location-park | 0.7153 | 0.6856 | 0.7001 |
|
| 177 |
+
| location-road/railway/highway/transit | 0.7090 | 0.7324 | 0.7205 |
|
| 178 |
+
| organization-company | 0.6963 | 0.7061 | 0.7012 |
|
| 179 |
+
| organization-education | 0.7994 | 0.7986 | 0.7990 |
|
| 180 |
+
| organization-government/governmentagency | 0.5524 | 0.4533 | 0.4980 |
|
| 181 |
+
| organization-media/newspaper | 0.6513 | 0.6656 | 0.6584 |
|
| 182 |
+
| organization-other | 0.5978 | 0.5375 | 0.5661 |
|
| 183 |
+
| organization-politicalparty | 0.6793 | 0.7315 | 0.7044 |
|
| 184 |
+
| organization-religion | 0.5575 | 0.6131 | 0.5840 |
|
| 185 |
+
| organization-showorganization | 0.6035 | 0.5839 | 0.5935 |
|
| 186 |
+
| organization-sportsleague | 0.6393 | 0.6610 | 0.6499 |
|
| 187 |
+
| organization-sportsteam | 0.7259 | 0.7796 | 0.7518 |
|
| 188 |
+
| other-astronomything | 0.7794 | 0.8024 | 0.7907 |
|
| 189 |
+
| other-award | 0.7180 | 0.6649 | 0.6904 |
|
| 190 |
+
| other-biologything | 0.6864 | 0.6238 | 0.6536 |
|
| 191 |
+
| other-chemicalthing | 0.5688 | 0.6036 | 0.5856 |
|
| 192 |
+
| other-currency | 0.6996 | 0.8423 | 0.7643 |
|
| 193 |
+
| other-disease | 0.6591 | 0.7410 | 0.6977 |
|
| 194 |
+
| other-educationaldegree | 0.6114 | 0.6198 | 0.6156 |
|
| 195 |
+
| other-god | 0.6486 | 0.7181 | 0.6816 |
|
| 196 |
+
| other-language | 0.6507 | 0.8313 | 0.7300 |
|
| 197 |
+
| other-law | 0.6934 | 0.7331 | 0.7127 |
|
| 198 |
+
| other-livingthing | 0.6019 | 0.6605 | 0.6298 |
|
| 199 |
+
| other-medical | 0.5124 | 0.5214 | 0.5169 |
|
| 200 |
+
| person-actor | 0.8384 | 0.8051 | 0.8214 |
|
| 201 |
+
| person-artist/author | 0.7122 | 0.7531 | 0.7321 |
|
| 202 |
+
| person-athlete | 0.8318 | 0.8422 | 0.8370 |
|
| 203 |
+
| person-director | 0.7083 | 0.7365 | 0.7221 |
|
| 204 |
+
| person-other | 0.6833 | 0.6737 | 0.6785 |
|
| 205 |
+
| person-politician | 0.6807 | 0.6836 | 0.6822 |
|
| 206 |
+
| person-scholar | 0.5397 | 0.5209 | 0.5301 |
|
| 207 |
+
| person-soldier | 0.5053 | 0.5920 | 0.5452 |
|
| 208 |
+
| product-airplane | 0.6617 | 0.6692 | 0.6654 |
|
| 209 |
+
| product-car | 0.7313 | 0.7132 | 0.7222 |
|
| 210 |
+
| product-food | 0.5787 | 0.5787 | 0.5787 |
|
| 211 |
+
| product-game | 0.7364 | 0.7140 | 0.7250 |
|
| 212 |
+
| product-other | 0.5567 | 0.4210 | 0.4795 |
|
| 213 |
+
| product-ship | 0.6842 | 0.6842 | 0.6842 |
|
| 214 |
+
| product-software | 0.6495 | 0.6648 | 0.6570 |
|
| 215 |
+
| product-train | 0.5942 | 0.5924 | 0.5933 |
|
| 216 |
+
| product-weapon | 0.6435 | 0.5353 | 0.5844 |
|
| 217 |
+
|
| 218 |
+
## Uses
|
| 219 |
+
|
| 220 |
+
### Direct Use for Inference
|
| 221 |
+
|
| 222 |
+
```python
|
| 223 |
+
from span_marker import SpanMarkerModel
|
| 224 |
+
|
| 225 |
+
# Download from the 🤗 Hub
|
| 226 |
+
model = SpanMarkerModel.from_pretrained("supreethrao/instructNER_fewnerd_xl")
|
| 227 |
+
# Run inference
|
| 228 |
+
entities = model.predict("The Sunday Edition is a television programme broadcast on the ITV Network in the United Kingdom focusing on political interview and discussion, produced by ITV Productions.")
|
| 229 |
+
```
|
| 230 |
+
|
| 231 |
+
### Downstream Use
|
| 232 |
+
You can finetune this model on your own dataset.
|
| 233 |
+
|
| 234 |
+
<details><summary>Click to expand</summary>
|
| 235 |
+
|
| 236 |
+
```python
|
| 237 |
+
from span_marker import SpanMarkerModel, Trainer
|
| 238 |
+
|
| 239 |
+
# Download from the 🤗 Hub
|
| 240 |
+
model = SpanMarkerModel.from_pretrained("supreethrao/instructNER_fewnerd_xl")
|
| 241 |
+
|
| 242 |
+
# Specify a Dataset with "tokens" and "ner_tag" columns
|
| 243 |
+
dataset = load_dataset("conll2003") # For example CoNLL2003
|
| 244 |
+
|
| 245 |
+
# Initialize a Trainer using the pretrained model & dataset
|
| 246 |
+
trainer = Trainer(
|
| 247 |
+
model=model,
|
| 248 |
+
train_dataset=dataset["train"],
|
| 249 |
+
eval_dataset=dataset["validation"],
|
| 250 |
+
)
|
| 251 |
+
trainer.train()
|
| 252 |
+
trainer.save_model("supreethrao/instructNER_fewnerd_xl-finetuned")
|
| 253 |
+
```
|
| 254 |
+
</details>
|
| 255 |
+
|
| 256 |
+
<!--
|
| 257 |
+
### Out-of-Scope Use
|
| 258 |
+
|
| 259 |
+
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
| 260 |
+
-->
|
| 261 |
+
|
| 262 |
+
<!--
|
| 263 |
+
## Bias, Risks and Limitations
|
| 264 |
+
|
| 265 |
+
*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
|
| 266 |
+
-->
|
| 267 |
+
|
| 268 |
+
<!--
|
| 269 |
+
### Recommendations
|
| 270 |
+
|
| 271 |
+
*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
|
| 272 |
+
-->
|
| 273 |
+
|
| 274 |
+
## Training Details
|
| 275 |
+
|
| 276 |
+
### Training Set Metrics
|
| 277 |
+
| Training set | Min | Median | Max |
|
| 278 |
+
|:----------------------|:----|:--------|:----|
|
| 279 |
+
| Sentence length | 1 | 24.4945 | 267 |
|
| 280 |
+
| Entities per sentence | 0 | 2.5832 | 88 |
|
| 281 |
+
|
| 282 |
+
### Training Hyperparameters
|
| 283 |
+
- learning_rate: 5e-05
|
| 284 |
+
- train_batch_size: 16
|
| 285 |
+
- eval_batch_size: 16
|
| 286 |
+
- seed: 42
|
| 287 |
+
- distributed_type: multi-GPU
|
| 288 |
+
- num_devices: 2
|
| 289 |
+
- total_train_batch_size: 32
|
| 290 |
+
- total_eval_batch_size: 32
|
| 291 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
| 292 |
+
- lr_scheduler_type: linear
|
| 293 |
+
- lr_scheduler_warmup_ratio: 0.1
|
| 294 |
+
- num_epochs: 3
|
| 295 |
+
- mixed_precision_training: Native AMP
|
| 296 |
+
|
| 297 |
+
### Framework Versions
|
| 298 |
+
- Python: 3.10.13
|
| 299 |
+
- SpanMarker: 1.5.0
|
| 300 |
+
- Transformers: 4.35.2
|
| 301 |
+
- PyTorch: 2.1.1
|
| 302 |
+
- Datasets: 2.15.0
|
| 303 |
+
- Tokenizers: 0.15.0
|
| 304 |
+
|
| 305 |
+
## Citation
|
| 306 |
+
|
| 307 |
+
### BibTeX
|
| 308 |
+
```
|
| 309 |
+
@software{Aarsen_SpanMarker,
|
| 310 |
+
author = {Aarsen, Tom},
|
| 311 |
+
license = {Apache-2.0},
|
| 312 |
+
title = {{SpanMarker for Named Entity Recognition}},
|
| 313 |
+
url = {https://github.com/tomaarsen/SpanMarkerNER}
|
| 314 |
+
}
|
| 315 |
+
```
|
| 316 |
+
|
| 317 |
+
<!--
|
| 318 |
+
## Glossary
|
| 319 |
+
|
| 320 |
+
*Clearly define terms in order to be accessible across audiences.*
|
| 321 |
+
-->
|
| 322 |
+
|
| 323 |
+
<!--
|
| 324 |
+
## Model Card Authors
|
| 325 |
+
|
| 326 |
+
*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
|
| 327 |
+
-->
|
| 328 |
+
|
| 329 |
+
<!--
|
| 330 |
+
## Model Card Contact
|
| 331 |
+
|
| 332 |
+
*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
|
| 333 |
+
-->
|
final_checkpoint/added_tokens.json
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"<end>": 50266,
|
| 3 |
+
"<start>": 50265
|
| 4 |
+
}
|
final_checkpoint/config.json
ADDED
|
@@ -0,0 +1,228 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"SpanMarkerModel"
|
| 4 |
+
],
|
| 5 |
+
"encoder": {
|
| 6 |
+
"_name_or_path": "numind/generic-entity_recognition_NER-v1",
|
| 7 |
+
"add_cross_attention": false,
|
| 8 |
+
"architectures": [
|
| 9 |
+
"RobertaModel"
|
| 10 |
+
],
|
| 11 |
+
"attention_probs_dropout_prob": 0.1,
|
| 12 |
+
"bad_words_ids": null,
|
| 13 |
+
"begin_suppress_tokens": null,
|
| 14 |
+
"bos_token_id": 0,
|
| 15 |
+
"chunk_size_feed_forward": 0,
|
| 16 |
+
"classifier_dropout": null,
|
| 17 |
+
"cross_attention_hidden_size": null,
|
| 18 |
+
"decoder_start_token_id": null,
|
| 19 |
+
"diversity_penalty": 0.0,
|
| 20 |
+
"do_sample": false,
|
| 21 |
+
"early_stopping": false,
|
| 22 |
+
"encoder_no_repeat_ngram_size": 0,
|
| 23 |
+
"eos_token_id": 2,
|
| 24 |
+
"exponential_decay_length_penalty": null,
|
| 25 |
+
"finetuning_task": null,
|
| 26 |
+
"forced_bos_token_id": null,
|
| 27 |
+
"forced_eos_token_id": null,
|
| 28 |
+
"hidden_act": "gelu",
|
| 29 |
+
"hidden_dropout_prob": 0.1,
|
| 30 |
+
"hidden_size": 768,
|
| 31 |
+
"id2label": {
|
| 32 |
+
"0": "O",
|
| 33 |
+
"1": "art-broadcastprogram",
|
| 34 |
+
"2": "art-film",
|
| 35 |
+
"3": "art-music",
|
| 36 |
+
"4": "art-other",
|
| 37 |
+
"5": "art-painting",
|
| 38 |
+
"6": "art-writtenart",
|
| 39 |
+
"7": "building-airport",
|
| 40 |
+
"8": "building-hospital",
|
| 41 |
+
"9": "building-hotel",
|
| 42 |
+
"10": "building-library",
|
| 43 |
+
"11": "building-other",
|
| 44 |
+
"12": "building-restaurant",
|
| 45 |
+
"13": "building-sportsfacility",
|
| 46 |
+
"14": "building-theater",
|
| 47 |
+
"15": "event-attack/battle/war/militaryconflict",
|
| 48 |
+
"16": "event-disaster",
|
| 49 |
+
"17": "event-election",
|
| 50 |
+
"18": "event-other",
|
| 51 |
+
"19": "event-protest",
|
| 52 |
+
"20": "event-sportsevent",
|
| 53 |
+
"21": "location-GPE",
|
| 54 |
+
"22": "location-bodiesofwater",
|
| 55 |
+
"23": "location-island",
|
| 56 |
+
"24": "location-mountain",
|
| 57 |
+
"25": "location-other",
|
| 58 |
+
"26": "location-park",
|
| 59 |
+
"27": "location-road/railway/highway/transit",
|
| 60 |
+
"28": "organization-company",
|
| 61 |
+
"29": "organization-education",
|
| 62 |
+
"30": "organization-government/governmentagency",
|
| 63 |
+
"31": "organization-media/newspaper",
|
| 64 |
+
"32": "organization-other",
|
| 65 |
+
"33": "organization-politicalparty",
|
| 66 |
+
"34": "organization-religion",
|
| 67 |
+
"35": "organization-showorganization",
|
| 68 |
+
"36": "organization-sportsleague",
|
| 69 |
+
"37": "organization-sportsteam",
|
| 70 |
+
"38": "other-astronomything",
|
| 71 |
+
"39": "other-award",
|
| 72 |
+
"40": "other-biologything",
|
| 73 |
+
"41": "other-chemicalthing",
|
| 74 |
+
"42": "other-currency",
|
| 75 |
+
"43": "other-disease",
|
| 76 |
+
"44": "other-educationaldegree",
|
| 77 |
+
"45": "other-god",
|
| 78 |
+
"46": "other-language",
|
| 79 |
+
"47": "other-law",
|
| 80 |
+
"48": "other-livingthing",
|
| 81 |
+
"49": "other-medical",
|
| 82 |
+
"50": "person-actor",
|
| 83 |
+
"51": "person-artist/author",
|
| 84 |
+
"52": "person-athlete",
|
| 85 |
+
"53": "person-director",
|
| 86 |
+
"54": "person-other",
|
| 87 |
+
"55": "person-politician",
|
| 88 |
+
"56": "person-scholar",
|
| 89 |
+
"57": "person-soldier",
|
| 90 |
+
"58": "product-airplane",
|
| 91 |
+
"59": "product-car",
|
| 92 |
+
"60": "product-food",
|
| 93 |
+
"61": "product-game",
|
| 94 |
+
"62": "product-other",
|
| 95 |
+
"63": "product-ship",
|
| 96 |
+
"64": "product-software",
|
| 97 |
+
"65": "product-train",
|
| 98 |
+
"66": "product-weapon"
|
| 99 |
+
},
|
| 100 |
+
"initializer_range": 0.02,
|
| 101 |
+
"intermediate_size": 3072,
|
| 102 |
+
"is_decoder": false,
|
| 103 |
+
"is_encoder_decoder": false,
|
| 104 |
+
"label2id": {
|
| 105 |
+
"O": 0,
|
| 106 |
+
"art-broadcastprogram": 1,
|
| 107 |
+
"art-film": 2,
|
| 108 |
+
"art-music": 3,
|
| 109 |
+
"art-other": 4,
|
| 110 |
+
"art-painting": 5,
|
| 111 |
+
"art-writtenart": 6,
|
| 112 |
+
"building-airport": 7,
|
| 113 |
+
"building-hospital": 8,
|
| 114 |
+
"building-hotel": 9,
|
| 115 |
+
"building-library": 10,
|
| 116 |
+
"building-other": 11,
|
| 117 |
+
"building-restaurant": 12,
|
| 118 |
+
"building-sportsfacility": 13,
|
| 119 |
+
"building-theater": 14,
|
| 120 |
+
"event-attack/battle/war/militaryconflict": 15,
|
| 121 |
+
"event-disaster": 16,
|
| 122 |
+
"event-election": 17,
|
| 123 |
+
"event-other": 18,
|
| 124 |
+
"event-protest": 19,
|
| 125 |
+
"event-sportsevent": 20,
|
| 126 |
+
"location-GPE": 21,
|
| 127 |
+
"location-bodiesofwater": 22,
|
| 128 |
+
"location-island": 23,
|
| 129 |
+
"location-mountain": 24,
|
| 130 |
+
"location-other": 25,
|
| 131 |
+
"location-park": 26,
|
| 132 |
+
"location-road/railway/highway/transit": 27,
|
| 133 |
+
"organization-company": 28,
|
| 134 |
+
"organization-education": 29,
|
| 135 |
+
"organization-government/governmentagency": 30,
|
| 136 |
+
"organization-media/newspaper": 31,
|
| 137 |
+
"organization-other": 32,
|
| 138 |
+
"organization-politicalparty": 33,
|
| 139 |
+
"organization-religion": 34,
|
| 140 |
+
"organization-showorganization": 35,
|
| 141 |
+
"organization-sportsleague": 36,
|
| 142 |
+
"organization-sportsteam": 37,
|
| 143 |
+
"other-astronomything": 38,
|
| 144 |
+
"other-award": 39,
|
| 145 |
+
"other-biologything": 40,
|
| 146 |
+
"other-chemicalthing": 41,
|
| 147 |
+
"other-currency": 42,
|
| 148 |
+
"other-disease": 43,
|
| 149 |
+
"other-educationaldegree": 44,
|
| 150 |
+
"other-god": 45,
|
| 151 |
+
"other-language": 46,
|
| 152 |
+
"other-law": 47,
|
| 153 |
+
"other-livingthing": 48,
|
| 154 |
+
"other-medical": 49,
|
| 155 |
+
"person-actor": 50,
|
| 156 |
+
"person-artist/author": 51,
|
| 157 |
+
"person-athlete": 52,
|
| 158 |
+
"person-director": 53,
|
| 159 |
+
"person-other": 54,
|
| 160 |
+
"person-politician": 55,
|
| 161 |
+
"person-scholar": 56,
|
| 162 |
+
"person-soldier": 57,
|
| 163 |
+
"product-airplane": 58,
|
| 164 |
+
"product-car": 59,
|
| 165 |
+
"product-food": 60,
|
| 166 |
+
"product-game": 61,
|
| 167 |
+
"product-other": 62,
|
| 168 |
+
"product-ship": 63,
|
| 169 |
+
"product-software": 64,
|
| 170 |
+
"product-train": 65,
|
| 171 |
+
"product-weapon": 66
|
| 172 |
+
},
|
| 173 |
+
"layer_norm_eps": 1e-05,
|
| 174 |
+
"length_penalty": 1.0,
|
| 175 |
+
"max_length": 20,
|
| 176 |
+
"max_position_embeddings": 514,
|
| 177 |
+
"min_length": 0,
|
| 178 |
+
"model_type": "roberta",
|
| 179 |
+
"no_repeat_ngram_size": 0,
|
| 180 |
+
"num_attention_heads": 12,
|
| 181 |
+
"num_beam_groups": 1,
|
| 182 |
+
"num_beams": 1,
|
| 183 |
+
"num_hidden_layers": 12,
|
| 184 |
+
"num_return_sequences": 1,
|
| 185 |
+
"output_attentions": false,
|
| 186 |
+
"output_hidden_states": false,
|
| 187 |
+
"output_scores": false,
|
| 188 |
+
"pad_token_id": 1,
|
| 189 |
+
"position_embedding_type": "absolute",
|
| 190 |
+
"prefix": null,
|
| 191 |
+
"problem_type": null,
|
| 192 |
+
"pruned_heads": {},
|
| 193 |
+
"remove_invalid_values": false,
|
| 194 |
+
"repetition_penalty": 1.0,
|
| 195 |
+
"return_dict": true,
|
| 196 |
+
"return_dict_in_generate": false,
|
| 197 |
+
"sep_token_id": null,
|
| 198 |
+
"suppress_tokens": null,
|
| 199 |
+
"task_specific_params": null,
|
| 200 |
+
"temperature": 1.0,
|
| 201 |
+
"tf_legacy_loss": false,
|
| 202 |
+
"tie_encoder_decoder": false,
|
| 203 |
+
"tie_word_embeddings": true,
|
| 204 |
+
"tokenizer_class": null,
|
| 205 |
+
"top_k": 50,
|
| 206 |
+
"top_p": 1.0,
|
| 207 |
+
"torch_dtype": "float32",
|
| 208 |
+
"torchscript": false,
|
| 209 |
+
"transformers_version": "4.35.2",
|
| 210 |
+
"type_vocab_size": 1,
|
| 211 |
+
"typical_p": 1.0,
|
| 212 |
+
"use_bfloat16": false,
|
| 213 |
+
"use_cache": true,
|
| 214 |
+
"vocab_size": 50272
|
| 215 |
+
},
|
| 216 |
+
"entity_max_length": 8,
|
| 217 |
+
"marker_max_length": 128,
|
| 218 |
+
"max_next_context": null,
|
| 219 |
+
"max_prev_context": null,
|
| 220 |
+
"model_max_length": 256,
|
| 221 |
+
"model_max_length_default": 512,
|
| 222 |
+
"model_type": "span-marker",
|
| 223 |
+
"span_marker_version": "1.5.0",
|
| 224 |
+
"torch_dtype": "float32",
|
| 225 |
+
"trained_with_document_context": false,
|
| 226 |
+
"transformers_version": "4.35.2",
|
| 227 |
+
"vocab_size": 50272
|
| 228 |
+
}
|
final_checkpoint/merges.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
final_checkpoint/model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b1f2a83a5a877bf7c79b2dd581c8ee4ded91ae0fa501cadf1d3f12deb1be8fde
|
| 3 |
+
size 499040084
|
final_checkpoint/special_tokens_map.json
ADDED
|
@@ -0,0 +1,51 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token": {
|
| 3 |
+
"content": "<s>",
|
| 4 |
+
"lstrip": false,
|
| 5 |
+
"normalized": true,
|
| 6 |
+
"rstrip": false,
|
| 7 |
+
"single_word": false
|
| 8 |
+
},
|
| 9 |
+
"cls_token": {
|
| 10 |
+
"content": "<s>",
|
| 11 |
+
"lstrip": false,
|
| 12 |
+
"normalized": true,
|
| 13 |
+
"rstrip": false,
|
| 14 |
+
"single_word": false
|
| 15 |
+
},
|
| 16 |
+
"eos_token": {
|
| 17 |
+
"content": "</s>",
|
| 18 |
+
"lstrip": false,
|
| 19 |
+
"normalized": true,
|
| 20 |
+
"rstrip": false,
|
| 21 |
+
"single_word": false
|
| 22 |
+
},
|
| 23 |
+
"mask_token": {
|
| 24 |
+
"content": "<mask>",
|
| 25 |
+
"lstrip": true,
|
| 26 |
+
"normalized": true,
|
| 27 |
+
"rstrip": false,
|
| 28 |
+
"single_word": false
|
| 29 |
+
},
|
| 30 |
+
"pad_token": {
|
| 31 |
+
"content": "<pad>",
|
| 32 |
+
"lstrip": false,
|
| 33 |
+
"normalized": true,
|
| 34 |
+
"rstrip": false,
|
| 35 |
+
"single_word": false
|
| 36 |
+
},
|
| 37 |
+
"sep_token": {
|
| 38 |
+
"content": "</s>",
|
| 39 |
+
"lstrip": false,
|
| 40 |
+
"normalized": true,
|
| 41 |
+
"rstrip": false,
|
| 42 |
+
"single_word": false
|
| 43 |
+
},
|
| 44 |
+
"unk_token": {
|
| 45 |
+
"content": "<unk>",
|
| 46 |
+
"lstrip": false,
|
| 47 |
+
"normalized": true,
|
| 48 |
+
"rstrip": false,
|
| 49 |
+
"single_word": false
|
| 50 |
+
}
|
| 51 |
+
}
|
final_checkpoint/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
final_checkpoint/tokenizer_config.json
ADDED
|
@@ -0,0 +1,75 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"add_prefix_space": true,
|
| 3 |
+
"added_tokens_decoder": {
|
| 4 |
+
"0": {
|
| 5 |
+
"content": "<s>",
|
| 6 |
+
"lstrip": false,
|
| 7 |
+
"normalized": true,
|
| 8 |
+
"rstrip": false,
|
| 9 |
+
"single_word": false,
|
| 10 |
+
"special": true
|
| 11 |
+
},
|
| 12 |
+
"1": {
|
| 13 |
+
"content": "<pad>",
|
| 14 |
+
"lstrip": false,
|
| 15 |
+
"normalized": true,
|
| 16 |
+
"rstrip": false,
|
| 17 |
+
"single_word": false,
|
| 18 |
+
"special": true
|
| 19 |
+
},
|
| 20 |
+
"2": {
|
| 21 |
+
"content": "</s>",
|
| 22 |
+
"lstrip": false,
|
| 23 |
+
"normalized": true,
|
| 24 |
+
"rstrip": false,
|
| 25 |
+
"single_word": false,
|
| 26 |
+
"special": true
|
| 27 |
+
},
|
| 28 |
+
"3": {
|
| 29 |
+
"content": "<unk>",
|
| 30 |
+
"lstrip": false,
|
| 31 |
+
"normalized": true,
|
| 32 |
+
"rstrip": false,
|
| 33 |
+
"single_word": false,
|
| 34 |
+
"special": true
|
| 35 |
+
},
|
| 36 |
+
"50264": {
|
| 37 |
+
"content": "<mask>",
|
| 38 |
+
"lstrip": true,
|
| 39 |
+
"normalized": true,
|
| 40 |
+
"rstrip": false,
|
| 41 |
+
"single_word": false,
|
| 42 |
+
"special": true
|
| 43 |
+
},
|
| 44 |
+
"50265": {
|
| 45 |
+
"content": "<start>",
|
| 46 |
+
"lstrip": false,
|
| 47 |
+
"normalized": false,
|
| 48 |
+
"rstrip": false,
|
| 49 |
+
"single_word": false,
|
| 50 |
+
"special": true
|
| 51 |
+
},
|
| 52 |
+
"50266": {
|
| 53 |
+
"content": "<end>",
|
| 54 |
+
"lstrip": false,
|
| 55 |
+
"normalized": false,
|
| 56 |
+
"rstrip": false,
|
| 57 |
+
"single_word": false,
|
| 58 |
+
"special": true
|
| 59 |
+
}
|
| 60 |
+
},
|
| 61 |
+
"bos_token": "<s>",
|
| 62 |
+
"clean_up_tokenization_spaces": true,
|
| 63 |
+
"cls_token": "<s>",
|
| 64 |
+
"entity_max_length": 8,
|
| 65 |
+
"eos_token": "</s>",
|
| 66 |
+
"errors": "replace",
|
| 67 |
+
"marker_max_length": 128,
|
| 68 |
+
"mask_token": "<mask>",
|
| 69 |
+
"model_max_length": 256,
|
| 70 |
+
"pad_token": "<pad>",
|
| 71 |
+
"sep_token": "</s>",
|
| 72 |
+
"tokenizer_class": "RobertaTokenizer",
|
| 73 |
+
"trim_offsets": true,
|
| 74 |
+
"unk_token": "<unk>"
|
| 75 |
+
}
|
final_checkpoint/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cdc21b4c9d1f74ac0be95d21943b06d50dbdfc012b15ba465d456f1e90484c21
|
| 3 |
+
size 4600
|
final_checkpoint/vocab.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 499040084
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b1f2a83a5a877bf7c79b2dd581c8ee4ded91ae0fa501cadf1d3f12deb1be8fde
|
| 3 |
size 499040084
|
runs/Nov27_07-50-08_trinity/events.out.tfevents.1701071419.trinity.224163.0
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dcdd56364d742c98d9ed208bcfd60f99eb58c37a29078af4258221972021cc34
|
| 3 |
+
size 58090
|
runs/Nov27_07-50-08_trinity/events.out.tfevents.1701076911.trinity.224163.1
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7aa30977e15166f604b04b4ef242ee6471c33e917341e69b9ce509dd8db09e3f
|
| 3 |
+
size 592
|
test_results.json
ADDED
|
@@ -0,0 +1,407 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"epoch": 3.0,
|
| 3 |
+
"test_art-broadcastprogram": {
|
| 4 |
+
"f1": 0.5963149078726968,
|
| 5 |
+
"number": 603,
|
| 6 |
+
"precision": 0.6023688663282571,
|
| 7 |
+
"recall": 0.5903814262023217
|
| 8 |
+
},
|
| 9 |
+
"test_art-film": {
|
| 10 |
+
"f1": 0.7645466847090664,
|
| 11 |
+
"number": 750,
|
| 12 |
+
"precision": 0.7760989010989011,
|
| 13 |
+
"recall": 0.7533333333333333
|
| 14 |
+
},
|
| 15 |
+
"test_art-music": {
|
| 16 |
+
"f1": 0.7685459940652819,
|
| 17 |
+
"number": 1029,
|
| 18 |
+
"precision": 0.7824773413897281,
|
| 19 |
+
"recall": 0.7551020408163265
|
| 20 |
+
},
|
| 21 |
+
"test_art-other": {
|
| 22 |
+
"f1": 0.37103174603174605,
|
| 23 |
+
"number": 562,
|
| 24 |
+
"precision": 0.4192825112107623,
|
| 25 |
+
"recall": 0.33274021352313166
|
| 26 |
+
},
|
| 27 |
+
"test_art-painting": {
|
| 28 |
+
"f1": 0.5555555555555555,
|
| 29 |
+
"number": 57,
|
| 30 |
+
"precision": 0.5882352941176471,
|
| 31 |
+
"recall": 0.5263157894736842
|
| 32 |
+
},
|
| 33 |
+
"test_art-writtenart": {
|
| 34 |
+
"f1": 0.6649020645844362,
|
| 35 |
+
"number": 968,
|
| 36 |
+
"precision": 0.6818675352877307,
|
| 37 |
+
"recall": 0.6487603305785123
|
| 38 |
+
},
|
| 39 |
+
"test_building-airport": {
|
| 40 |
+
"f1": 0.8205128205128206,
|
| 41 |
+
"number": 364,
|
| 42 |
+
"precision": 0.8063660477453581,
|
| 43 |
+
"recall": 0.8351648351648352
|
| 44 |
+
},
|
| 45 |
+
"test_building-hospital": {
|
| 46 |
+
"f1": 0.7633986928104576,
|
| 47 |
+
"number": 364,
|
| 48 |
+
"precision": 0.7281795511221946,
|
| 49 |
+
"recall": 0.8021978021978022
|
| 50 |
+
},
|
| 51 |
+
"test_building-hotel": {
|
| 52 |
+
"f1": 0.7137546468401488,
|
| 53 |
+
"number": 265,
|
| 54 |
+
"precision": 0.7032967032967034,
|
| 55 |
+
"recall": 0.7245283018867924
|
| 56 |
+
},
|
| 57 |
+
"test_building-library": {
|
| 58 |
+
"f1": 0.7464387464387464,
|
| 59 |
+
"number": 355,
|
| 60 |
+
"precision": 0.7550432276657061,
|
| 61 |
+
"recall": 0.7380281690140845
|
| 62 |
+
},
|
| 63 |
+
"test_building-other": {
|
| 64 |
+
"f1": 0.5853370122191565,
|
| 65 |
+
"number": 2543,
|
| 66 |
+
"precision": 0.5867246147767681,
|
| 67 |
+
"recall": 0.5839559575304758
|
| 68 |
+
},
|
| 69 |
+
"test_building-restaurant": {
|
| 70 |
+
"f1": 0.5667447306791569,
|
| 71 |
+
"number": 232,
|
| 72 |
+
"precision": 0.6205128205128205,
|
| 73 |
+
"recall": 0.521551724137931
|
| 74 |
+
},
|
| 75 |
+
"test_building-sportsfacility": {
|
| 76 |
+
"f1": 0.6921487603305786,
|
| 77 |
+
"number": 420,
|
| 78 |
+
"precision": 0.6113138686131386,
|
| 79 |
+
"recall": 0.7976190476190477
|
| 80 |
+
},
|
| 81 |
+
"test_building-theater": {
|
| 82 |
+
"f1": 0.7270788912579957,
|
| 83 |
+
"number": 455,
|
| 84 |
+
"precision": 0.7060041407867494,
|
| 85 |
+
"recall": 0.7494505494505495
|
| 86 |
+
},
|
| 87 |
+
"test_event-attack/battle/war/militaryconflict": {
|
| 88 |
+
"f1": 0.7660377358490565,
|
| 89 |
+
"number": 1098,
|
| 90 |
+
"precision": 0.7945205479452054,
|
| 91 |
+
"recall": 0.7395264116575592
|
| 92 |
+
},
|
| 93 |
+
"test_event-disaster": {
|
| 94 |
+
"f1": 0.5603864734299517,
|
| 95 |
+
"number": 207,
|
| 96 |
+
"precision": 0.5603864734299517,
|
| 97 |
+
"recall": 0.5603864734299517
|
| 98 |
+
},
|
| 99 |
+
"test_event-election": {
|
| 100 |
+
"f1": 0.22040816326530616,
|
| 101 |
+
"number": 182,
|
| 102 |
+
"precision": 0.42857142857142855,
|
| 103 |
+
"recall": 0.14835164835164835
|
| 104 |
+
},
|
| 105 |
+
"test_event-other": {
|
| 106 |
+
"f1": 0.4629404617253949,
|
| 107 |
+
"number": 866,
|
| 108 |
+
"precision": 0.48846153846153845,
|
| 109 |
+
"recall": 0.4399538106235566
|
| 110 |
+
},
|
| 111 |
+
"test_event-protest": {
|
| 112 |
+
"f1": 0.42245989304812837,
|
| 113 |
+
"number": 166,
|
| 114 |
+
"precision": 0.3798076923076923,
|
| 115 |
+
"recall": 0.4759036144578313
|
| 116 |
+
},
|
| 117 |
+
"test_event-sportsevent": {
|
| 118 |
+
"f1": 0.6179955171309639,
|
| 119 |
+
"number": 1566,
|
| 120 |
+
"precision": 0.619781631342325,
|
| 121 |
+
"recall": 0.6162196679438059
|
| 122 |
+
},
|
| 123 |
+
"test_location-GPE": {
|
| 124 |
+
"f1": 0.8349881570447639,
|
| 125 |
+
"number": 20405,
|
| 126 |
+
"precision": 0.8157255048616305,
|
| 127 |
+
"recall": 0.8551825532957609
|
| 128 |
+
},
|
| 129 |
+
"test_location-bodiesofwater": {
|
| 130 |
+
"f1": 0.7472984206151289,
|
| 131 |
+
"number": 1169,
|
| 132 |
+
"precision": 0.7267582861762328,
|
| 133 |
+
"recall": 0.7690333618477331
|
| 134 |
+
},
|
| 135 |
+
"test_location-island": {
|
| 136 |
+
"f1": 0.7157894736842105,
|
| 137 |
+
"number": 646,
|
| 138 |
+
"precision": 0.7504244482173175,
|
| 139 |
+
"recall": 0.6842105263157895
|
| 140 |
+
},
|
| 141 |
+
"test_location-mountain": {
|
| 142 |
+
"f1": 0.7324981577008107,
|
| 143 |
+
"number": 681,
|
| 144 |
+
"precision": 0.735207100591716,
|
| 145 |
+
"recall": 0.7298091042584435
|
| 146 |
+
},
|
| 147 |
+
"test_location-other": {
|
| 148 |
+
"f1": 0.36490474912798493,
|
| 149 |
+
"number": 2191,
|
| 150 |
+
"precision": 0.4427083333333333,
|
| 151 |
+
"recall": 0.31036056595162026
|
| 152 |
+
},
|
| 153 |
+
"test_location-park": {
|
| 154 |
+
"f1": 0.7001114827201783,
|
| 155 |
+
"number": 458,
|
| 156 |
+
"precision": 0.715261958997722,
|
| 157 |
+
"recall": 0.6855895196506551
|
| 158 |
+
},
|
| 159 |
+
"test_location-road/railway/highway/transit": {
|
| 160 |
+
"f1": 0.7204861111111112,
|
| 161 |
+
"number": 1700,
|
| 162 |
+
"precision": 0.708997722095672,
|
| 163 |
+
"recall": 0.7323529411764705
|
| 164 |
+
},
|
| 165 |
+
"test_loss": 0.022335968911647797,
|
| 166 |
+
"test_organization-company": {
|
| 167 |
+
"f1": 0.7011596788581624,
|
| 168 |
+
"number": 3896,
|
| 169 |
+
"precision": 0.6962794229309036,
|
| 170 |
+
"recall": 0.7061088295687885
|
| 171 |
+
},
|
| 172 |
+
"test_organization-education": {
|
| 173 |
+
"f1": 0.7990314769975787,
|
| 174 |
+
"number": 2066,
|
| 175 |
+
"precision": 0.7994186046511628,
|
| 176 |
+
"recall": 0.7986447241045499
|
| 177 |
+
},
|
| 178 |
+
"test_organization-government/governmentagency": {
|
| 179 |
+
"f1": 0.49800072700836057,
|
| 180 |
+
"number": 1511,
|
| 181 |
+
"precision": 0.5524193548387096,
|
| 182 |
+
"recall": 0.45334215751158174
|
| 183 |
+
},
|
| 184 |
+
"test_organization-media/newspaper": {
|
| 185 |
+
"f1": 0.6583701324769169,
|
| 186 |
+
"number": 1232,
|
| 187 |
+
"precision": 0.6513105639396346,
|
| 188 |
+
"recall": 0.6655844155844156
|
| 189 |
+
},
|
| 190 |
+
"test_organization-other": {
|
| 191 |
+
"f1": 0.5660735468564649,
|
| 192 |
+
"number": 4439,
|
| 193 |
+
"precision": 0.59784515159108,
|
| 194 |
+
"recall": 0.5375084478486145
|
| 195 |
+
},
|
| 196 |
+
"test_organization-politicalparty": {
|
| 197 |
+
"f1": 0.704431247144815,
|
| 198 |
+
"number": 1054,
|
| 199 |
+
"precision": 0.6792951541850221,
|
| 200 |
+
"recall": 0.7314990512333965
|
| 201 |
+
},
|
| 202 |
+
"test_organization-religion": {
|
| 203 |
+
"f1": 0.583982990786676,
|
| 204 |
+
"number": 672,
|
| 205 |
+
"precision": 0.557510148849797,
|
| 206 |
+
"recall": 0.6130952380952381
|
| 207 |
+
},
|
| 208 |
+
"test_organization-showorganization": {
|
| 209 |
+
"f1": 0.5935228023793787,
|
| 210 |
+
"number": 769,
|
| 211 |
+
"precision": 0.603494623655914,
|
| 212 |
+
"recall": 0.5838751625487646
|
| 213 |
+
},
|
| 214 |
+
"test_organization-sportsleague": {
|
| 215 |
+
"f1": 0.6499442586399109,
|
| 216 |
+
"number": 882,
|
| 217 |
+
"precision": 0.6392543859649122,
|
| 218 |
+
"recall": 0.6609977324263039
|
| 219 |
+
},
|
| 220 |
+
"test_organization-sportsteam": {
|
| 221 |
+
"f1": 0.7518034704620783,
|
| 222 |
+
"number": 2473,
|
| 223 |
+
"precision": 0.7259036144578314,
|
| 224 |
+
"recall": 0.779619894864537
|
| 225 |
+
},
|
| 226 |
+
"test_other-astronomything": {
|
| 227 |
+
"f1": 0.7906976744186047,
|
| 228 |
+
"number": 678,
|
| 229 |
+
"precision": 0.7793696275071633,
|
| 230 |
+
"recall": 0.8023598820058997
|
| 231 |
+
},
|
| 232 |
+
"test_other-award": {
|
| 233 |
+
"f1": 0.6903954802259886,
|
| 234 |
+
"number": 919,
|
| 235 |
+
"precision": 0.717978848413631,
|
| 236 |
+
"recall": 0.6648531011969532
|
| 237 |
+
},
|
| 238 |
+
"test_other-biologything": {
|
| 239 |
+
"f1": 0.6536203522504893,
|
| 240 |
+
"number": 1874,
|
| 241 |
+
"precision": 0.6864357017028773,
|
| 242 |
+
"recall": 0.6237993596584845
|
| 243 |
+
},
|
| 244 |
+
"test_other-chemicalthing": {
|
| 245 |
+
"f1": 0.5856459330143541,
|
| 246 |
+
"number": 1014,
|
| 247 |
+
"precision": 0.5687732342007435,
|
| 248 |
+
"recall": 0.6035502958579881
|
| 249 |
+
},
|
| 250 |
+
"test_other-currency": {
|
| 251 |
+
"f1": 0.7643384440658716,
|
| 252 |
+
"number": 799,
|
| 253 |
+
"precision": 0.6995841995841996,
|
| 254 |
+
"recall": 0.8423028785982478
|
| 255 |
+
},
|
| 256 |
+
"test_other-disease": {
|
| 257 |
+
"f1": 0.6976744186046512,
|
| 258 |
+
"number": 749,
|
| 259 |
+
"precision": 0.6591448931116389,
|
| 260 |
+
"recall": 0.7409879839786382
|
| 261 |
+
},
|
| 262 |
+
"test_other-educationaldegree": {
|
| 263 |
+
"f1": 0.615595075239398,
|
| 264 |
+
"number": 363,
|
| 265 |
+
"precision": 0.6114130434782609,
|
| 266 |
+
"recall": 0.6198347107438017
|
| 267 |
+
},
|
| 268 |
+
"test_other-god": {
|
| 269 |
+
"f1": 0.6816143497757848,
|
| 270 |
+
"number": 635,
|
| 271 |
+
"precision": 0.6486486486486487,
|
| 272 |
+
"recall": 0.7181102362204724
|
| 273 |
+
},
|
| 274 |
+
"test_other-language": {
|
| 275 |
+
"f1": 0.7300291545189505,
|
| 276 |
+
"number": 753,
|
| 277 |
+
"precision": 0.6507276507276507,
|
| 278 |
+
"recall": 0.8313413014608234
|
| 279 |
+
},
|
| 280 |
+
"test_other-law": {
|
| 281 |
+
"f1": 0.7126673532440783,
|
| 282 |
+
"number": 472,
|
| 283 |
+
"precision": 0.6933867735470942,
|
| 284 |
+
"recall": 0.7330508474576272
|
| 285 |
+
},
|
| 286 |
+
"test_other-livingthing": {
|
| 287 |
+
"f1": 0.6298342541436465,
|
| 288 |
+
"number": 863,
|
| 289 |
+
"precision": 0.6019007391763463,
|
| 290 |
+
"recall": 0.660486674391657
|
| 291 |
+
},
|
| 292 |
+
"test_other-medical": {
|
| 293 |
+
"f1": 0.5168539325842697,
|
| 294 |
+
"number": 397,
|
| 295 |
+
"precision": 0.5123762376237624,
|
| 296 |
+
"recall": 0.5214105793450882
|
| 297 |
+
},
|
| 298 |
+
"test_overall_accuracy": 0.9256893595441806,
|
| 299 |
+
"test_overall_f1": 0.703084859534267,
|
| 300 |
+
"test_overall_precision": 0.7034273336857051,
|
| 301 |
+
"test_overall_recall": 0.7027427186979075,
|
| 302 |
+
"test_person-actor": {
|
| 303 |
+
"f1": 0.8214397008413836,
|
| 304 |
+
"number": 1637,
|
| 305 |
+
"precision": 0.8384223918575063,
|
| 306 |
+
"recall": 0.8051313378130727
|
| 307 |
+
},
|
| 308 |
+
"test_person-artist/author": {
|
| 309 |
+
"f1": 0.7320701754385964,
|
| 310 |
+
"number": 3463,
|
| 311 |
+
"precision": 0.7121791370835608,
|
| 312 |
+
"recall": 0.7531042448743863
|
| 313 |
+
},
|
| 314 |
+
"test_person-athlete": {
|
| 315 |
+
"f1": 0.8370089593383873,
|
| 316 |
+
"number": 2884,
|
| 317 |
+
"precision": 0.8318493150684931,
|
| 318 |
+
"recall": 0.8422330097087378
|
| 319 |
+
},
|
| 320 |
+
"test_person-director": {
|
| 321 |
+
"f1": 0.7221238938053098,
|
| 322 |
+
"number": 554,
|
| 323 |
+
"precision": 0.7083333333333334,
|
| 324 |
+
"recall": 0.7364620938628159
|
| 325 |
+
},
|
| 326 |
+
"test_person-other": {
|
| 327 |
+
"f1": 0.6784606547960942,
|
| 328 |
+
"number": 8767,
|
| 329 |
+
"precision": 0.6833275483049867,
|
| 330 |
+
"recall": 0.6736625983802897
|
| 331 |
+
},
|
| 332 |
+
"test_person-politician": {
|
| 333 |
+
"f1": 0.6821515892420537,
|
| 334 |
+
"number": 2857,
|
| 335 |
+
"precision": 0.6807249912861624,
|
| 336 |
+
"recall": 0.6835841792089604
|
| 337 |
+
},
|
| 338 |
+
"test_person-scholar": {
|
| 339 |
+
"f1": 0.5301369863013699,
|
| 340 |
+
"number": 743,
|
| 341 |
+
"precision": 0.5397489539748954,
|
| 342 |
+
"recall": 0.5208613728129206
|
| 343 |
+
},
|
| 344 |
+
"test_person-soldier": {
|
| 345 |
+
"f1": 0.5451957295373664,
|
| 346 |
+
"number": 647,
|
| 347 |
+
"precision": 0.5052770448548812,
|
| 348 |
+
"recall": 0.5919629057187017
|
| 349 |
+
},
|
| 350 |
+
"test_product-airplane": {
|
| 351 |
+
"f1": 0.6654111738857501,
|
| 352 |
+
"number": 792,
|
| 353 |
+
"precision": 0.66167290886392,
|
| 354 |
+
"recall": 0.6691919191919192
|
| 355 |
+
},
|
| 356 |
+
"test_product-car": {
|
| 357 |
+
"f1": 0.7221812822402359,
|
| 358 |
+
"number": 687,
|
| 359 |
+
"precision": 0.7313432835820896,
|
| 360 |
+
"recall": 0.7132459970887919
|
| 361 |
+
},
|
| 362 |
+
"test_product-food": {
|
| 363 |
+
"f1": 0.5787037037037037,
|
| 364 |
+
"number": 432,
|
| 365 |
+
"precision": 0.5787037037037037,
|
| 366 |
+
"recall": 0.5787037037037037
|
| 367 |
+
},
|
| 368 |
+
"test_product-game": {
|
| 369 |
+
"f1": 0.7250257466529352,
|
| 370 |
+
"number": 493,
|
| 371 |
+
"precision": 0.7364016736401674,
|
| 372 |
+
"recall": 0.7139959432048681
|
| 373 |
+
},
|
| 374 |
+
"test_product-other": {
|
| 375 |
+
"f1": 0.4794617563739376,
|
| 376 |
+
"number": 1608,
|
| 377 |
+
"precision": 0.5567434210526315,
|
| 378 |
+
"recall": 0.4210199004975124
|
| 379 |
+
},
|
| 380 |
+
"test_product-ship": {
|
| 381 |
+
"f1": 0.6842105263157895,
|
| 382 |
+
"number": 380,
|
| 383 |
+
"precision": 0.6842105263157895,
|
| 384 |
+
"recall": 0.6842105263157895
|
| 385 |
+
},
|
| 386 |
+
"test_product-software": {
|
| 387 |
+
"f1": 0.6570316842690384,
|
| 388 |
+
"number": 889,
|
| 389 |
+
"precision": 0.6494505494505495,
|
| 390 |
+
"recall": 0.6647919010123734
|
| 391 |
+
},
|
| 392 |
+
"test_product-train": {
|
| 393 |
+
"f1": 0.5933014354066984,
|
| 394 |
+
"number": 314,
|
| 395 |
+
"precision": 0.5942492012779552,
|
| 396 |
+
"recall": 0.5923566878980892
|
| 397 |
+
},
|
| 398 |
+
"test_product-weapon": {
|
| 399 |
+
"f1": 0.584426946631671,
|
| 400 |
+
"number": 624,
|
| 401 |
+
"precision": 0.6435452793834296,
|
| 402 |
+
"recall": 0.5352564102564102
|
| 403 |
+
},
|
| 404 |
+
"test_runtime": 242.3784,
|
| 405 |
+
"test_samples_per_second": 189.732,
|
| 406 |
+
"test_steps_per_second": 5.933
|
| 407 |
+
}
|