GitXplorerGitXplorer
s

germeval-ner-t5

public
3 stars
0 forks
0 issues

Commits

List of commits on branch main.
Unverified
8b601430f6776917d51764e1c8a5dbcc0e4a51dd

configs: rename base to small (according to model name)

sstefan-it committed 2 years ago
Verified
fac2d2e84ad3c4531d4d01295309a5125ba23495

logs: fix parsing script

sstefan-it committed 2 years ago
Verified
88fa1fe3a212a0f8fc50bc545b59f970011eacea

logs: add initial version of flair log parser

sstefan-it committed 2 years ago
Unverified
a9e37412b00eda82bd62c4fecd995ddf9aabdfe6

readme: clarify test set result

sstefan-it committed 2 years ago
Unverified
7ccd0e222f4d51bb4307a8cdd3f55d28f1e749e4

readme: add more useful links

sstefan-it committed 2 years ago
Unverified
3ae62006f7bd7d92b8ab1692035b8a8e8dcd020b

readme: add baseline and hardware/requirement sections

sstefan-it committed 2 years ago

README

The README file for this repository.

Evaluating German T5 Models on GermEval 2014 (NER)

This repository presents an on-going evaluation of German T5 models on the GermEval 2014 NER downstream task.

Changelog

  • 03.02.2023: Initial version.

Fine-Tuning T5

A few approaches exist for fine-tuning T5 models for token classification tasks:

These approaches tackle the token classification task as a sequence-to-sequence task.

However, it is also possible to use obly the encoder of a T5 model for downstream tasks as presented in:

The proposed "EncT5" architecture was not evaluated on token classification tasks.

This repository uses the Flair library and encoder-only fine-tuning is performed for the GermEval 2014 NER dataset. The recently released T5 models for German are used as LM backbones.

Results

We perform a basic hyper-parameter search over and report micro F1-Score, averaged over 5 runs (with different seeds). Score in brackets indicates result on development split.

Model Size Configuration Run 1 Run 2 Run 3 Run 4 Run 5 Avg.
Small bs16-e10-lr0.00011 (87.24) / 85.53 (86.40) / 85.63 (86.50) / 85.47 (86.32) / 85.57 (86.77) / 85.38 (86.65) / 85.52
Large bs16-e10-lr0.00011 (87.16) / 86.46 (87.07) / 85.76 (87.46) / 85.57 (87.05) / 86.91 (87.15) / 86.11 (87.18) / 86.16

For hyper-parameter search, the script flair-fine-tuner.py is used in combination with a configuration file (passed as argument). All configuration files are located under ./configs that were used for the experiments here.

Baselines:

Hardware/Requirements

Latest Flair version (commit 6da65a4) is used for experiments.

All models are fine-tuned on A10 (24GB) instances from Lambda Cloud.