File size: 2,982 Bytes
a15af03
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
---
datasets:
- samirmsallem/text_parts_de
language:
- de
base_model:
- deepset/gbert-large
pipeline_tag: text-classification
library_name: transformers
tags:
- science
- german
- text
metrics:
- accuracy
model-index:
- name: checkpoints
  results:
  - task:
      name: Text Classification
      type: text-classification
    dataset:
      name: samirmsallem/text_parts_de
      type: samirmsallem/text_parts_de
    metrics:
    - name: Accuracy
      type: accuracy
      value: 0.7850140056022409
---

## Sequence classification model for text parts classification in German scientific texts

**gbert-large-text_parts** is a sequence classification model in the scientific domain in German, finetuned from the model [gbert-large](https://huggingface.co/deepset/gbert-large). 
It was trained using a custom annotated dataset of around 4,800 training and 1,400 test examples containing introduction, main part and conclusion text sequences from scientific papers and theses in german.

The two German BERT models — gbert-base and gbert-large — were evaluated on the same text classification task. This repository contains the model and results for [gbert-large](https://huggingface.co/deepset/gbert-large).

- gbert-base reached a maximum evaluation accuracy of 70.31%, with a minimum evaluation loss of 2.13.
- gbert-large achieved a higher peak accuracy of 78.50% and a lower evaluation loss of 1.67.

Despite its superior peak performance, gbert-large exhibited a consistent degradation in evaluation accuracy and an increase in loss over subsequent epochs. 
This pattern is characteristic of overfitting, where the model fits the training data too closely and fails to generalize well to unseen data.
In contrast, gbert-base maintained more stable performance over time, making it a more robust choice in scenarios where computational resources or dataset size are limited.

|Text Classification Tag| Text Classification Label | Description                             |
| :----:                |    :----:                 |    :----:   |
| 0                     | EINLEITUNG          | The text is an introduction (Einleitung). |
| 1                     | HAUPTTEIL              | The text is a main part (Hauptteil). |
| 2                     | SCHLUSS              | The text is a conclusion (Schluss). |


### Training
Training was conducted on a 7 epoch fine-tuning approach:

| Epoch | Loss    | Accuracy |
|-------|---------|----------|
| 1.0   | **1.6681**  | **0.7850** |
| 2.0   | 2.0727  | 0.7780   |
| 3.0   | 2.2932  | 0.7703   |
| 4.0   | 2.4115  | 0.7682   |
| 5.0   | 3.8431  | 0.6611   |
| 6.0   | 2.9657  | 0.7129   |
| 7.0   | 4.3850  | 0.6492   |





Training was conducted using a standard Text classification objective. The model achieves an accuracy of approximately 78,5% on the evaluation set.

Here are the overall final metrics on the test dataset after 1 epoch of training:
  - **Accuracy**: 0.7850140056022409
  - **Loss**: 1.668112874031067