Endre søk
RefereraExporteraLink to record
Permanent link

Direct link
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annet format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annet språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf
Utvärdering av olika språkmodeller för identifiering av information i svenska texter
Jönköping University, Tekniska Högskolan, JTH, Avdelningen för datateknik och informatik.
Jönköping University, Tekniska Högskolan, JTH, Avdelningen för datateknik och informatik.
2021 (svensk)Independent thesis Basic level (degree of Bachelor), 180 hpOppgaveAlternativ tittel
Evaluation of different language models for identifying entities in Swedish texts (engelsk)
Abstract [en]

The purpose of the study was to investigate how different types of Natural Language Processing (NLP) models can be fine-tuned for use within Named Entity Recognition (NER) and how these models work differently. The aim of this was to identify which type of model works best when it comes to entity extraction. To test the models, they were fine-tuned using PyTorch and Huggingface Transformers libraries. Here, it was also investigated which other techniques are not based on machine learning that can be used to solve the problem of entity extraction. The results of these tests showed that the KB/BERT-base-swedish-cased model worked best. Also, there are other techniques that can be used to solve the problem of entity extraction but that it is machine learning that is the most effective. KB/BERT-base-swedish-cased was shown to be the best model but it also proved that it may be due to the purpose that you need to use the model within. The study limitations have been that there was no large data set that could be used for training and that there was a lack of time to produce a new dataset. 

sted, utgiver, år, opplag, sider
2021. , s. 45
Emneord [sv]
Maskininlärning, Named Entity Recognition, Natural Language Processing, Transformers, BERT, ALBERT, ELECTRA, HFST-SweNER
HSV kategori
Identifikatorer
URN: urn:nbn:se:hj:diva-55101ISRN: JU-JTH-DTA-1-20210159OAI: oai:DiVA.org:hj-55101DiVA, id: diva2:1612215
Eksternt samarbeid
Cybercom Group AB
Fag / kurs
JTH, Computer Engineering
Veileder
Examiner
Tilgjengelig fra: 2021-11-18 Laget: 2021-11-17 Sist oppdatert: 2021-11-18bibliografisk kontrollert

Open Access i DiVA

Fulltekst mangler i DiVA

Av organisasjonen

Søk utenfor DiVA

GoogleGoogle Scholar

urn-nbn

Altmetric

urn-nbn
Totalt: 479 treff
RefereraExporteraLink to record
Permanent link

Direct link
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annet format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annet språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf