Information Technology for Foreign Languages Remote Learning with Adaptation to the User Based on Machine Learning

DC ElementWertSprache
dc.contributor.authorSopin, Taras
dc.contributor.authorVysotska, Victoria
dc.contributor.authorMarkiv, Oksana
dc.contributor.authorChyrun, Lyubomyr
dc.contributor.authorAndrunyk, Vasyl
dc.contributor.authorChyrun, Sofia
dc.contributor.authorNaum, Oleh
dc.contributor.editorEmmerich, M.
dc.contributor.editorLeiden University
dc.contributor.editorLeiden Institute of Advanced Computer Science
dc.contributor.editorNiels Bohrweg 1
dc.contributor.editorLeiden
dc.contributor.editorVysotska, V.
dc.contributor.editorOsnabruck University
dc.contributor.editorFriedrich-Janssen-Str. 1
dc.contributor.editorOsnabruck
dc.contributor.editorLytvynenko, V.
dc.contributor.editorKherson National Technical University
dc.contributor.editorBeryslavske Shosse
dc.contributor.editor24
dc.contributor.editorKherson
dc.date.accessioned2024-01-04T10:29:05Z-
dc.date.available2024-01-04T10:29:05Z-
dc.date.issued2023
dc.identifier.issn1613-0073
dc.identifier.urihttp://osnascholar.ub.uni-osnabrueck.de/handle/unios/72969-
dc.descriptionCited by: 0; Conference name: 2023 Modern Machine Learning Technologies and Data Science Workshop, MoMLeT and DS 2023; Conference date: 3 June 2023; Conference code: 189914
dc.description.abstractSince the goal of the work is to improve the process of remote learning of foreign languages, it was chosen to create an application for translation based on the choice of the native language and the language being studied. An educational process using a set of telecommunication technologies aimed at enabling students to learn the basic amount of information they need without direct contact between students and teachers during the learning process (which can take place both synchronously and asynchronously), and can be both an independent form of education, as well as a supplement to another more traditional form of education (full-time, part-time, extramural or externship), if necessary, giving a person the opportunity to study a foreign language training course. So, on the basis of this concept, a translation application was developed, which accurately translates both ordinary language and phraseological units, slang expressions, etc. The model is used as the basis of training, so let's analyze the model according to the main indicators. The model was pre-trained on BookCorpus, a dataset consisting of 11,038 unpublished books and the English Wikipedia (excluding lists, tables and titles). The texts are written in lowercase and tokenized using WordPiece and a dictionary size of 30,000. With probability 0.5, sentence A and sentence B match two consecutive sentences in the original corpus, and in other cases it is another random sentence in the corpus. Note that a sentence here is a continuous stretch of text, usually longer than one sentence. The only limitation is that the result with two "sentences" has a total length of less than 512 tokens. The masking procedure details for each sentence are as follows: 15% of tokens are masked; in 80% of cases masked tokens are replaced by [MASK]; 10% of the time, masked tokens are replaced by a random token from the one they replace; in the remaining 10% of cases, masked markers remain unchanged. The model was trained on 4 Cloud TPUs in a Pod configuration (16 TPUs in total) for one million steps with a batch size of 256. The sequence length was limited to 128 markers for 90% of the steps and 512 for the remaining 10%. Adam optimizer is used with learning rate: β1=0.9, and β2=0.999, weight decay 0.01, learning rate warm-up for 10,000 steps and learning rate linear decrease after. After training the network, the mean squared error decreased from 34.2 to 3.3. Also, training the network made it possible to reduce overtraining and improve its ability to generalize to new data. In the trained network, the number of layers and neurons was increased, which allowed it to reproduce more complex dependencies in the input data. Training the network made it possible to improve its results on test data, increase its ability to generalize, optimize its structure and parameters, choose a more effective activation function, and reduce the risk of overtraining. © 2023 Copyright for this paper by its authors.
dc.language.isoen
dc.publisherCEUR-WS
dc.relation.ispartofCEUR Workshop Proceedings
dc.subjectE-learning
dc.subjectEducation computing
dc.subjectForeign language
dc.subjectForeign language training
dc.subjectLearning rates
dc.subjectLearning systems
dc.subjectmachine learning
dc.subjectMachine-learning
dc.subjectNetwork training
dc.subjectNeural networks
dc.subjectneuron network
dc.subjectneuron network training
dc.subjectNeuron networks
dc.subjectNeurons
dc.subjectOn-machines
dc.subjectRemote learning
dc.subjectStudents
dc.subjectTranslation (languages)
dc.titleInformation Technology for Foreign Languages Remote Learning with Adaptation to the User Based on Machine Learning
dc.typeconference paper
dc.identifier.scopus2-s2.0-85164952144
dc.identifier.urlhttps://www.scopus.com/inward/record.uri?eid=2-s2.0-85164952144&partnerID=40&md5=8ffd3b0335c77d7326bf6aef3adee021
dc.description.volume3426
dc.description.startpage526 – 550
dcterms.isPartOf.abbreviationCEUR Workshop Proc.
local.import.remainsaffiliations : Lviv Polytechnic National University, S. Bandera Street, 12, Lviv, 79013, Ukraine; Osnabrück University, Friedrich-Janssen-Str. 1, Osnabrück, 49076, Germany; Ivan Franko National University of Lviv, University Street, 1, Lviv, 79000, Ukraine; Ivan Franko Drohobych State Pedagogical University, I. Franko Street, 24, Drohobych, 82100, Ukraine
local.import.remainspublication_stage : Final
Zur Kurzanzeige

Seitenaufrufe

1
Letzte Woche
0
Letzter Monat
0
geprüft am 04.06.2024

Google ScholarTM

Prüfen