{"created":"2021-03-01T06:00:59.282178+00:00","id":8658,"links":{},"metadata":{"_buckets":{"deposit":"f734a41c-b452-425e-a399-b05cce6af161"},"_deposit":{"id":"8658","owners":[],"pid":{"revision_id":0,"type":"depid","value":"8658"},"status":"published"},"_oai":{"id":"oai:kitami-it.repo.nii.ac.jp:00008658","sets":["2:6"]},"author_link":["90014"],"item_7_biblio_info_6":{"attribute_name":"書誌情報","attribute_value_mlt":[{"bibliographicIssueDates":{"bibliographicIssueDate":"2018-09","bibliographicIssueDateType":"Issued"}}]},"item_7_date_granted_63":{"attribute_name":"学位授与年月日","attribute_value_mlt":[{"subitem_dategranted":"2018-09-10"}]},"item_7_degree_grantor_61":{"attribute_name":"学位授与機関","attribute_value_mlt":[{"subitem_degreegrantor":[{"subitem_degreegrantor_name":"北見工業大学"}]}]},"item_7_degree_name_60":{"attribute_name":"学位名","attribute_value_mlt":[{"subitem_degreename":"博士(工学)"}]},"item_7_description_4":{"attribute_name":"抄録","attribute_value_mlt":[{"subitem_description":"Language models are an indispensable element of Natural Language Processing (NLP) research. They are used in machine translation, speech recognition, part-of-speech tagging, handwriting recognition, syntactic parsing, information retrieval and others. In short, language models are probability distributions over sequences of words. There are countless numbers of NLP solutions, algorithms and programs applying language models in specific tasks. Unfortunately, often these are not optimized, but rely on default, most commonly used sets of parameters. For example, many of them use numerous objective functions with different variables but without proper weights applied to them. Users usually set these variables themselves, which causes the results not to exceed a certain mediocre level. In case of small number of variables, users can adjust them manually, but optimization of objective functions with massive number of variables, especially multi-objective functions is difficult and time consuming. This was the motivation to propose an application of a Genetic Algorithms (GAs) to optimize the weighting process.\nGAs are subset of Evolutionary Algorithms (EAs), inspired by the process of natural selection known from nature. They use bio-inspired operators such as selection, crossover and mutation to generate solutions for optimization and search problems. This way GAs represent randomized heuristic search strategies simulating natural selection process, where the population is composed of candidate solutions. They are focused on evolving a population from which strong and diverse candidates can emerge via mutation and crossover (mating). There exist different types of GAs, moreover the same type of GA can bring different quality of solutions, depending on multiple variables, which include starting population, number of generations or fitness function. Finding the best starting parameters and type of GA the most appropriate for a given optimization problem is a next challenge. For that reason, I created a library that automatically applies multiple types of GAs in optimization purposes.\nThe library was created in C++ language, with the use of .NET environment. Its main goal is to be used with different secondary programs and applications, without significant interfering in the original structure of the solution. Basic function of library allows the use of several different kinds of GAs like: Simple GA, Uniform Crossover GA, n-point Crossover GA, GA with sexual selection, GA with chromosome aging and so forth. User can freely define starting parameters for GA including: population size, starting population, number of generations, type of mutation and crossover. Advanced functions of the library allow the use of multithreaded processing for running several GAs in the same time. Basic option of multithreading runs the same type of GA with different starting parameters, advanced version allows to exchange information between different threads every set number of generations. In case of large number of variables to compute, it is also possible to separate a mutation and crossover for several threads running at the same time.\nThe most important functionality of the library is its easy adjustability in optimization of different kinds of applications. The library is used to run the original program in every generation of GA with new weights for variables generated from natural selection. Time of program running is closely related with original program processing time. It depends on the type of original solution and the time of processing one generation is similar to one run of the optimized program.\nDuring creating and testing the library, numerous experiments have been carried out. In preliminary experiments the library was used for optimization of construction of mechanical elements. Later the application was tested on natural language processing and related solutions. One part of the research was optimizing Quantitative Learner’s Motivation Model. The goal of this experiment was to optimize the formula for prediction of learning motivation by means of different weights for three values: interest, usefulness in the future and satisfaction. For this optimization, an application in C# using GA library was created. Data sets for the experiments were acquired from questionnaires enquiring about the above three elements in actual university classes. The results of the experiment showed improvement in the estimation of student’s learning motivation up to over 17 percentage points of Fscore.\nThe final experiment aimed to optimize the implementation of Support Vector Ma-chines (SVMs) for the problem of pattern recognition in natural language data. SVMs are a machine learning algorithm based on statistical learning theory. They are applied to large number of real-world applications, such as text categorization, hand-written character recognition, etc. Original program was created in C++. For this application numerous different types of GAs were tested with different number of generations, weight range and starting parameters. Optimization was successful, with different scale of improvement based on previously mentioned conditions, with the highest achieved improvement of over 6 percentage points of recall comparing to baseline and reaching 78%. All experiments data are included in this work.","subitem_description_language":"en","subitem_description_type":"Abstract"}]},"item_7_dissertation_number_64":{"attribute_name":"学位授与番号","attribute_value_mlt":[{"subitem_dissertationnumber":"10106甲第170号"}]},"item_7_identifier_registration":{"attribute_name":"ID登録","attribute_value_mlt":[{"subitem_identifier_reg_text":"10.19000/00008636","subitem_identifier_reg_type":"JaLC"}]},"item_7_select_15":{"attribute_name":"著者版フラグ","attribute_value_mlt":[{"subitem_select_item":"ETD"}]},"item_7_text_66":{"attribute_name":"研究科・専攻名","attribute_value_mlt":[{"subitem_text_value":"生産基盤工学専攻"}]},"item_access_right":{"attribute_name":"アクセス権","attribute_value_mlt":[{"subitem_access_right":"open access","subitem_access_right_uri":"http://purl.org/coar/access_right/c_abf2"}]},"item_creator":{"attribute_name":"著者","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"PAWEL, CEZARY LEMPA","creatorNameLang":"en"}],"nameIdentifiers":[{"nameIdentifier":"90014","nameIdentifierScheme":"WEKO"}]}]},"item_files":{"attribute_name":"ファイル情報","attribute_type":"file","attribute_value_mlt":[{"accessrole":"open_date","date":[{"dateType":"Available","dateValue":"2018-09-13"}],"displaytype":"detail","filename":"Doctoral_Thesis_2018_Pawel_Cezary_Lempa.pdf","filesize":[{"value":"2.3 MB"}],"format":"application/pdf","licensetype":"license_note","mimetype":"application/pdf","url":{"label":"Doctoral_Thesis_2018_Pawel_Cezary_Lempa","objectType":"fulltext","url":"https://kitami-it.repo.nii.ac.jp/record/8658/files/Doctoral_Thesis_2018_Pawel_Cezary_Lempa.pdf"},"version_id":"5546f14b-fee1-4dbb-81a8-8f2eb974654d"}]},"item_language":{"attribute_name":"言語","attribute_value_mlt":[{"subitem_language":"eng"}]},"item_resource_type":{"attribute_name":"資源タイプ","attribute_value_mlt":[{"resourcetype":"doctoral thesis","resourceuri":"http://purl.org/coar/resource_type/c_db06"}]},"item_title":"Development of Optimization Method with the Use of Genetic Algorithms for Natural Language and Related Models","item_titles":{"attribute_name":"タイトル","attribute_value_mlt":[{"subitem_title":"Development of Optimization Method with the Use of Genetic Algorithms for Natural Language and Related Models","subitem_title_language":"en"},{"subitem_title":"遺伝的アルゴリズムを用いた自然言語とその関連モデルの最適化手法の開発","subitem_title_language":"ja"}]},"item_type_id":"7","owner":"1","path":["6"],"pubdate":{"attribute_name":"PubDate","attribute_value":"2018-09-13"},"publish_date":"2018-09-13","publish_status":"0","recid":"8658","relation_version_is_last":true,"title":["Development of Optimization Method with the Use of Genetic Algorithms for Natural Language and Related Models"],"weko_creator_id":"1","weko_shared_id":-1},"updated":"2022-12-13T02:22:12.388890+00:00"}