Fine-Tuning Large Language Models for Turkish Flutter Code Generation
| dc.contributor.author | Uluirmak, B.A. | |
| dc.contributor.author | Kurban, R. | |
| dc.date.accessioned | 2026-02-21T00:43:17Z | |
| dc.date.available | 2026-02-21T00:43:17Z | |
| dc.date.issued | 2025 | |
| dc.description.abstract | The rapid advancement of large language models (LLMs) for code generation has largely centered on English programming queries. This paper focuses on a low-resource language scenario, specifically Turkish, in the context of Flutter mobile app development. Two representative LLMs (a 4B-parameter multilingual model and a 3B code-specialized model) on a new Turkish question-and-answer dataset for Flutter/Dart are fine-tuned in this study. Fine-tuning with parameter-efficient techniques yields dramatic improvements in code generation quality: Bilingual Evaluation Understudy (BLEU), Recall-Oriented Understudy for Gisting Evaluation (ROUGE-L), Metric for Evaluation of Translation with Explicit Ordering (METEOR), Bidirectional Encoder Representations from Transformers Score (BERTScore), and CodeBLEU scores show significant increases. The rate of correct solutions increased from ~30–70% (for base models) to 80–90% after fine-tuning. The performance trade-offs between models are analyzed, revealing that the multilingual model slightly outperforms the code-focused model in accuracy after fine-tuning. However, the code-focused model demonstrates faster inference speeds. These results demonstrate that even with very limited non-English training data, customizing LLMs can bridge the gap in code generation, enabling high-quality assistance for Turkish developers comparable to that for English. The dataset was released on GitHub to facilitate further research in multilingual code generation. © 2025, Sakarya University. All rights reserved. | en_US |
| dc.identifier.doi | 10.35377/saucis…1722643 | |
| dc.identifier.issn | 2636-8129 | |
| dc.identifier.scopus | 2-s2.0-105027667311 | |
| dc.identifier.uri | https://doi.org/10.35377/saucis…1722643 | |
| dc.identifier.uri | https://hdl.handle.net/20.500.12573/5785 | |
| dc.language.iso | en | en_US |
| dc.publisher | Sakarya University | en_US |
| dc.relation.ispartof | Sakarya University Journal of Computer and Information Sciences | en_US |
| dc.rights | info:eu-repo/semantics/closedAccess | en_US |
| dc.subject | Code Generation | en_US |
| dc.subject | Fine-Tuning | en_US |
| dc.subject | Flutter | en_US |
| dc.subject | Large Language Models | en_US |
| dc.subject | Low-Resource Languages | en_US |
| dc.title | Fine-Tuning Large Language Models for Turkish Flutter Code Generation | en_US |
| dc.type | Article | en_US |
| dspace.entity.type | Publication | |
| gdc.author.scopusid | 60093425600 | |
| gdc.author.scopusid | 24729361200 | |
| gdc.description.department | Abdullah Gül University | en_US |
| gdc.description.departmenttemp | [Uluirmak] Bugra Alperen, Department of Computer Engineering, Abdullah Gül Üniversitesi, Kayseri, Kayseri, Turkey; [Kurban] Rifat, Department of Computer Engineering, Abdullah Gül Üniversitesi, Kayseri, Kayseri, Turkey | en_US |
| gdc.description.endpage | 650 | en_US |
| gdc.description.issue | 4 | en_US |
| gdc.description.publicationcategory | Makale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanı | en_US |
| gdc.description.scopusquality | N/A | |
| gdc.description.startpage | 637 | en_US |
| gdc.description.volume | 8 | en_US |
| gdc.description.wosquality | N/A | |
| gdc.index.type | Scopus | |
| gdc.opencitations.count | 0 | |
| gdc.plumx.scopuscites | 0 | |
| gdc.scopus.citedcount | 0 | |
| gdc.virtual.author | Kurban, Rifat | |
| relation.isAuthorOfPublication | f55f9796-680f-4dd5-9c98-e43d0ffee812 | |
| relation.isAuthorOfPublication.latestForDiscovery | f55f9796-680f-4dd5-9c98-e43d0ffee812 | |
| relation.isOrgUnitOfPublication | 52f507ab-f278-4a1f-824c-44da2a86bd51 | |
| relation.isOrgUnitOfPublication | ef13a800-4c99-4124-81e0-3e25b33c0c2b | |
| relation.isOrgUnitOfPublication | 665d3039-05f8-4a25-9a3c-b9550bffecef | |
| relation.isOrgUnitOfPublication.latestForDiscovery | 52f507ab-f278-4a1f-824c-44da2a86bd51 |
