22 results on '"Park, Chanjun"'
Search Results
2. Analysis of the Effectiveness of Model, Data, and User-Centric Approaches for Chat Application: A Case Study of BlenderBot 2.0.
- Author
-
Park, Chanjun, Lee, Jungseob, Son, Suhyune, Park, Kinam, Jang, Jungsun, and Lim, Heuiseok
- Subjects
LANGUAGE models ,ONLINE chat ,NATURAL language processing ,CHATBOTS ,HATE speech ,INTERNET searching ,ACCURACY of information - Abstract
BlenderBot 2.0 represents a significant advancement in open-domain chatbots by incorporating real-time information and retaining user information across multiple sessions through an internet search module. Despite its innovations, there are still areas for improvement. This paper examines BlenderBot 2.0's limitations and errors from three perspectives: model, data, and user interaction. From the data perspective, we highlight the challenges associated with the crowdsourcing process, including unclear guidelines for workers, insufficient measures for filtering hate speech, and the lack of a robust process for verifying the accuracy of internet-sourced information. From the user perspective, we identify nine types of limitations and conduct a thorough investigation into their causes. For each perspective, we propose practical methods for improvement and discuss potential directions for future research. Additionally, we extend our analysis to include perspectives in the era of large language models (LLMs), further broadening our understanding of the challenges and opportunities present in current AI technologies. This multifaceted analysis not only sheds light on BlenderBot 2.0's current limitations but also charts a path forward for the development of more sophisticated and reliable open-domain chatbots within the broader context of LLM advancements. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. Towards Harnessing the Most of ChatGPT for Korean Grammatical Error Correction.
- Author
-
Park, Chanjun, Koo, Seonmin, Kim, Gyeongmin, and Lim, Heuiseok
- Subjects
CHATGPT ,KOREAN language ,LANGUAGE models ,GENERATIVE pre-trained transformers - Abstract
In this study, we conduct a pioneering and comprehensive examination of ChatGPT's (GPT-3.5 Turbo) capabilities within the realm of Korean Grammatical Error Correction (K-GEC). Given the Korean language's agglutinative nature and its rich linguistic intricacies, the task of accurately correcting errors while preserving Korean-specific sentiments is notably challenging. Utilizing a systematic categorization of Korean grammatical errors, we delve into a meticulous, case-specific analysis to identify the strengths and limitations of a ChatGPT-based correction system. We also critically assess influential parameters like temperature and specific error criteria, illuminating potential strategies to enhance ChatGPT's efficacy in K-GEC tasks. Our findings offer valuable contributions to the expanding domain of NLP research centered on the Korean language. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
4. Enhancing Machine Translation Quality Estimation via Fine-Grained Error Analysis and Large Language Model.
- Author
-
Jung, Dahyun, Park, Chanjun, Eo, Sugyeong, and Lim, Heuiseok
- Subjects
LANGUAGE models ,MACHINE translating ,NATURAL language processing - Abstract
Fine-grained error span detection is a sub-task within quality estimation that aims to identify and assess the spans and severity of errors present in translated sentences. In prior quality estimation, the focus has predominantly been on evaluating translations at the sentence and word levels. However, such an approach fails to recognize the severity of specific segments within translated sentences. To the best of our knowledge, this is the first study that concentrates on enhancing models for this fine-grained error span detection task in machine translation. This study introduces a framework that sequentially performs sentence-level error detection, word-level error span extraction, and severity assessment. We present a detailed analysis for each of the methodologies we propose, substantiating the effectiveness of our system, focusing on two language pairs: English-to-German and Chinese-to-English. Our results suggest that task granularity enhances performance and that a prompt-based fine-tuning approach can offer optimal performance in the classification tasks. Furthermore, we demonstrate that employing a large language model to edit the fine-tuned model's output constitutes a top strategy for achieving robust quality estimation performance. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
5. Primary care and emergency department utilization patterns: Differences between White and Black low‐acuity patients.
- Author
-
Park, Chanjun Syd, Sams, Malik, Nobay, Flavia, Morgan, Adrienne, Adler, David, and Abar, Beau
- Subjects
STATISTICS ,HOSPITAL emergency services ,HEALTH services accessibility ,MEDICAL triage ,ACADEMIC medical centers ,SELF-evaluation ,MULTIPLE regression analysis ,PRIMARY health care ,SEVERITY of illness index ,COMPARATIVE studies ,T-test (Statistics) ,QUESTIONNAIRES ,DESCRIPTIVE statistics ,CHI-squared test ,SOCIODEMOGRAPHIC factors ,WHITE people ,STATISTICAL sampling ,AFRICAN Americans - Abstract
The article discusses about the rising utilization of emergency rooms (ERs) for nonacute medical care in the United States and its impact on patient safety and quality of care. Topic include demographic characteristics of patients seeking low-acuity care in the ER, highlighting the disparities across races and ethnicities and exploring their contact with primary care providers and barriers to accessing other sources of care.
- Published
- 2023
- Full Text
- View/download PDF
6. A Survey on Evaluation Metrics for Machine Translation.
- Author
-
Lee, Seungjun, Lee, Jungseob, Moon, Hyeonseok, Park, Chanjun, Seo, Jaehyung, Eo, Sugyeong, Koo, Seonmin, and Lim, Heuiseok
- Subjects
MACHINE translating ,DEEP learning - Abstract
The success of Transformer architecture has seen increased interest in machine translation (MT). The translation quality of neural network-based MT transcends that of translations derived using statistical methods. This growth in MT research has entailed the development of accurate automatic evaluation metrics that allow us to track the performance of MT. However, automatically evaluating and comparing MT systems is a challenging task. Several studies have shown that traditional metrics (e.g., BLEU, TER) show poor performance in capturing semantic similarity between MT outputs and human reference translations. To date, to improve performance, various evaluation metrics have been proposed using the Transformer architecture. However, a systematic and comprehensive literature review on these metrics is still missing. Therefore, it is necessary to survey the existing automatic evaluation metrics of MT to enable both established and new researchers to quickly understand the trend of MT evaluation over the past few years. In this survey, we present the trend of automatic evaluation metrics. To better understand the developments in the field, we provide the taxonomy of the automatic evaluation metrics. Then, we explain the key contributions and shortcomings of the metrics. In addition, we select the representative metrics from the taxonomy, and conduct experiments to analyze related problems. Finally, we discuss the limitation of the current automatic metric studies through the experimentation and our suggestions for further research to improve the automatic evaluation metrics. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
7. The ASR Post-Processor Performance Challenges of BackTranScription (BTS): Data-Centric and Model-Centric Approaches.
- Author
-
Park, Chanjun, Seo, Jaehyung, Lee, Seolhwa, Lee, Chanhee, and Lim, Heuiseok
- Subjects
AUTOMATIC speech recognition ,MACHINE translating - Abstract
Training an automatic speech recognition (ASR) post-processor based on sequence-to-sequence (S2S) requires a parallel pair (e.g., speech recognition result and human post-edited sentence) to construct the dataset, which demands a great amount of human labor. BackTransScription (BTS) proposes a data-building method to mitigate the limitations of the existing S2S based ASR post-processors, which can automatically generate vast amounts of training datasets, reducing time and cost in data construction. Despite the emergence of this novel approach, the BTS-based ASR post-processor still has research challenges and is mostly untested in diverse approaches. In this study, we highlight these challenges through detailed experiments by analyzing the data-centric approach (i.e., controlling the amount of data without model alteration) and the model-centric approach (i.e., model modification). In other words, we attempt to point out problems with the current trend of research pursuing a model-centric approach and alert against ignoring the importance of the data. Our experiment results show that the data-centric approach outperformed the model-centric approach by +11.69, +17.64, and +19.02 in the F1-score, BLEU, and GLEU tests. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
8. BERTOEIC: Solving TOEIC Problems Using Simple and Efficient Data Augmentation Techniques with Pretrained Transformer Encoders.
- Author
-
Lee, Jeongwoo, Moon, Hyeonseok, Park, Chanjun, Seo, Jaehyung, Eo, Sugyeong, and Lim, Heuiseok
- Subjects
DATA augmentation ,PROBLEM solving ,DEEP learning ,NATURAL language processing ,READING comprehension ,NATURAL languages - Abstract
Recent studies have attempted to understand natural language and infer answers. Machine reading comprehension is one of the representatives, and several related datasets have been opened. However, there are few official open datasets for the Test of English for International Communication (TOEIC), which is widely used for evaluating people's English proficiency, and research for further advancement is not being actively conducted. We consider that the reason why deep learning research for TOEIC is difficult is due to the data scarcity problem, so we therefore propose two data augmentation methods to improve the model in a low resource environment. Considering the attributes of the semantic and grammar problem type in TOEIC, the proposed methods can augment the data similar to the real TOEIC problem by using POS-tagging and Lemmatizing. In addition, we confirmed the importance of understanding semantics and grammar in TOEIC through experiments on each proposed methodology and experiments according to the amount of data. The proposed methods address the data shortage problem of TOEIC and enable an acceptable human-level performance. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
9. Empirical Analysis of Parallel Corpora and In-Depth Analysis Using LIWC.
- Author
-
Park, Chanjun, Shim, Midan, Eo, Sugyeong, Lee, Seolhwa, Seo, Jaehyung, Moon, Hyeonseok, and Lim, Heuiseok
- Subjects
MACHINE translating ,CORPORA ,TRANSLATING & interpreting ,WORD frequency ,LANGUAGE & languages - Abstract
The machine translation system aims to translate source language into target language. Recent studies on MT systems mainly focus on neural machine translation. One factor that significantly affects the performance of NMT is the availability of high-quality parallel corpora. However, high-quality parallel corpora concerning Korean are relatively scarce compared to those associated with other high-resource languages, such as German or Italian. To address this problem, AI Hub recently released seven types of parallel corpora for Korean. In this study, we conduct an in-depth verification of the quality of corresponding parallel corpora through Linguistic Inquiry and Word Count (LIWC) and several relevant experiments. LIWC is a word-counting software program that can analyze corpora in multiple ways and extract linguistic features as a dictionary base. To the best of our knowledge, this study is the first to use LIWC to analyze parallel corpora in the field of NMT. Our findings suggest the direction of further research toward obtaining the improved quality parallel corpora through our correlation analysis in LIWC and NMT performance. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
10. Return on Advertising Spend Prediction with Task Decomposition-Based LSTM Model.
- Author
-
Moon, Hyeonseok, Lee, Taemin, Seo, Jaehyung, Park, Chanjun, Eo, Sugyeong, Aiyanyo, Imatitikua D., Park, Jeongbae, So, Aram, Ok, Kyoungwha, and Park, Kinam
- Subjects
ADVERTISING spending ,ADVERTISING effectiveness ,MARKETING effectiveness ,ADVERTISING revenue ,FORECASTING - Abstract
Return on advertising spend (ROAS) refers to the ratio of revenue generated by advertising projects to its expense. It is used to assess the effectiveness of advertising marketing. Several simulation-based controlled experiments, such as geo experiments, have been proposed recently. This refers to calculating ROAS by dividing a geographic region into a control group and a treatment group and comparing the ROAS generated in each group. However, the data collected through these experiments can only be used to analyze previously constructed data, making it difficult to use in an inductive process that predicts future profits or costs. Furthermore, to obtain ROAS for each advertising group, data must be collected under a new experimental setting each time, suggesting that there is a limitation in using previously collected data. Considering these, we present a method for predicting ROAS that does not require controlled experiments in data acquisition and validates its effectiveness through comparative experiments. Specifically, we propose a task deposition method that divides the end-to-end prediction task into the two-stage process: occurrence prediction and occurred ROAS regression. Through comparative experiments, we reveal that these approaches can effectively deal with the advertising data, in which the label is mainly set to zero-label. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
11. AI Student: A Machine Reading Comprehension System for the Korean College Scholastic Ability Test.
- Author
-
Kim, Gyeongmin, Lee, Soomin, Park, Chanjun, and Jo, Jaechoon
- Subjects
ACADEMIC ability ,READING comprehension ,ABILITY testing ,QUESTION answering systems ,DATA augmentation ,MACHINERY - Abstract
Machine reading comprehension is a question answering mechanism in which a machine reads, understands, and answers questions from a given text. These reasoning skills can be sufficiently grafted into the Korean College Scholastic Ability Test (CSAT) to bring about new scientific and educational advances. In this paper, we propose a novel Korean CSAT Question and Answering (KCQA) model and effectively utilize four easy data augmentation strategies with round trip translation to augment the insufficient data in the training dataset. To evaluate the effectiveness of KCQA, 30 students appeared for the test under conditions identical to the proposed model. Our qualitative and quantitative analysis along with experimental results revealed that KCQA achieved better performance than humans with a higher F1 score of 3.86. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
12. Emergency Medical Services Communication Barriers and the Deaf American Sign Language User.
- Author
-
Rotoli, Jason M., Hancock, Sarah, Park, Chanjun, Demers-Mcletchie, Susan, Panko, Tiffany L., Halle, Trevor, Wills, Jennifer, Scarpino, Julie, Merrill, Johannah, Cushman, Jeremy, and Jones, Courtney
- Subjects
ONLINE education ,FRUSTRATION ,PROFESSIONS ,HUMAN comfort ,COMMUNICATION barriers ,DEAFNESS ,RESEARCH methodology ,CROSS-sectional method ,STAKEHOLDER analysis ,SIGN language ,LIPREADING ,SURVEYS ,PRE-tests & post-tests ,HUMAN services programs ,EMERGENCY medical services ,COMMUNICATION ,DESCRIPTIVE statistics ,PEOPLE with disabilities ,PATIENT care ,DEAF culture - Abstract
Objective: We sought to identify current Emergency Medical Services (EMS) practitioner comfort levels and communication strategies when caring for the Deaf American Sign Language (ASL) user. Additionally, we created and evaluated the effect of an educational intervention and visual communication tool on EMS practitioner comfort levels and communication. Methods: This was a descriptive study assessing communication barriers at baseline and after the implementation of a novel educational intervention with cross-sectional surveys conducted at three time points (pre-, immediate-post, and three months post-intervention). Descriptive statistics characterized the study sample and we quantified responses from the baseline survey and both post-intervention surveys. Results: There were 148 EMS practitioners who responded to the baseline survey. The majority of participants (74%; 109/148) previously responded to a 9-1-1 call for a Deaf patient and 24% (35/148) reported previous training regarding the Deaf community. The majority felt that important details were lost during communication (83%; 90/109), reported that the Deaf patient appeared frustrated during an encounter (72%; 78/109), and felt that communication limited patient care (67%; 73/109). When interacting with a Deaf person, the most common communication strategies included written text (90%; 98/109), friend/family member (90%; 98/109), lip reading (55%; 60/109), and spoken English (50%; 55/109). Immediately after the training, most participants reported that the educational training expanded their knowledge of Deaf culture (93%; 126/135), communication strategies to use (93%; 125/135), and common pitfalls to avoid (96%; 129/135) when caring for Deaf patients. At 3 months, all participants (100%, 79/79) reported that the educational module was helpful. Some participants (19%, 15/79) also reported using the communication tool with other non-English speaking patients. Conclusions: The majority of EMS practitioners reported difficulty communicating with Deaf ASL users and acknowledged a sense of patient frustration. Nearly all participants felt the educational training was beneficial and clinically relevant; three months later, all participants found it to still be helpful. Additionally, the communication tool may be applicable to other populations that use English as a second language. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
13. Dense-to-Question and Sparse-to-Answer: Hybrid Retriever System for Industrial Frequently Asked Questions.
- Author
-
Seo, Jaehyung, Lee, Taemin, Moon, Hyeonseok, Park, Chanjun, Eo, Sugyeong, Aiyanyo, Imatitikua D., Park, Kinam, So, Aram, Ahn, Sungmin, and Park, Jeongbae
- Subjects
DEEP learning ,HYBRID systems ,INFORMATION retrieval ,BRAND loyalty ,NATURAL language processing ,ARTIFICIAL intelligence - Abstract
The term "Frequently asked questions" (FAQ) refers to a query that is asked repeatedly and produces a manually constructed response. It is one of the most important factors influencing customer repurchase and brand loyalty; thus, most industry domains invest heavily in it. This has led to deep-learning-based retrieval models being studied. However, training a model and creating a database specializing in each industry domain comes at a high cost, especially when using a chatbot-based conversation system, as a large amount of resources must be continuously input for the FAQ system's maintenance. It is also difficult for small- and medium-sized companies and national institutions to build individualized training data and databases and obtain satisfactory results. As a result, based on the deep learning information retrieval module, we propose a method of returning responses to customer inquiries using only data that can be easily obtained from companies. We hybridize dense embedding and sparse embedding in this work to make it more robust in professional terms, and we propose new functions to adjust the weight ratio and scale the results returned by the two modules. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
14. Gender Affirming Surgery: A Comprehensive, Systematic Review of All Peer-reviewed Literature and Methods of Assessing Patient-centered Outcomes (Part 1: Breast/Chest, Face, and Voice).
- Author
-
Oles, Norah, Darrach, Halley, Landford, Wilmina, Garza, Matthew, Twose, Claire, Park, Chanjun S., Tran, Phuong, Schechter, Loren S., Lau, Brandyn, and Coon, Devin
- Published
- 2022
- Full Text
- View/download PDF
15. Gender Affirming Surgery: A Comprehensive, Systematic Review of All Peer-reviewed Literature and Methods of Assessing Patient-centered Outcomes (Part 2: Genital Reconstruction).
- Author
-
Oles, Norah, Darrach, Halley, Landford, Wilmina, Garza, Matthew, Twose, Claire, Park, Chanjun S., Tran, Phuong, Schechter, Loren S., Lau, Brandyn, and Coon, Devin
- Published
- 2022
- Full Text
- View/download PDF
16. Development of Neural Network Model With Explainable AI for Measuring Uranium Enrichment.
- Author
-
Ryu, Jichang, Park, Chanjun, Park, Jungsuk, Cho, Namchan, Park, Jaehyun, and Cho, Gyuseong
- Subjects
URANIUM enrichment ,ARTIFICIAL neural networks ,NEURAL development ,URANIUM ,PRINCIPAL components analysis ,ARTIFICIAL intelligence - Abstract
In this work, we have developed a neural network (NN) model that can analyze enrichment from depleted (0.2%) to low enriched uranium (4.5%) when UO2 waste with very low radioactivity was contained in a 1-L Marinelli beaker, even when the measurement time is as short as 10 s using a low-resolution detector. The average count rate was about 3800 cps. Measurement of uranium enrichment is necessary for quantitative analysis of uranium radioactivity for disposal of uranium waste. Previously studied uranium enrichment methods (infinite thickness (IT) method, peak ratio (PR) method, and relative-efficiency (RE) method) are difficult to use for field measurement due to many limitations of the algorithms. Among existing methods, the RE method is accurate but requires a long measurement time; there is also a limitation in which a high-resolution detector is essential. In this work, we proposed a model to predict uranium enrichment using a low-resolution detector and an artificial NN model. Furthermore, we validated the results of the NN models using an explainable AI algorithm and principal component analysis (PCA). When the measurement time was less than 60 s, the existing method failed to analyze uranium enrichment, but the proposed model can predict enrichment of uranium within 5% of relative error when 5 g of uranium powder was mixed with various waste (ash, soil, and concrete). [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
17. Neural spelling correction: translating incorrect sentences to correct sentences for multimedia.
- Author
-
Park, Chanjun, Kim, Kuekyeng, Yang, YeongWook, Kang, Minho, and Lim, Heuiseok
- Subjects
MACHINE translating ,SPELLING errors ,ORTHOGRAPHY & spelling ,MULTIMEDIA systems ,CORPORA - Abstract
The aim of a spelling correction task is to detect spelling errors and automatically correct them. In this paper we aim to perform the Korean spelling correction task from a machine translation perspective, allowing it to overcome the limitations of cost, time and data. Based on a sequence to sequence model, the model aligns its source sentence with an 'error filled sentence' and its target sentence aligned to the correct counter part. Thus, 'translating' the error sentence to a correct sentence. For this research, we have also proposed three new data generation methods allowing the creation of multiple spelling correction parallel corpora from just a single monolingual corpus. Additionally, we discovered that applying the Copy Mechanism not only resolves the problem of overcorrection but even improves it. For this paper, we evaluated our model upon these aspects: Performance comparisons to other models and evaluation on overcorrection. The results show the proposed model to even out-perform other systems currently in commercial use. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
18. Variational Reward Estimator Bottleneck: Towards Robust Reward Estimator for Multidomain Task-Oriented Dialogue.
- Author
-
Park, Jeiyoon, Lee, Chanhee, Park, Chanjun, Kim, Kuekyeng, and Lim, Heuiseok
- Subjects
REINFORCEMENT learning ,QUANTITATIVE research - Abstract
Despite its significant effectiveness in adversarial training approaches to multidomain task-oriented dialogue systems, adversarial inverse reinforcement learning of the dialogue policy frequently fails to balance the performance of the reward estimator and policy generator. During the optimization process, the reward estimator frequently overwhelms the policy generator, resulting in excessively uninformative gradients. We propose the variational reward estimator bottleneck (VRB), which is a novel and effective regularization strategy that aims to constrain unproductive information flows between inputs and the reward estimator. The VRB focuses on capturing discriminative features by exploiting information bottleneck on mutual information. Quantitative analysis on a multidomain task-oriented dialogue dataset demonstrates that the VRB significantly outperforms previous studies. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
19. Comparative Analysis of Current Approaches to Quality Estimation for Neural Machine Translation.
- Author
-
Eo, Sugyeong, Park, Chanjun, Moon, Hyeonseok, Seo, Jaehyung, and Lim, Heuiseok
- Subjects
MACHINE translating ,DATA augmentation ,COMPARATIVE studies ,TRANSLATING & interpreting ,TASK performance - Abstract
Quality estimation (QE) has recently gained increasing interest as it can predict the quality of machine translation results without a reference translation. QE is an annual shared task at the Conference on Machine Translation (WMT), and most recent studies have applied the multilingual pretrained language model (mPLM) to address this task. Recent studies have focused on the performance improvement of this task using data augmentation with finetuning based on a large-scale mPLM. In this study, we eliminate the effects of data augmentation and conduct a pure performance comparison between various mPLMs. Separate from the recent performance-driven QE research involved in competitions addressing a shared task, we utilize the comparison for sub-tasks from WMT20 and identify an optimal mPLM. Moreover, we demonstrate QE using the multilingual BART model, which has not yet been utilized, and conduct comparative experiments and analyses with cross-lingual language models (XLMs), multilingual BERT, and XLM-RoBERTa. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
20. Exploring the Data Efficiency of Cross-Lingual Post-Training in Pretrained Language Models.
- Author
-
Lee, Chanhee, Yang, Kisu, Whang, Taesun, Park, Chanjun, Matteson, Andrew, Lim, Heuiseok, and Valencia-Garcia, Rafael
- Subjects
NATURAL language processing ,KOREAN language ,COMPUTER assisted language instruction ,DEEP learning - Abstract
Language model pretraining is an effective method for improving the performance of downstream natural language processing tasks. Even though language modeling is unsupervised and thus collecting data for it is relatively less expensive, it is still a challenging process for languages with limited resources. This results in great technological disparity between high- and low-resource languages for numerous downstream natural language processing tasks. In this paper, we aim to make this technology more accessible by enabling data efficient training of pretrained language models. It is achieved by formulating language modeling of low-resource languages as a domain adaptation task using transformer-based language models pretrained on corpora of high-resource languages. Our novel cross-lingual post-training approach selectively reuses parameters of the language model trained on a high-resource language and post-trains them while learning language-specific parameters in the low-resource language. We also propose implicit translation layers that can learn linguistic differences between languages at a sequence level. To evaluate our method, we post-train a RoBERTa model pretrained in English and conduct a case study for the Korean language. Quantitative results from intrinsic and extrinsic evaluations show that our method outperforms several massively multilingual and monolingual pretrained language models in most settings and improves the data efficiency by a factor of up to 32 compared to monolingual training. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
21. Planar multiresonance reactive shield for reducing electromagnetic interference in portable wireless power charging application.
- Author
-
Park, Jaehyoung, Park, Chanjun, Shin, Yujun, Kim, Dongwook, Park, Bumjin, Cho, Jaeyong, Choi, Junsung, and Ahn, Seungyoung
- Subjects
RESONANCE ,WIRELESS power transmission ,STORAGE batteries ,TOPOLOGY ,ENERGY transfer - Abstract
Wireless power transfer (WPT) technologies are currently growing in popularity because of problems with batteries in portable devices, such as the added weight and the inconvenience of conventional wired charging methods. However, wireless charging systems generate leakage magnetic fields, causing malfunctions in adjacent electronic devices. Technologies to suppress electromagnetic interference (EMI) from portable WPT systems are required to prevent this problem. In this paper, we propose a topology design for a multiresonance reactive shield to suppress the EMI from WPT systems. The proposed shield employs closed loops, which are located peripheral to the WPT coil and resonance matching capacitor. The power efficiency and shielding performance of the proposed method were compared with those of the conventional WPT system coil shielding topologies. The proposed shield exhibited about 27.37 dB more 3rd harmonic EMI suppression than other methods while maintaining higher power transfer efficiency. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
22. Decoding Strategies for Improving Low-Resource Machine Translation.
- Author
-
Park, Chanjun, Yang, Yeongwook, Park, Kinam, and Lim, Heuiseok
- Subjects
NATURAL language processing ,MACHINE translating ,TRANSLATIONS ,MACHINE performance ,APPLICATION software - Abstract
Pre-processing and post-processing are significant aspects of natural language processing (NLP) application software. Pre-processing in neural machine translation (NMT) includes subword tokenization to alleviate the problem of unknown words, parallel corpus filtering that only filters data suitable for training, and data augmentation to ensure that the corpus contains sufficient content. Post-processing includes automatic post editing and the application of various strategies during decoding in the translation process. Most recent NLP researches are based on the Pretrain-Finetuning Approach (PFA). However, when small and medium-sized organizations with insufficient hardware attempt to provide NLP services, throughput and memory problems often occur. These difficulties increase when utilizing PFA to process low-resource languages, as PFA requires large amounts of data, and the data for low-resource languages are often insufficient. Utilizing the current research premise that NMT model performance can be enhanced through various pre-processing and post-processing strategies without changing the model, we applied various decoding strategies to Korean–English NMT, which relies on a low-resource language pair. Through comparative experiments, we proved that translation performance could be enhanced without changes to the model. We experimentally examined how performance changed in response to beam size changes and n-gram blocking, and whether performance was enhanced when a length penalty was applied. The results showed that various decoding strategies enhance the performance and compare well with previous Korean–English NMT approaches. Therefore, the proposed methodology can improve the performance of NMT models, without the use of PFA; this presents a new perspective for improving machine translation performance. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.