9,918 results on '"Wu, Yuan"'
Search Results
2. Measurement of two-neutrino double electron capture half-life of $^{124}$Xe with PandaX-4T
- Author
-
PandaX Collaboration, Bo, Zihao, Chen, Wei, Chen, Xun, Chen, Yunhua, Cheng, Zhaokan, Cui, Xiangyi, Fan, Yingjie, Fang, Deqing, Gao, Zhixing, Geng, Lisheng, Giboni, Karl, Guo, Xunan, Guo, Xuyuan, Guo, Zichao, Han, Chencheng, Han, Ke, He, Changda, He, Jinrong, Huang, Di, Huang, Houqi, Huang, Junting, Hou, Ruquan, Hou, Yu, Ji, Xiangdong, Ji, Xiangpan, Ju, Yonglin, Li, Chenxiang, Li, Jiafu, Li, Mingchuan, Li, Shuaijie, Li, Tao, Li, Zhiyuan, Lin, Qing, Liu, Jianglai, Lu, Congcong, Lu, Xiaoying, Luo, Lingyin, Luo, Yunyang, Ma, Wenbo, Ma, Yugang, Mao, Yajun, Meng, Yue, Ning, Xuyang, Pang, Binyu, Qi, Ningchun, Qian, Zhicheng, Ren, Xiangxiang, Shan, Dong, Shang, Xiaofeng, Shao, Xiyuan, Shen, Guofang, Shen, Manbin, Sun, Wenliang, Tao, Yi, Wang, Anqing, Wang, Guanbo, Wang, Hao, Wang, Jiamin, Wang, Lei, Wang, Meng, Wang, Qiuhong, Wang, Shaobo, Wang, Siguang, Wang, Wei, Wang, Xiuli, Wang, Xu, Wang, Zhou, Wei, Yuehuan, Wu, Weihao, Wu, Yuan, Xiao, Mengjiao, Xiao, Xiang, Xiong, Kaizhi, Xu, Yifan, Yao, Shunyu, Yan, Binbin, Yan, Xiyu, Yang, Yong, Ye, Peihua, Yu, Chunxu, Yuan, Ying, Yuan, Zhe, Yun, Youhui, Zeng, Xinning, Zhang, Minzhen, Zhang, Peng, Zhang, Shibo, Zhang, Shu, Zhang, Tao, Zhang, Wei, Zhang, Yang, Zhang, Yingxin, Zhang, Yuanyuan, Zhao, Li, Zhou, Jifang, Zhou, Jiaxu, Zhou, Jiayi, Zhou, Ning, Zhou, Xiaopeng, Zhou, Yubo, and Zhou, Zhizhen
- Subjects
Nuclear Experiment - Abstract
Detailed studies of two-neutrino double electron capture (2$\nu$DEC) is a crucial step towards searching for the neutrino-less mode to explore the Majorana nature of neutrinos. We have measured precisely the half-life of the 2$\nu$DEC process in $^{124}$Xe, utilizing a total exposure of 1.73 tonne$\cdot$year from the commissioning run and the first science run of the PandaX-4T experiment. A time-dependent background model in the $\mathcal{O}$(10 keV) energy is constructed for the first time in PandaX-4T data. With an unbinned maximum likelihood fit, we determine the half-life of the 2$\nu$DEC process to be $(1.03\pm0.15_{\rm stat}\pm0.06_{\rm sys})\times 10^{22}$$\,$yr. Furthermore, we have evaluated the branching ratio for both electrons captured from the $K$ shell ($KK$) to be $(65\pm5)\%$, which aligns with the $^{124}$Xe nuclear model calculations within 1.5$\,$$\sigma$., Comment: 18 pages, 5 figures, 3 tables
- Published
- 2024
3. Atom-light-correlated quantum interferometer with memory-induced phase comb
- Author
-
Huang, Wenfeng, Liang, Xinyun, Zhao, Jie, Wu, Zeliang, Zhang, Keye, Yuan, Chun-Hua, Wu, Yuan, Fan, Bixuan, Zhang, Weiping, and Chen, Liqing
- Subjects
Quantum Physics - Abstract
Precise phase measurements by interferometers are crucial in science for detecting subtle changes, such as gravitational waves. However, phase sensitivity is typically limited by the standard quantum limit (SQL) with uncorrelated particles N. This limit can be surpassed using quantum correlations, but achieving high-quality correlations in large systems is challenging. Here, we propose and demonstrate an atom-light hybrid quantum interferometry whose sensitivity is enhanced beyond the SQL with atom-light quantum correlation and newly developed phase comb superposition via atomic-memory-assisted multiple quantum amplification. Finally, a phase sensitivity beyond the SQL of up to $8.3\pm 0.2$ dB is achieved, especially at $N=4 \times10^{13}/s$, resulting in both atomic and optical phase sensitivities of $6\times10^{-8} rad/\sqrt{Hz}$. This technique can advance sensitive quantum measurements in various fields., Comment: 11 pages, 3 figures
- Published
- 2024
4. Large Language Model Evaluation via Matrix Nuclear-Norm
- Author
-
Li, Yahan, Xia, Tingyu, Chang, Yi, and Wu, Yuan
- Subjects
Computer Science - Computation and Language - Abstract
As large language models (LLMs) continue to evolve, efficient evaluation metrics are vital for assessing their ability to compress information and reduce redundancy. While traditional metrics like Matrix Entropy offer valuable insights, they are computationally intensive for large-scale models due to their \( O(n^3) \) time complexity with Singular Value Decomposition (SVD). To mitigate this issue, we introduce the Matrix Nuclear-Norm, which not only serves as a metric to quantify the data compression proficiency of LLM but also provides a convex approximation of matrix rank to capture both predictive discriminability and diversity. By employing the \( L_{1,2}\text{-norm} \) to further approximate the nuclear norm, we can effectively assess the model's information compression capabilities. This approach reduces the time complexity to \( O(n^2) \) and eliminates the need for SVD computation. Consequently, the Matrix Nuclear-Norm achieves speeds 8 to 24 times faster than Matrix Entropy for the CEREBRAS-GPT model as sizes increase from 111M to 6.7B. This performance gap becomes more pronounced with larger models, as validated in tests with other models like Pythia. Additionally, evaluations on benchmarks and model responses confirm that our proposed Matrix Nuclear-Norm is a reliable, scalable, and efficient tool for assessing LLMs' performance, striking a balance between accuracy and computational efficiency. The code is available at https://github.com/MLGroupJLU/MatrixNuclearNorm., Comment: 22 pages
- Published
- 2024
5. Rethinking Data Selection at Scale: Random Selection is Almost All You Need
- Author
-
Xia, Tingyu, Yu, Bowen, Dang, Kai, Yang, An, Wu, Yuan, Tian, Yuan, Chang, Yi, and Lin, Junyang
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence - Abstract
Supervised fine-tuning (SFT) is crucial for aligning Large Language Models (LLMs) with human instructions. The primary goal during SFT is to select a small yet representative subset of training data from the larger pool, such that fine-tuning with this subset achieves results comparable to or even exceeding those obtained using the entire dataset. However, most existing data selection techniques are designed for small-scale data pools, which fail to meet the demands of real-world SFT scenarios. In this paper, we replicated several self-scoring methods those that do not rely on external model assistance on two million scale datasets, and found that nearly all methods struggled to significantly outperform random selection when dealing with such large-scale data pools. Moreover, our comparisons suggest that, during SFT, diversity in data selection is more critical than simply focusing on high quality data. We also analyzed the limitations of several current approaches, explaining why they perform poorly on large-scale datasets and why they are unsuitable for such contexts. Finally, we found that filtering data by token length offers a stable and efficient method for improving results. This approach, particularly when training on long text data, proves highly beneficial for relatively weaker base models, such as Llama3.
- Published
- 2024
6. Sequential Design with Derived Win Statistics
- Author
-
Zhang, Baoshan and Wu, Yuan
- Subjects
Statistics - Methodology - Abstract
The Win Ratio has gained significant traction in cardiovascular trials as a novel method for analyzing composite endpoints (Pocock and others, 2012). Compared with conventional approaches based on time to the first event, the Win Ratio accommodates the varying priorities and types of outcomes among components, potentially offering greater statistical power by fully utilizing the information contained within each outcome. However, studies using Win Ratio have largely been confined to fixed design, limiting flexibility for early decisions, such as stopping for futility or efficacy. Our study proposes a sequential design framework incorporating multiple interim analyses based on Win Ratio or Net Benefit statistics. Moreover, we provide rigorous proof of the canonical joint distribution for sequential Win Ratio and Net Benefit statistics, and an algorithm for sample size determination is developed. We also provide results from a finite sample simulation study, which show that our proposed method controls Type I error maintains power level, and has a smaller average sample size than the fixed design. A real study of cardiovascular study is applied to illustrate the proposed method., Comment: 36 pages, 1 figure, 5 tables
- Published
- 2024
7. Design and Experimental Application of a Radon Diffusion Chamber for Determining Diffusion Coefficients in Membrane Materials
- Author
-
Wu, Liang-Yu, Si, Lin, Wu, Yuan, Gao, Zhi-Xing, Heng, Yue-Kun, Li, Yuan, Liu, Jiang-Lai, Luo, Xiao-Lan, Ma, Fei, Meng, Yue, Qian, Xiao-Hui, Qian, Zhi-Cheng, Wang, Hao, Yun, You-Hui, Zhang, Gao-Feng, and Zhao, Jie
- Subjects
Physics - Instrumentation and Detectors ,High Energy Physics - Experiment - Abstract
In recent years, the issue of radon emanation and diffusion has become a critical concern for rare decay experiments, such as JUNO and PandaX-4T. This paper introduces a detector design featuring a symmetric radon detector cavity for the quantitative assessment of membrane materials' radon blocking capabilities. The performance of this design is evaluated through the application of Fick's Law and the diffusion equation considering material solubility. Our detector has completed measurements of radon diffusion coefficients for four types of membrane materials currently used in experiments, which also confirms the rationality of this detector design. The findings are instrumental in guiding the selection and evaluation of optimal materials for radon shielding to reduce radon background, contributing to boost sensitivities of rare event research., Comment: 7 pages, 10 figures and 2 tables
- Published
- 2024
8. CHBench: A Chinese Dataset for Evaluating Health in Large Language Models
- Author
-
Guo, Chenlu, Xu, Nuo, Chang, Yi, and Wu, Yuan
- Subjects
Computer Science - Computation and Language - Abstract
With the rapid development of large language models (LLMs), assessing their performance on health-related inquiries has become increasingly essential. It is critical that these models provide accurate and trustworthy health information, as their application in real-world contexts--where misinformation can have serious consequences for individuals seeking medical advice and support--depends on their reliability. In this work, we present CHBench, the first comprehensive Chinese Health-related Benchmark designed to evaluate LLMs' capabilities in understanding physical and mental health across diverse scenarios. CHBench includes 6,493 entries related to mental health and 2,999 entries focused on physical health, covering a broad spectrum of topics. This dataset serves as a foundation for evaluating Chinese LLMs' capacity to comprehend and generate accurate health-related information. Our extensive evaluations of four popular Chinese LLMs demonstrate that there remains considerable room for improvement in their understanding of health-related information. The code is available at https://github.com/TracyGuo2001/CHBench., Comment: 11 pages
- Published
- 2024
9. XTRUST: On the Multilingual Trustworthiness of Large Language Models
- Author
-
Li, Yahan, Wang, Yi, Chang, Yi, and Wu, Yuan
- Subjects
Computer Science - Computation and Language - Abstract
Large language models (LLMs) have demonstrated remarkable capabilities across a range of natural language processing (NLP) tasks, capturing the attention of both practitioners and the broader public. A key question that now preoccupies the AI community concerns the capabilities and limitations of these models, with trustworthiness emerging as a central issue, particularly as LLMs are increasingly applied in sensitive fields like healthcare and finance, where errors can have serious consequences. However, most previous studies on the trustworthiness of LLMs have been limited to a single language, typically the predominant one in the dataset, such as English. In response to the growing global deployment of LLMs, we introduce XTRUST, the first comprehensive multilingual trustworthiness benchmark. XTRUST encompasses a diverse range of topics, including illegal activities, hallucination, out-of-distribution (OOD) robustness, physical and mental health, toxicity, fairness, misinformation, privacy, and machine ethics, across 10 different languages. Using XTRUST, we conduct an empirical evaluation of the multilingual trustworthiness of five widely used LLMs, offering an in-depth analysis of their performance across languages and tasks. Our results indicate that many LLMs struggle with certain low-resource languages, such as Arabic and Russian, highlighting the considerable room for improvement in the multilingual trustworthiness of current language models. The code is available at https://github.com/LluckyYH/XTRUST., Comment: 21 pages
- Published
- 2024
10. Codec-SUPERB @ SLT 2024: A lightweight benchmark for neural audio codec models
- Author
-
Wu, Haibin, Chen, Xuanjun, Lin, Yi-Cheng, Chang, Kaiwei, Du, Jiawei, Lu, Ke-Han, Liu, Alexander H., Chung, Ho-Lam, Wu, Yuan-Kuei, Yang, Dongchao, Liu, Songxiang, Wu, Yi-Chiao, Tan, Xu, Glass, James, Watanabe, Shinji, and Lee, Hung-yi
- Subjects
Electrical Engineering and Systems Science - Audio and Speech Processing ,Computer Science - Sound - Abstract
Neural audio codec models are becoming increasingly important as they serve as tokenizers for audio, enabling efficient transmission or facilitating speech language modeling. The ideal neural audio codec should maintain content, paralinguistics, speaker characteristics, and audio information even at low bitrates. Recently, numerous advanced neural codec models have been proposed. However, codec models are often tested under varying experimental conditions. As a result, we introduce the Codec-SUPERB challenge at SLT 2024, designed to facilitate fair and lightweight comparisons among existing codec models and inspire advancements in the field. This challenge brings together representative speech applications and objective metrics, and carefully selects license-free datasets, sampling them into small sets to reduce evaluation computation costs. This paper presents the challenge's rules, datasets, five participant systems, results, and findings.
- Published
- 2024
11. Deep Height Decoupling for Precise Vision-based 3D Occupancy Prediction
- Author
-
Wu, Yuan, Yan, Zhiqiang, Wang, Zhengxue, Li, Xiang, Hui, Le, and Yang, Jian
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
The task of vision-based 3D occupancy prediction aims to reconstruct 3D geometry and estimate its semantic classes from 2D color images, where the 2D-to-3D view transformation is an indispensable step. Most previous methods conduct forward projection, such as BEVPooling and VoxelPooling, both of which map the 2D image features into 3D grids. However, the current grid representing features within a certain height range usually introduces many confusing features that belong to other height ranges. To address this challenge, we present Deep Height Decoupling (DHD), a novel framework that incorporates explicit height prior to filter out the confusing features. Specifically, DHD first predicts height maps via explicit supervision. Based on the height distribution statistics, DHD designs Mask Guided Height Sampling (MGHS) to adaptively decouple the height map into multiple binary masks. MGHS projects the 2D image features into multiple subspaces, where each grid contains features within reasonable height ranges. Finally, a Synergistic Feature Aggregation (SFA) module is deployed to enhance the feature representation through channel and spatial affinities, enabling further occupancy refinement. On the popular Occ3D-nuScenes benchmark, our method achieves state-of-the-art performance even with minimal input frames. Code is available at https://github.com/yanzq95/DHD.
- Published
- 2024
12. Searching for MeV-scale Axion-like Particles and Dark Photons with PandaX-4T
- Author
-
PandaX Collaboration, Li, Tao, Bo, Zihao, Chen, Wei, Chen, Xun, Chen, Yunhua, Cheng, Zhaokan, Cui, Xiangyi, Fan, Yingjie, Fang, Deqing, Gao, Zhixing, Geng, Lisheng, Giboni, Karl, Guo, Xunan, Guo, Xuyuan, Guo, Zichao, Han, Chencheng, He, Ke HanChangda, He, Jinrong, Huang, Di, Huang, Houqi, Huang, Junting, Hou, Ruquan, Hou, Yu, Ji, Xiangdong, Ji, Xiangpan, Ju, Yonglin, Li, Chenxiang, Li, Jiafu, Li, Mingchuan, Li, Shuaijie, Li, Zhiyuan, Lin, Qing, Liu, Jianglai, Lu, Congcong, Lu, Xiaoying, Luo, Lingyin, Luo, Yunyang, Ma, Wenbo, Ma, Yugang, Mao, Yajun, Meng, Yue, Ning, Xuyang, Pang, Binyu, Qi, Ningchun, Qian, Zhicheng, Ren, Xiangxiang, Shan, Dong, Shang, Xiaofeng, Shao, Xiyuan, Shen, Guofang, Shen, Manbin, Sun, Wenliang, Tao, Yi, Wang, Anqing, Wang, Guanbo, Wang, Hao, Wang, Jiamin, Wang, Lei, Wang, Meng, Wang, Qiuhong, Wang, Shaobo, Wang, Siguang, Wang, Wei, Wang, Xiuli, Wang, Xu, Wang, Zhou, Wei, Yuehuan, Wu, Weihao, Wu, Yuan, Xiao, Mengjiao, Xiao, Xiang, Xiong, Kaizhi, Xu, Yifan, Yao, Shunyu, Yan, Binbin, Yan, Xiyu, Yang, Yong, Ye, Peihua, Yu, Chunxu, Yuan, Ying, Yuan, Zhe, Yun, Youhui, Zeng, Xinning, Zhang, Minzhen, Zhang, Peng, Zhang, Shibo, Zhang, Shu, Zhang, Tao, Zhang, Wei, Zhang, Yang, Zhang, Yingxin, Zhang, Yuanyuan, Zhao, Li, Zhou, Jifang, Zhou, Jiaxu, Zhou, Jiayi, Zhou, Ning, Zhou, Xiaopeng, Zhou, Yubo, and Zhou, Zhizhen
- Subjects
High Energy Physics - Experiment - Abstract
Axion-like particles (ALPs) and dark photons (DPs) are viable dark matter particle candidates. We have searched for possible ALP/DP signals in the PandaX-4T liquid xenon detector using 94.8 days of data. A binned likelihood fit is constructed to search for possible mono-energetic peaks induced by the absorption processes between ALPs/DPs and atomic electrons of xenon. A detailed temporal model of decays associated with xenon isotopes is introduced to constrain the number of background events. No signal excess over background expectations is observed, and we have established the most stringent exclusion limits for most ALP/DP masses ranging from 150 keV/$c^2$ to 1 MeV/$c^2$.
- Published
- 2024
13. SpeechPrompt: Prompting Speech Language Models for Speech Processing Tasks
- Author
-
Chang, Kai-Wei, Wu, Haibin, Wang, Yu-Kai, Wu, Yuan-Kuei, Shen, Hua, Tseng, Wei-Cheng, Kang, Iu-thing, Li, Shang-Wen, and Lee, Hung-yi
- Subjects
Electrical Engineering and Systems Science - Audio and Speech Processing ,Computer Science - Artificial Intelligence ,Computer Science - Computation and Language ,Computer Science - Machine Learning - Abstract
Prompting has become a practical method for utilizing pre-trained language models (LMs). This approach offers several advantages. It allows an LM to adapt to new tasks with minimal training and parameter updates, thus achieving efficiency in both storage and computation. Additionally, prompting modifies only the LM's inputs and harnesses the generative capabilities of language models to address various downstream tasks in a unified manner. This significantly reduces the need for human labor in designing task-specific models. These advantages become even more evident as the number of tasks served by the LM scales up. Motivated by the strengths of prompting, we are the first to explore the potential of prompting speech LMs in the domain of speech processing. Recently, there has been a growing interest in converting speech into discrete units for language modeling. Our pioneer research demonstrates that these quantized speech units are highly versatile within our unified prompting framework. Not only can they serve as class labels, but they also contain rich phonetic information that can be re-synthesized back into speech signals for speech generation tasks. Specifically, we reformulate speech processing tasks into speech-to-unit generation tasks. As a result, we can seamlessly integrate tasks such as speech classification, sequence generation, and speech generation within a single, unified prompting framework. The experiment results show that the prompting method can achieve competitive performance compared to the strong fine-tuning method based on self-supervised learning models with a similar number of trainable parameters. The prompting method also shows promising results in the few-shot setting. Moreover, with the advanced speech LMs coming into the stage, the proposed prompting framework attains great potential., Comment: Published in IEEE/ACM Transactions on Audio, Speech, and Language Processing (TASLP)
- Published
- 2024
- Full Text
- View/download PDF
14. Exploring New Physics with PandaX-4T Low Energy Electronic Recoil Data
- Author
-
PandaX Collaboration, Zeng, Xinning, Bo, Zihao, Chen, Wei, Chen, Xun, Chen, Yunhua, Cheng, Zhaokan, Cui, Xiangyi, Fan, Yingjie, Fang, Deqing, Gao, Zhixing, Geng, Lisheng, Giboni, Karl, Guo, Xunan, Guo, Xuyuan, Guo, Zichao, Han, Chencheng, He, Ke HanChangda, He, Jinrong, Huang, Di, Huang, Houqi, Huang, Junting, Hou, Ruquan, Hou, Yu, Ji, Xiangdong, Ji, Xiangpan, Ju, Yonglin, Li, Chenxiang, Li, Jiafu, Li, Mingchuan, Li, Shuaijie, Li, Tao, Li, Zhiyuan, Lin, Qing, Liu, Jianglai, Lu, Congcong, Lu, Xiaoying, Luo, Lingyin, Luo, Yunyang, Ma, Wenbo, Ma, Yugang, Mao, Yajun, Meng, Yue, Ning, Xuyang, Pang, Binyu, Qi, Ningchun, Qian, Zhicheng, Ren, Xiangxiang, Shan, Dong, Shang, Xiaofeng, Shao, Xiyuan, Shen, Guofang, Shen, Manbin, Sun, Wenliang, Tao, Yi, Wang, Anqing, Wang, Guanbo, Wang, Hao, Wang, Jiamin, Wang, Lei, Wang, Meng, Wang, Qiuhong, Wang, Shaobo, Wang, Siguang, Wang, Wei, Wang, Xiuli, Wang, Xu, Wang, Zhou, Wei, Yuehuan, Wu, Weihao, Wu, Yuan, Xiao, Mengjiao, Xiao, Xiang, Xiong, Kaizhi, Xu, Yifan, Yao, Shunyu, Yan, Binbin, Yan, Xiyu, Yang, Yong, Ye, Peihua, Yu, Chunxu, Yuan, Ying, Yuan, Zhe, Yun, Youhui, Zhang, Minzhen, Zhang, Peng, Zhang, Shibo, Zhang, Shu, Zhang, Tao, Zhang, Wei, Zhang, Yang, Zhang, Yingxin, Zhang, Yuanyuan, Zhao, Li, Zhou, Jifang, Zhou, Jiaxu, Zhou, Jiayi, Zhou, Ning, Zhou, Xiaopeng, Zhou, Yubo, and Zhou, Zhizhen
- Subjects
High Energy Physics - Experiment - Abstract
New particles beyond the Standard Model of particle physics, such as axions, can be effectively searched through their interactions with electrons. We use the large liquid xenon detector PandaX-4T to search for novel electronic recoil signals induced by solar axions, neutrinos with anomalous magnetic moment, axion-like particles, dark photons, and light fermionic dark matter. A detailed background model is established with the latest datasets with 1.54 $\rm tonne \cdot year$ exposure. No significant excess above the background has been observed, and we have obtained competitive constraints for axion couplings, neutrino magnetic moment, and fermionic dark matter interactions.
- Published
- 2024
15. Digital Semantic Communications: An Alternating Multi-Phase Training Strategy with Mask Attack
- Author
-
Gong, Mingze, Wang, Shuoyao, Bi, Suzhi, Wu, Yuan, and Qian, Liping
- Subjects
Electrical Engineering and Systems Science - Signal Processing - Abstract
Semantic communication (SemComm) has emerged as new paradigm shifts.Most existing SemComm systems transmit continuously distributed signals in analog fashion.However, the analog paradigm is not compatible with current digital communication frameworks. In this paper, we propose an alternating multi-phase training strategy (AMP) to enable the joint training of the networks in the encoder and decoder through non-differentiable digital processes. AMP contains three training phases, aiming at feature extraction (FE), robustness enhancement (RE), and training-testing alignment (TTA), respectively. AMP contains three training phases, aiming at feature extraction (FE), robustness enhancement (RE), and training-testing alignment (TTA), respectively. In particular, in the FE stage, we learn the representation ability of semantic information by end-to-end training the encoder and decoder in an analog manner. When we take digital communication into consideration, the domain shift between digital and analog demands the fine-tuning for encoder and decoder. To cope with joint training process within the non-differentiable digital processes, we propose the alternation between updating the decoder individually and jointly training the codec in RE phase. To boost robustness further, we investigate a mask-attack (MATK) in RE to simulate an evident and severe bit-flipping effect in a differentiable manner. To address the training-testing inconsistency introduced by MATK, we employ an additional TTA phase, fine-tuning the decoder without MATK. Combining with AMP and an information restoration network, we propose a digital SemComm system for image transmission, named AMP-SC. Comparing with the representative benchmark, AMP-SC achieves $0.82 \sim 1.65$dB higher average reconstruction performance among various representative datasets at different scales and a wide range of signal-to-noise ratio.
- Published
- 2024
16. BA-LoRA: Bias-Alleviating Low-Rank Adaptation to Mitigate Catastrophic Inheritance in Large Language Models
- Author
-
Chang, Yupeng, Chang, Yi, and Wu, Yuan
- Subjects
Computer Science - Computation and Language ,Computer Science - Machine Learning - Abstract
Large language models (LLMs) have demonstrated remarkable proficiency across various natural language processing (NLP) tasks. However, adapting LLMs to downstream applications requires computationally intensive and memory-demanding fine-tuning procedures. To alleviate these burdens, parameter-efficient fine-tuning (PEFT) techniques have emerged as a promising approach to tailor LLMs with minimal computational overhead. While PEFT methods offer substantial advantages, they do not fully address the pervasive issue of bias propagation from pre-training data. This work introduces Bias-Alleviating Low-Rank Adaptation (BA-LoRA), a novel PEFT method designed to counteract bias inheritance. BA-LoRA incorporates three distinct regularization terms: (1) a consistency regularizer, (2) a diversity regularizer, and (3) a singular value decomposition regularizer. These regularizers aim to enhance the models' consistency, diversity, and generalization capabilities during fine-tuning. We conduct extensive experiments on natural language understanding (NLU) and natural language generation (NLG) tasks using prominent LLMs such as LLaMA, Mistral, and Gemma. The results demonstrate that BA-LoRA outperforms LoRA and its state-of-the-art variants. Moreover, our method effectively mitigates the adverse effects of pre-training bias, leading to more reliable and robust model outputs. The code is available at https://github.com/cyp-jlu-ai/BA-LoRA., Comment: 23 pages
- Published
- 2024
17. Dark Matter Search Results from 1.54 Tonne$\cdot$Year Exposure of PandaX-4T
- Author
-
PandaX Collaboration, Bo, Zihao, Chen, Wei, Chen, Xun, Chen, Yunhua, Cheng, Zhaokan, Cui, Xiangyi, Fan, Yingjie, Fang, Deqing, Gao, Zhixing, Geng, Lisheng, Giboni, Karl, Guo, Xunan, Guo, Xuyuan, Guo, Zichao, Han, Chencheng, Han, Ke, He, Changda, He, Jinrong, Huang, Di, Huang, Houqi, Huang, Junting, Hou, Ruquan, Hou, Yu, Ji, Xiangdong, Ji, Xiangpan, Ju, Yonglin, Li, Chenxiang, Li, Jiafu, Li, Mingchuan, Li, Shuaijie, Li, Tao, Li, Zhiyuan, Lin, Qing, Liu, Jianglai, Lu, Congcong, Lu, Xiaoying, Luo, Lingyin, Luo, Yunyang, Ma, Wenbo, Ma, Yugang, Mao, Yajun, Meng, Yue, Ning, Xuyang, Pang, Binyu, Qi, Ningchun, Qian, Zhicheng, Ren, Xiangxiang, Shan, Dong, Shang, Xiaofeng, Shao, Xiyuan, Shen, Guofang, Shen, Manbin, Sun, Wenliang, Tao, Yi, Wang, Anqing, Wang, Guanbo, Wang, Hao, Wang, Jiamin, Wang, Lei, Wang, Meng, Wang, Qiuhong, Wang, Shaobo, Wang, Siguang, Wang, Wei, Wang, Xiuli, Wang, Xu, Wang, Zhou, Wei, Yuehuan, Wu, Weihao, Wu, Yuan, Xiao, Mengjiao, Xiao, Xiang, Xiong, Kaizhi, Xu, Yifan, Yao, Shunyu, Yan, Binbin, Yan, Xiyu, Yang, Yong, Ye, Peihua, Yu, Chunxu, Yuan, Ying, Yuan, Zhe, Yun, Youhui, Zeng, Xinning, Zhang, Minzhen, Zhang, Peng, Zhang, Shibo, Zhang, Shu, Zhang, Tao, Zhang, Wei, Zhang, Yang, Zhang, Yingxin, Zhang, Yuanyuan, Zhao, Li, Zhou, Jifang, Zhou, Jiaxu, Zhou, Jiayi, Zhou, Ning, Zhou, Xiaopeng, Zhou, Yubo, and Zhou, Zhizhen
- Subjects
High Energy Physics - Experiment - Abstract
In this letter, we report the dark matter search results from the commissioning run and the first science run of the PandaX-4T experiment. A blind analysis is carried out on the entire data set. The data processing is improved compared to previous work, unifying the low-level signal reconstruction in a wide energy range up to 120 keV. With a total exposure of 1.54 tonne$\cdot$year, no significant excess of nuclear recoil events is found. The lowest 90% confidence level exclusion on the spin-independent cross section is $1.6 \times 10^{-47} \mathrm{cm}^2$ at a dark matter mass of 40 GeV$/c^2$. Our results represent the most stringent constraint for a dark matter mass above 100 GeV$/c^2$.
- Published
- 2024
18. Attention-based SIC Ordering and Power Allocation for Non-orthogonal Multiple Access Networks
- Author
-
Huang, Liang, Zhu, Bincheng, Nan, Runkai, Chi, Kaikai, and Wu, Yuan
- Subjects
Computer Science - Networking and Internet Architecture - Abstract
Non-orthogonal multiple access (NOMA) emerges as a superior technology for enhancing spectral efficiency compared to orthogonal multiple access. In NOMA networks, successive interference cancellation (SIC) plays a crucial role in decoding user signals sequentially. The challenge lies in the joint optimization of SIC ordering and power allocation, due to the factorial nature of ordering combinations. This study introduces an innovative solution, the Attention-based SIC Ordering and Power Allocation (ASOPA) framework, targeting an uplink NOMA network with dynamic SIC ordering. ASOPA aims to maximize weighted proportional fairness by employing deep reinforcement learning, strategically decomposing the problem into two manageable subproblems: SIC ordering optimization and optimal power allocation. Our approach utilizes an attention-based neural network, which processes instantaneous channel gains and user weights to determine the SIC decoding sequence for each user. Once the SIC ordering is established, the power allocation subproblem transforms into a convex optimization problem, enabling efficient calculation. Extensive simulations validate ASOPA's efficacy, demonstrating a performance closely paralleling the exhaustive method, with over 97% confidence in normalized network utility. Notably, ASOPA maintains a low execution latency of approximately 50 milliseconds in a ten-user NOMA network, aligning with static SIC ordering algorithms. Furthermore, ASOPA demonstrates superior performance in various NOMA network configurations, including scenarios with imperfect channel state information, multiple base stations, and multiple-antenna setups. Such results underscore ASOPA's robustness and effectiveness, highlighting its ability to excel across various NOMA network environments. The complete source code for ASOPA is accessible at https://github.com/Jil-Menzerna/ASOPA.
- Published
- 2024
19. First Indication of Solar $^8$B Neutrino Flux through Coherent Elastic Neutrino-Nucleus Scattering in PandaX-4T
- Author
-
PandaX Collaboration, Bo, Zihao, Chen, Wei, Chen, Xun, Chen, Yunhua, Cheng, Zhaokan, Cui, Xiangyi, Fan, Yingjie, Fang, Deqing, Gao, Zhixing, Geng, Lisheng, Giboni, Karl, Guo, Xunan, Guo, Xuyuan, Guo, Zichao, Han, Chencheng, Han, Ke, He, Changda, He, Jinrong, Huang, Di, Huang, Houqi, Huang, Junting, Hou, Ruquan, Hou, Yu, Ji, Xiangdong, Ji, Xiangpan, Ju, Yonglin, Li, Chenxiang, Li, Jiafu, Li, Mingchuan, Li, Shuaijie, Li, Tao, Li, Zhiyuan, Lin, Qing, Liu, Jianglai, Lu, Congcong, Lu, Xiaoying, Luo, Lingyin, Luo, Yunyang, Ma, Wenbo, Ma, Yugang, Mao, Yajun, Meng, Yue, Ning, Xuyang, Pang, Binyu, Qi, Ningchun, Qian, Zhicheng, Ren, Xiangxiang, Shan, Dong, Shang, Xiaofeng, Shao, Xiyuan, Shen, Guofang, Shen, Manbin, Sun, Wenliang, Tao, Yi, Wang, Anqing, Wang, Guanbo, Wang, Hao, Wang, Jiamin, Wang, Lei, Wang, Meng, Wang, Qiuhong, Wang, Shaobo, Wang, Siguang, Wang, Wei, Wang, Xiuli, Wang, Xu, Wang, Zhou, Wei, Yuehuan, Wu, Weihao, Wu, Yuan, Xiao, Mengjiao, Xiao, Xiang, Xiong, Kaizhi, Xu, Yifan, Yao, Shunyu, Yan, Binbin, Yan, Xiyu, Yang, Yong, Ye, Peihua, Yu, Chunxu, Yuan, Ying, Yuan, Zhe, Yun, Youhui, Zeng, Xinning, Zhang, Minzhen, Zhang, Peng, Zhang, Shibo, Zhang, Shu, Zhang, Tao, Zhang, Wei, Zhang, Yang, Zhang, Yingxin, Zhang, Yuanyuan, Zhao, Li, Zhou, Jifang, Zhou, Jiaxu, Zhou, Jiayi, Zhou, Ning, Zhou, Xiaopeng, Zhou, Yubo, and Zhou, Zhizhen
- Subjects
High Energy Physics - Experiment ,Astrophysics - Solar and Stellar Astrophysics ,Nuclear Experiment - Abstract
The PandaX-4T liquid xenon detector at the China Jinping Underground Laboratory is used to measure the solar $^8$B neutrino flux by detecting neutrinos through coherent scattering with xenon nuclei. Data samples requiring the coincidence of scintillation and ionization signals (paired), as well as unpaired ionization-only signals (US2), are selected with energy threshold of approximately 1.1 keV (0.33 keV) nuclear recoil energy. Combining the commissioning run and the first science run of PandaX-4T, a total exposure of 1.20 and 1.04 tonne$\cdot$year are collected for the paired and US2, respectively. After unblinding, 3 and 332 events are observed with an expectation of 2.8$\pm$0.5 and 251$\pm$32 background events, for the paired and US2 data, respectively. A combined analysis yields a best-fit $^8$B neutrino signal of 3.5 (75) events from the paired (US2) data sample, with $\sim$37\% uncertainty, and the background-only hypothesis is disfavored at 2.64$\sigma$ significance. This gives a solar $^8$B neutrino flux of ($8.4\pm3.1$)$\times$10$^6$ cm$^{-2}$s$^{-1}$, consistent with the standard solar model prediction. It is also the first indication of solar $^8$B neutrino ``fog'' in a dark matter direct detection experiment., Comment: Accepted by Physical Review Letters
- Published
- 2024
20. Realization of Conditional Operations through Transition Pathway Engineering
- Author
-
Zhang, Sheng, Duan, Peng, Wang, Yun-Jie, Wang, Tian-Le, Wang, Peng, Zhao, Ren-Ze, Yang, Xiao-Yan, Zhao, Ze-An, Guo, Liang-Liang, Chen, Yong, Zhang, Hai-Feng, Du, Lei, Tao, Hao-Ran, Li, Zhi-Fei, Wu, Yuan, Jia, Zhi-Long, Kong, Wei-Cheng, Chen, Zhao-Yun, Wu, Yu-Chun, and Guo, Guo-Ping
- Subjects
Quantum Physics - Abstract
In the NISQ era, achieving large-scale quantum computing demands compact circuits to mitigate decoherence and gate error accumulation. Quantum operations with diverse degrees of freedom hold promise for circuit compression, but conventional approaches encounter challenges in simultaneously adjusting multiple parameters. Here, we propose a transition composite gate (TCG) scheme grounded on state-selective transition path engineering, enabling more expressive conditional operations. We experimentally validate a controlled unitary (CU) gate as an example, with independent and continuous parameters. By adjusting the parameters of $\rm X^{12}$ gate, we obtain the CU family with a fidelity range of 95.2% to 99.0% leveraging quantum process tomography (QPT). To demonstrate the capability of circuit compression, we use TCG scheme to prepare 3-qubit Greenberger-Horne-Zeilinger (GHZ) and W states, with the fidelity of 96.77% and 95.72%. TCG can achieve the reduction in circuit depth of about 40% and 44% compared with the use of CZ gates only. Moreover, we show that short-path TCG (SPTCG) can further reduce the state-preparation circuit time cost. The TCG scheme exhibits advantages in certain quantum circuits and shows significant potential for large-scale quantum algorithms., Comment: 21 pages, 12 figures
- Published
- 2024
21. Stochastic Adversarial Networks for Multi-Domain Text Classification
- Author
-
Wang, Xu and Wu, Yuan
- Subjects
Computer Science - Computation and Language ,Computer Science - Machine Learning - Abstract
Adversarial training has been instrumental in advancing multi-domain text classification (MDTC). Traditionally, MDTC methods employ a shared-private paradigm, with a shared feature extractor for domain-invariant knowledge and individual private feature extractors for domain-specific knowledge. Despite achieving state-of-the-art results, these methods grapple with the escalating model parameters due to the continuous addition of new domains. To address this challenge, we introduce the Stochastic Adversarial Network (SAN), which innovatively models the parameters of the domain-specific feature extractor as a multivariate Gaussian distribution, as opposed to a traditional weight vector. This design allows for the generation of numerous domain-specific feature extractors without a substantial increase in model parameters, maintaining the model's size on par with that of a single domain-specific extractor. Furthermore, our approach integrates domain label smoothing and robust pseudo-label regularization to fortify the stability of adversarial training and to refine feature discriminability, respectively. The performance of our SAN, evaluated on two leading MDTC benchmarks, demonstrates its competitive edge against the current state-of-the-art methodologies. The code is available at https://github.com/wangxu0820/SAN., Comment: Technical report
- Published
- 2024
22. Language Models can Evaluate Themselves via Probability Discrepancy
- Author
-
Xia, Tingyu, Yu, Bowen, Wu, Yuan, Chang, Yi, and Zhou, Chang
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence - Abstract
In this paper, we initiate our discussion by demonstrating how Large Language Models (LLMs), when tasked with responding to queries, display a more even probability distribution in their answers if they are more adept, as opposed to their less skilled counterparts. Expanding on this foundational insight, we propose a new self-evaluation method ProbDiff for assessing the efficacy of various LLMs. This approach obviates the necessity for an additional evaluation model or the dependence on external, proprietary models like GPT-4 for judgment. It uniquely utilizes the LLMs being tested to compute the probability discrepancy between the initial response and its revised versions. A higher discrepancy for a given query between two LLMs indicates a relatively weaker capability. Our findings reveal that ProbDiff achieves results on par with those obtained from evaluations based on GPT-4, spanning a range of scenarios that include natural language generation (NLG) tasks such as translation, summarization, and our proposed Xiaohongshu blog writing task, and benchmarks for LLM evaluation like AlignBench, MT-Bench, and AlpacaEval, across LLMs of varying magnitudes., Comment: ACL 2024 Findings
- Published
- 2024
23. You Only Need Half: Boosting Data Augmentation by Using Partial Content
- Author
-
Hu, Juntao and Wu, Yuan
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
We propose a novel data augmentation method termed You Only Need hAlf (YONA), which simplifies the augmentation process. YONA bisects an image, substitutes one half with noise, and applies data augmentation techniques to the remaining half. This method reduces the redundant information in the original image, encourages neural networks to recognize objects from incomplete views, and significantly enhances neural networks' robustness. YONA is distinguished by its properties of parameter-free, straightforward application, enhancing various existing data augmentation strategies, and thereby bolstering neural networks' robustness without additional computational cost. To demonstrate YONA's efficacy, extensive experiments were carried out. These experiments confirm YONA's compatibility with diverse data augmentation methods and neural network architectures, yielding substantial improvements in CIFAR classification tasks, sometimes outperforming conventional image-level data augmentation methods. Furthermore, YONA markedly increases the resilience of neural networks to adversarial attacks. Additional experiments exploring YONA's variants conclusively show that masking half of an image optimizes performance. The code is available at https://github.com/HansMoe/YONA., Comment: Technical report,16 pages
- Published
- 2024
24. NegativePrompt: Leveraging Psychology for Large Language Models Enhancement via Negative Emotional Stimuli
- Author
-
Wang, Xu, Li, Cheng, Chang, Yi, Wang, Jindong, and Wu, Yuan
- Subjects
Computer Science - Computation and Language - Abstract
Large Language Models (LLMs) have become integral to a wide spectrum of applications, ranging from traditional computing tasks to advanced artificial intelligence (AI) applications. This widespread adoption has spurred extensive research into LLMs across various disciplines, including the social sciences. Notably, studies have revealed that LLMs possess emotional intelligence, which can be further developed through positive emotional stimuli. This discovery raises an intriguing question: can negative emotions similarly influence LLMs, potentially enhancing their performance? In response to this question, we introduce NegativePrompt, a novel approach underpinned by psychological principles, involving ten specifically designed negative emotional stimuli. We embark on rigorous experimental evaluations of five LLMs including Flan-T5-Large, Vicuna, Llama 2, ChatGPT, and GPT-4, across a set of 45 tasks. The results are revealing: NegativePrompt markedly enhances the performance of LLMs, evidenced by relative improvements of 12.89% in Instruction Induction tasks and 46.25% in BIG-Bench tasks. Moreover, we conduct attention visualization experiments to decipher the underlying mechanisms of NegativePrompt's influence. Our research contributes significantly to the understanding of LLMs and emotion interaction, demonstrating the practical efficacy of NegativePrompt as an emotion-driven method and offering novel insights for the enhancement of LLMs in real-world applications. The code is available at https://github.com/wangxu0820/NegativePrompt., Comment: This paper has been accepted by IJCAI 2024
- Published
- 2024
25. Vision Transformer-based Adversarial Domain Adaptation
- Author
-
Li, Yahan and Wu, Yuan
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Machine Learning - Abstract
Unsupervised domain adaptation (UDA) aims to transfer knowledge from a labeled source domain to an unlabeled target domain. The most recent UDA methods always resort to adversarial training to yield state-of-the-art results and a dominant number of existing UDA methods employ convolutional neural networks (CNNs) as feature extractors to learn domain invariant features. Vision transformer (ViT) has attracted tremendous attention since its emergence and has been widely used in various computer vision tasks, such as image classification, object detection, and semantic segmentation, yet its potential in adversarial domain adaptation has never been investigated. In this paper, we fill this gap by employing the ViT as the feature extractor in adversarial domain adaptation. Moreover, we empirically demonstrate that ViT can be a plug-and-play component in adversarial domain adaptation, which means directly replacing the CNN-based feature extractor in existing UDA methods with the ViT-based feature extractor can easily obtain performance improvement. The code is available at https://github.com/LluckyYH/VT-ADA., Comment: 6 pages
- Published
- 2024
26. Search for Cosmic-ray Boosted Sub-MeV Dark-Matter-Electron Scattering in PandaX-4T
- Author
-
Shang, Xiaofeng, Abdukerim, Abdusalam, Bo, Zihao, Chen, Wei, Chen, Xun, Cheng, Chen, Cheng, Zhaokan, Cui, Xiangyi, Fan, Yingjie, Fang, Deqing, Geng, Lisheng, Giboni, Karl, Guo, Xuyuan, Han, Chencheng, Han, Ke, He, Changda, He, Jinrong, Huang, Di, Huang, Junting, Huang, Zhou, Hou, Ruquan, Hou, Yu, Ji, Xiangdong, Ju, Yonglin, Li, Chenxiang, Li, Jiafu, Li, Mingchuan, Li, Shuaijie, Li, Tao, Lin, Qing, Liu, Jianglai, Lu, Congcong, Lu, Xiaoying, Luo, Lingyin, Luo, Yunyang, Ma, Wenbo, Ma, Yugang, Mao, Yajun, Meng, Yue, Ning, Xuyang, Pang, Binyu, Qi, Ningchun, Qian, Zhicheng, Ren, Xiangxiang, Shaheed, Nasir, Shao, Xiyuan, Shen, Guofang, Si, Lin, Sun, Wenliang, Tan, Andi, Tao, Yi, Wang, Anqing, Wang, Meng, Wang, Qiuhong, Wang, Shaobo, Wang, Siguang, Wang, Wei, Wang, Xiuli, Wang, Xu, Wang, Zhou, Wei, Yuehuan, Wu, Mengmeng, Wu, Weihao, Wu, Yuan, Xiao, Mengjiao, Xiao, Xiang, Yan, Binbin, Yan, Xiyu, Yang, Yong, Yu, Chunxu, Yuan, Ying, Yuan, Zhe, Yun, Youhui, Zeng, Xinning, Zhang, Minzhen, Zhang, Peng, Zhang, Shibo, Zhang, Shu, Zhang, Tao, Zhang, Wei, Zhang, Yang, Zhang, Yingxin, Zhang, Yuanyuan, Zhao, Li, Zhou, Jifang, Zhou, Ning, Zhou, Xiaopeng, Zhou, Yong, Zhou, Yubo, Zhou, Zhizhen, Ge, Shao-Feng, and Xia, Chen
- Subjects
High Energy Physics - Experiment ,High Energy Physics - Phenomenology - Abstract
We report the first search for the elastic scatterings between cosmic-ray boosted sub-MeV dark matter and electrons in the PandaX-4T liquid xenon experiment. Sub-MeV dark matter particles can be accelerated by scattering with electrons in the cosmic rays and produce detectable electron recoil signals in the detector. Using the commissioning data from PandaX-4T of 0.63~tonne$\cdot$year exposure, we set new constraints on DM-electron scattering cross sections for DM masses ranging from 10~eV/$c^2$ to 3~keV/$c^2$., Comment: 6 pages, 3 figures
- Published
- 2024
- Full Text
- View/download PDF
27. Margin Discrepancy-based Adversarial Training for Multi-Domain Text Classification
- Author
-
Wu, Yuan
- Subjects
Computer Science - Computation and Language ,Computer Science - Machine Learning - Abstract
Multi-domain text classification (MDTC) endeavors to harness available resources from correlated domains to enhance the classification accuracy of the target domain. Presently, most MDTC approaches that embrace adversarial training and the shared-private paradigm exhibit cutting-edge performance. Unfortunately, these methods face a non-negligible challenge: the absence of theoretical guarantees in the design of MDTC algorithms. The dearth of theoretical underpinning poses a substantial impediment to the advancement of MDTC algorithms. To tackle this problem, we first provide a theoretical analysis of MDTC by decomposing the MDTC task into multiple domain adaptation tasks. We incorporate the margin discrepancy as the measure of domain divergence and establish a new generalization bound based on Rademacher complexity. Subsequently, we propose a margin discrepancy-based adversarial training (MDAT) approach for MDTC, in accordance with our theoretical analysis. To validate the efficacy of the proposed MDAT method, we conduct empirical studies on two MDTC benchmarks. The experimental results demonstrate that our MDAT approach surpasses state-of-the-art baselines on both datasets., Comment: 16 pages
- Published
- 2024
28. Oxygen vacancies enriched Ir/WOx catalysts for the directly chem-catalytic conversion of cellulose to ethanol
- Author
-
Weng, Yu-Jing, Ding, Zhao-Ying, Li, Ying-Chao, Wu, Yuan-Feng, Xu, Yuan-Yang, Chen, Rui, Zhao, Xiao-Lei, Wang, Hai-Yong, Zhang, Da-Lei, and Zhang, Yu-Long
- Published
- 2024
- Full Text
- View/download PDF
29. Frontiers in high entropy alloys and high entropy functional materials
- Author
-
Zhang, Wen-Tao, Wang, Xue-Qian, Zhang, Feng-Qi, Cui, Xiao-Ya, Fan, Bing-Bing, Guo, Jia-Ming, Guo, Zhi-Min, Huang, Rui, Huang, Wen, Li, Xu-Bo, Li, Meng-Ru, Ma, Yan, Shen, Zhi-Hua, Sun, Yong-Gang, Wang, De-Zhuang, Wang, Fei-Yang, Wang, Li-Qiang, Wang, Nan, Wang, Tian-Li, Wang, Wei, Wang, Xiao-Yang, Wang, Yi-Han, Yu, Fu-Jie, Yin, Yu-Zhen, Zhang, Ling-Kun, Zhang, Yi, Zhang, Jian-Yang, Zhao, Qi, Zhao, Yu-Ping, Zhu, Xin-Dong, Sohail, Yasir, Chen, Ya-Nan, Feng, Tao, Gao, Qi-Long, He, Hai-Yan, Huang, Yong-Jiang, Jiao, Zeng-Bao, Ji, Hua, Jiang, Yao, Li, Qiang, Li, Xiao-Ming, Liao, Wei-Bing, Lin, Huai-Jun, Liu, Hui, Liu, Qi, Liu, Qing-Feng, Liu, Wei-Di, Liu, Xiong-Jun, Lu, Yang, Lu, Yi-Ping, Ma, Wen, Miao, Xue-Fei, Pan, Jie, Wang, Qing, Wu, Hong-Hui, Wu, Yuan, Yang, Tao, Yang, Wei-Ming, Yu, Qian, Zhang, Jin-Yu, Chen, Zhi-Gang, Mao, Liang, Ren, Yang, Shen, Bao-Long, Wang, Xun-Li, Jia, Zhe, Zhu, He, Wu, Zhen-Duo, and Lan, Si
- Published
- 2024
- Full Text
- View/download PDF
30. Association of the retinol to all-trans retinoic acid pathway with autism spectrum disorder
- Author
-
Feng, Yu-Ru, Zhang, Qian, Miao, Jing-Kun, Yang, Ting, Chen, Jie, Chen, Hong-Yu, Mou, Qiu-Hong, Xiang, Xue-Li, Long, Dan, Wei, Qiu-Hong, Wu, Yuan, and Li, Ting-Yu
- Published
- 2024
- Full Text
- View/download PDF
31. Efficient hydrogen transfer carriers: hydrogenation mechanism of dibenzyltoluene catalyzed by Mg-based metal hydride
- Author
-
Deng, Hai-Yu, Jiang, Li-Jun, Wang, Shao-Hua, Jiang, Wen-Quan, Wu, Yuan-Fang, Guo, Xiu-Mei, Wang, Shu-Mao, and Hao, Lei
- Published
- 2024
- Full Text
- View/download PDF
32. Efficient Copper Adsorption from Wastewater Using Silica Nanoparticles Derived from Incinerated Coconut Shell Ash
- Author
-
Arumugam, Maathiniyaar, Gopinath, Subash C. B., Anbu, Periasamy, Packirisamy, Vinitha, Yaakub, Ahmad Radi Wan, and Wu, Yuan Seng
- Published
- 2024
- Full Text
- View/download PDF
33. Codec-SUPERB: An In-Depth Analysis of Sound Codec Models
- Author
-
Wu, Haibin, Chung, Ho-Lam, Lin, Yi-Cheng, Wu, Yuan-Kuei, Chen, Xuanjun, Pai, Yu-Chi, Wang, Hsiu-Hsuan, Chang, Kai-Wei, Liu, Alexander H., and Lee, Hung-yi
- Subjects
Electrical Engineering and Systems Science - Audio and Speech Processing ,Computer Science - Sound - Abstract
The sound codec's dual roles in minimizing data transmission latency and serving as tokenizers underscore its critical importance. Recent years have witnessed significant developments in codec models. The ideal sound codec should preserve content, paralinguistics, speakers, and audio information. However, the question of which codec achieves optimal sound information preservation remains unanswered, as in different papers, models are evaluated on their selected experimental settings. This study introduces Codec-SUPERB, an acronym for Codec sound processing Universal PERformance Benchmark. It is an ecosystem designed to assess codec models across representative sound applications and signal-level metrics rooted in sound domain knowledge.Codec-SUPERB simplifies result sharing through an online leaderboard, promoting collaboration within a community-driven benchmark database, thereby stimulating new development cycles for codecs. Furthermore, we undertake an in-depth analysis to offer insights into codec models from both application and signal perspectives, diverging from previous codec papers mainly concentrating on signal-level comparisons. Finally, we will release codes, the leaderboard, and data to accelerate progress within the community., Comment: Github: https://github.com/voidful/Codec-SUPERB
- Published
- 2024
34. Multi-Scale Semantic Segmentation with Modified MBConv Blocks
- Author
-
Chen, Xi, Cai, Yang, Wu, Yuan, Xiong, Bo, and Park, Taesung
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Recently, MBConv blocks, initially designed for efficiency in resource-limited settings and later adapted for cutting-edge image classification performances, have demonstrated significant potential in image classification tasks. Despite their success, their application in semantic segmentation has remained relatively unexplored. This paper introduces a novel adaptation of MBConv blocks specifically tailored for semantic segmentation. Our modification stems from the insight that semantic segmentation requires the extraction of more detailed spatial information than image classification. We argue that to effectively perform multi-scale semantic segmentation, each branch of a U-Net architecture, regardless of its resolution, should possess equivalent segmentation capabilities. By implementing these changes, our approach achieves impressive mean Intersection over Union (IoU) scores of 84.5% and 84.0% on the Cityscapes test and validation datasets, respectively, demonstrating the efficacy of our proposed modifications in enhancing semantic segmentation performance.
- Published
- 2024
35. PandaX-xT: a Multi-ten-tonne Liquid Xenon Observatory at the China Jinping Underground Laboratory
- Author
-
PandaX Collaboration, Abdukerim, Abdusalam, Bo, Zihao, Chen, Wei, Chen, Xun, Cheng, Chen, Cheng, Zhaokan, Cui, Xiangyi, Fan, Yingjie, Fang, Deqing, Geng, Lisheng, Giboni, Karl, Gu, Linhui, Guo, Xunan, Guo, Xuyuan, Guo, Zhichao, Han, Chencheng, Han, Ke, He, Changda, He, Jinrong, Huang, Di, Huang, Junting, Huang, Zhou, Hou, Ruquan, Hou, Yu, Ji, Xiangdong, Ju, Yonglin, Li, Chenxiang, Li, Jiafu, Li, Mingchuan, Li, Shuaijie, Li, Tao, Lin, Qing, Liu, Jianglai, Lu, Congcong, Lu, Xiaoying, Luo, Lingyin, Luo, Yunyang, Ma, Wenbo, Ma, Yugang, Mao, Yajun, Meng, Yue, Ning, Xuyang, Pang, Binyu, Qi, Ningchun, Qian, Zhicheng, Ren, Xiangxiang, Shaheed, Nasir, Shang, Xiaofeng, Shao, Xiyuan, Shen, Guofang, Si, Lin, Sun, Wenliang, Tao, Yi, Wang, Anqing, Wang, Meng, Wang, Qiuhong, Wang, Shaobo, Wang, Siguang, Wang, Wei, Wang, Xiuli, Wang, Xu, Wang, Zhou, Wei, Yuehuan, Wu, Mengmeng, Wu, Weihao, Wu, Yuan, Xiao, Mengjiao, Xiao, Xiang, Yan, Binbin, Yan, Xiyu, Yang, Yong, Yu, Chunxu, Yuan, Ying, Yuan, Zhe, Yun, Youhui, Zeng, Xinning, Zhang, Minzhen, Zhang, Peng, Zhang, Shibo, Zhang, Shu, Zhang, Tao, Zhang, Wei, Zhang, Yang, Zhang, Yingxin, Zhang, Yuanyuan, Zhao, Li, Zhou, Jifang, Zhou, Ning, Zhou, Xiaopeng, Zhou, Yong, Zhou, Yubo, and Zhou, Zhizhen
- Subjects
High Energy Physics - Experiment ,Astrophysics - Cosmology and Nongalactic Astrophysics ,High Energy Physics - Phenomenology - Abstract
We propose a major upgrade to the existing PandaX-4T experiment in the China Jinping Underground Laboratory. The new experiment, PandaX-xT, will be a multi-ten-tonne liquid xenon, ultra-low background, and general-purpose observatory. The full-scaled PandaX-xT contains a 43-tonne liquid xenon active target. Such an experiment will significantly advance our fundamental understanding of particle physics and astrophysics. The sensitivity of dark matter direct detection will be improved by nearly two orders of magnitude compared to the current best limits, approaching the so-called "neutrino floor" for a dark matter mass above 10 GeV/$c^2$, providing a decisive test to the Weakly Interacting Massive Particle paradigm. By searching for the neutrinoless double beta decay of $^{136}$Xe isotope in the detector, the effective Majorana neutrino mass can be measured to a [10 -- 41] meV/$c^2$ sensitivity, providing a key test to the Dirac/Majorana nature of neutrino s. Astrophysical neutrinos and other ultra-rare interactions can also be measured and searched for with an unprecedented background level, opening up new windows of discovery. Depending on the findings, PandaX-xT will seek the next stage upgrade utilizing isotopic separation on natural xenon.
- Published
- 2024
36. A Survey on Data Augmentation in Large Model Era
- Author
-
Zhou, Yue, Guo, Chenlu, Wang, Xu, Chang, Yi, and Wu, Yuan
- Subjects
Computer Science - Machine Learning ,Computer Science - Computation and Language ,Computer Science - Computer Vision and Pattern Recognition - Abstract
Large models, encompassing large language and diffusion models, have shown exceptional promise in approximating human-level intelligence, garnering significant interest from both academic and industrial spheres. However, the training of these large models necessitates vast quantities of high-quality data, and with continuous updates to these models, the existing reservoir of high-quality data may soon be depleted. This challenge has catalyzed a surge in research focused on data augmentation methods. Leveraging large models, these data augmentation techniques have outperformed traditional approaches. This paper offers an exhaustive review of large model-driven data augmentation methods, adopting a comprehensive perspective. We begin by establishing a classification of relevant studies into three main categories: image augmentation, text augmentation, and paired data augmentation. Following this, we delve into various data post-processing techniques pertinent to large model-based data augmentation. Our discussion then expands to encompass the array of applications for these data augmentation methods within natural language processing, computer vision, and audio signal processing. We proceed to evaluate the successes and limitations of large model-based data augmentation across different scenarios. Concluding our review, we highlight prospective challenges and avenues for future exploration in the field of data augmentation. Our objective is to furnish researchers with critical insights, ultimately contributing to the advancement of more sophisticated large models. We consistently maintain the related open-source materials at: https://github.com/MLGroup-JLU/LLM-data-aug-survey., Comment: 33 pages; https://github.com/MLGroup-JLU/LLM-data-aug-survey
- Published
- 2024
37. Measurement of Solar $pp$ Neutrino Flux using Electron Recoil Data from PandaX-4T Commissioning Run
- Author
-
PandaX Collaboration, Lu, Xiaoying, Abdukerim, Abdusalam, Bo, Zihao, Chen, Wei, Chen, Xun, Chen, Yunhua, Cheng, Chen, Cheng, Zhaokan, Cui, Xiangyi, Fan, Yingjie, Fang, Deqing, Geng, Lisheng, Giboni, Karl, Guo, Xuyuan, Han, Chencheng, Han, Ke, He, Changda, He, Jinrong, Huang, Di, Huang, Junting, Huang, Zhou, Hou, Ruquan, Hou, Yu, Ji, Xiangdong, Ju, Yonglin, Li, Chenxiang, Li, Jiafu, Li, Mingchuan, Li, Shuaijie, Li, Tao, Lin, Qing, Liu, Jianglai, Lu, Congcong, Luo, Lingyin, Luo, Yunyang, Ma, Wenbo, Ma, Yugang, Mao, Yajun, Meng, Yue, Ning, Xuyang, Pang, Binyu, Qi, Ningchun, Qian, Zhicheng, Ren, Xiangxiang, Shaheed, Nasir, Shang, Xiaofeng, Shao, Xiyuan, Shen, Guofang, Shen, Manbin, Si, Lin, Sun, Wenliang, Tao, Yi, Wang, Anqing, Wang, Meng, Wang, Qiuhong, Wang, Shaobo, Wang, Siguang, Wang, Wei, Wang, Xiuli, Wang, Xu, Wang, Zhou, Wei, Yuehuan, Wu, Mengmeng, Wu, Weihao, Wu, Yuan, Xiao, Mengjiao, Xiao, Xiang, Xiong, Kaizhi, Yan, Binbin, Yan, Xiyu, Yang, Yong, Yu, Chunxu, Yuan, Ying, Yuan, Zhe, Yun, Youhui, Zeng, Xinning, Zhang, Minzhen, Zhang, Peng, Zhang, Shibo, Zhang, Shu, Zhang, Tao, Zhang, Wei, Zhang, Yang, Zhang, Yingxin, Zhang, Yuanyuan, Zhao, Li, Zhou, Jifang, Zhou, Ning, Zhou, Xiaopeng, Zhou, Yubo, and Zhou, Zhizhen
- Subjects
High Energy Physics - Experiment - Abstract
The proton-proton ($pp$) fusion chain dominates the neutrino production from the Sun. The uncertainty of the predicted $pp$ neutrino flux is at the sub-percent level, whereas that of the best measurement is $\mathcal{O}(10\%)$. In this paper, we present the first result to measure the solar $pp$ neutrinos in the electron recoil energy range from 24 to 144 keV, using the PandaX-4T commissioning data with 0.63 tonne$\times$year exposure. The $pp$ neutrino flux is determined to be $(8.0 \pm 3.9 \,{\rm{(stat)}} \pm 10.0 \,{\rm{(syst)}} )\times 10^{10}\, $$\rm{s}^{-1} \rm{cm}^{-2}$, consistent with Standard Solar Model and existing measurements, corresponding to a flux upper limit of $23.3\times 10^{10}\, $$\rm{s}^{-1} \rm{cm}^{-2}$ at 90\% C.L.., Comment: 6 pages, 5 figures
- Published
- 2024
38. Factors resisting protein adsorption on hydrophilic/hydrophobic self-assembled monolayers terminated with hydrophilic hydroxyl groups
- Author
-
Mao, Dangxin, Wu, Yuan-Yan, and Tu, Yusong
- Subjects
Physics - Biological Physics - Abstract
The hydroxyl-terminated self-assembled monolayer (OH-SAM), as a surface resistant to protein adsorption, exhibits substantial potential in applications such as ship navigation and medical implants, and the appropriate strategies for designing anti-fouling surfaces are crucial. Here, we employ molecular dynamics simulations and alchemical free energy calculations to systematically analyze the factors influencing resistance to protein adsorption on the SAMs terminated with single or double OH groups at three packing densities ({\Sigma} = 2.0 nm-2, 4.5 nm-2, and 6.5 nm-2), respectively. For the first time, we observe that the compactness and order of interfacial water enhance its physical barrier effect, subsequently enhancing the resistance of SAM to protein adsorption. Notably, the weak spatial hindrance effect of SAM leads to the embedding of protein into SAM, resulting in a lack of resistance of SAM towards protein. Furthermore, the number of hydroxyl groups per unit area of double OH-terminated SAM at {\Sigma} = 6.5 nm-2 is approximately 2 to 3 times that of single OH-terminated SAM at {\Sigma} = 6.5 nm-2 and 4.5 nm-2, consequently yielding a weaker resistance of double OH-terminated SAM towards protein. Meanwhile, due to the structure of SAM itself, i.e., the formation of a nearly perfect ice-like hydrogen bond structure, the SAM exhibits the weakest resistance towards protein. This study will complement and improve the mechanism of OH-SAM resistance to protein adsorption, especially the traditional barrier effect of interfacial water.
- Published
- 2024
39. Elastic strain-induced amorphization in high-entropy alloys.
- Author
-
Lu, Zhaoping, Yang, Wei, Bu, Yeqiang, Wu, Yuan, Lei, Zhifeng, Yuan, Xiaoyuan, Liu, Leqing, Wang, Peng, Liu, Xiongjun, Wu, Honghui, Liu, Jiabin, Wang, Hongtao, and Ritchie, Robert
- Abstract
Elastic stability is the basis for understanding structural responses to external stimuli in crystalline solids, including melting, incipient plasticity and fracture. In this work, elastic stability is investigated in a series of high-entropy alloys (HEAs) using in situ mechanical tests and atomic-resolution characterization in transmission electron microscopy. Under tensile loading, the HEA lattices are observed to undergo a sudden loss of ordering as the elastic strain reached ∽ 10%. Such elastic strain-induced amorphization stands in intrinsic contrast to previously reported dislocation-mediated elastic instability and defect accumulation-mediated amorphization, introducing a form of elastic instability. Together with the first principle calculations and atomic-resolution chemical mapping, we identify that the elastic strain-induced amorphization is closely related to the depressed dislocation nucleation due to the local atomic environment inhomogeneity of HEAs. Our findings provide insights for the understanding of the fundamental nature of physical mechanical phenomena like elastic instability and incipient plasticity.
- Published
- 2024
40. Regularized Conditional Alignment for Multi-Domain Text Classification
- Author
-
Hu, Juntao and Wu, Yuan
- Subjects
Computer Science - Computation and Language ,Computer Science - Cryptography and Security ,Computer Science - Machine Learning - Abstract
The most successful multi-domain text classification (MDTC) approaches employ the shared-private paradigm to facilitate the enhancement of domain-invariant features through domain-specific attributes. Additionally, they employ adversarial training to align marginal feature distributions. Nevertheless, these methodologies encounter two primary challenges: (1) Neglecting class-aware information during adversarial alignment poses a risk of misalignment; (2) The limited availability of labeled data across multiple domains fails to ensure adequate discriminative capacity for the model. To tackle these issues, we propose a method called Regularized Conditional Alignment (RCA) to align the joint distributions of domains and classes, thus matching features within the same category and amplifying the discriminative qualities of acquired features. Moreover, we employ entropy minimization and virtual adversarial training to constrain the uncertainty of predictions pertaining to unlabeled data and enhance the model's robustness. Empirical results on two benchmark datasets demonstrate that our RCA approach outperforms state-of-the-art MDTC techniques., Comment: This paper has been accepted by ICASSP 2024
- Published
- 2023
41. A Comparison of Functional Features in Chinese and US Mobile Apps for Diabetes Self-Management: A Systematic Search in App Stores and Content Analysis
- Author
-
Wu, Yuan, Zhou, Yiling, Wang, Xuan, Zhang, Qi, Yao, Xun, Li, Xiaodan, Li, Jianshu, Tian, Haoming, and Li, Sheyu
- Subjects
Information technology ,T58.5-58.64 ,Public aspects of medicine ,RA1-1270 - Abstract
BackgroundMobile health interventions are widely used for self-management of diabetes, which is one of the most burdensome noncommunicable chronic diseases worldwide. However, little is known about the distribution of characteristics and functions of in-store mobile apps for diabetes. ObjectiveThis study aimed to investigate the distribution of characteristics and functions of the in-store mobile apps for self-management of diabetes in the United States and China using a predefined functional taxonomy, which was developed and published in our previous study. MethodsWe identified apps by searching diabetes in English or Chinese in the Apple iTunes Store and Android Markets (both in the United States and China) and included apps for diabetes self-management. We examined the validity and reliability of the predefined functional taxonomy with 3 dimensions: clinical module, functional module, and potential risk. We then classified all functions in the included apps according to the predefined taxonomy and compared the differences in the features of these apps between the United States and China. ResultsWe included 171 mobile diabetes apps, with 133 from the United States and 38 from China. Apps from both countries faced the challenges of evidence-based information, proper risk assessment, and declaration, especially Chinese apps. More Chinese apps provide app-based communication functions (general communication: Chinese vs US apps, 39%, 15/38 vs 18.0%, 24/133; P=.006 and patient-clinician communication: Chinese vs US apps, 68%, 26/38 vs 6.0%, 8/133; P.99) are neglected by the 2 countries. ConclusionsThe distribution of characteristics and functions of in-store mobile apps for diabetes self-management in the United States was different from China. The design of in-store diabetes apps needs to be monitored closely.
- Published
- 2019
- Full Text
- View/download PDF
42. Exploring the Effect of Learning Motivation and Self-Regulated Learning Climate on Undergraduates' Self-Regulated Learning in Higher Education
- Author
-
Ma, Yujie and Guo, Wu Yuan
- Abstract
Self-regulated learning (SRL) has numerous benefits for students. For this reason, it is necessary to understand what factors can affect SRL. The literature has indicated that many factors can influence SRL as a whole. However, few studies have investigated the relationship between factors that can affect SRL and different phases of SRL. This research aims to investigate the influence of learning motivation and self-regulated learning climate on three phases of SRL (forethought, performance, and self-reflection) in higher education context. Thirteen Chinese undergraduates participated in semi-structured interviews. The results reveal that both learning motivation and self-regulated learning climate have significant impact on forethought, performance, and self-reflection. This study also demonstrates that personality may play a role of mediator in the effect of learning motivation on SRL. The influence of self-regulated learning climate on SRL is related to the level of interaction between teachers and students. These findings can help to further develop undergraduates' SRL ability.
- Published
- 2023
43. Novel high-entropy ultra-high temperature ceramics with enhanced ablation resistance
- Author
-
Zhang, Pan, Liu, Xiong-Jun, He, Guang-Yu, Chiang, Fu-Kuo, Wang, Hui, Wu, Yuan, Jiang, Sui-He, Zhang, Xiao-Bin, and Lu, Zhao-Ping
- Published
- 2024
- Full Text
- View/download PDF
44. Phase field modeling of grain stability of nanocrystalline alloys by explicitly incorporating mismatch strain
- Author
-
Zhou, Min, Wu, Hong-Hui, Wu, Yuan, Wang, Hui, Liu, Xiong-Jun, Jiang, Sui-He, Zhang, Xiao-Bin, and Lu, Zhao-Ping
- Published
- 2024
- Full Text
- View/download PDF
45. Expression of LC3A, LC3B and p62/SQSTM1 autophagy proteins in hepatocellular carcinoma (HCC) tissues and the predicted microRNAs involved in the autophagy-related pathway
- Author
-
Wong, Magdelyn Mei-Theng, Aziz, Norazlin Abdul, Ch’ng, Ewe Seng, Armon, Subasri, Chook, Jack-Bee, Bong, Jan-Jin, Peh, Suat-Cheng, Wu, Yuan Seng, and Teow, Sin-Yeang
- Published
- 2024
- Full Text
- View/download PDF
46. Antiviral Activity of Withanolide A Against Different Infectivity Phases of Dengue Virus Serotype 2 in Vero Cell Line
- Author
-
Al Quwatli, Lujin, Lee, Michelle Felicia, Wu, Yuan Seng, Poh, Chit Laa, Batumalaie, Kalaivani, Ahemad, Nafees, Fuloria, Neeraj Kumar, Fuloria, Shivkanya, Sekar, Mahendran, Subramaniyan, Vetriselvan, Sarke, Moklesur Rahman, and Mac Guad, Rhanye
- Published
- 2024
- Full Text
- View/download PDF
47. Deep Image Semantic Communication Model for Artificial Intelligent Internet of Things
- Author
-
Qian, Li Ping, Zhang, Yi, Lyu, Sikai, Zhu, Huijie, Wu, Yuan, Shen, Xuemin Sherman, and Yang, Xiaoniu
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence - Abstract
With the rapid development of Artificial Intelligent Internet of Things (AIoT), the image data from AIoT devices has been witnessing the explosive increasing. In this paper, a novel deep image semantic communication model is proposed for the efficient image communication in AIoT. Particularly, at the transmitter side, a high-precision image semantic segmentation algorithm is proposed to extract the semantic information of the image to achieve significant compression of the image data. At the receiver side, a semantic image restoration algorithm based on Generative Adversarial Network (GAN) is proposed to convert the semantic image to a real scene image with detailed information. Simulation results demonstrate that the proposed image semantic communication model can improve the image compression ratio and recovery accuracy by 71.93% and 25.07% on average in comparison with WebP and CycleGAN, respectively. More importantly, our demo experiment shows that the proposed model reduces the total delay by 95.26% in the image communication, when comparing with the original image transmission.
- Published
- 2023
48. Filling the Missing: Exploring Generative AI for Enhanced Federated Learning over Heterogeneous Mobile Edge Devices
- Author
-
Li, Peichun, Zhang, Hanwen, Wu, Yuan, Qian, Liping, Yu, Rong, Niyato, Dusit, and Shen, Xuemin
- Subjects
Computer Science - Machine Learning ,Computer Science - Distributed, Parallel, and Cluster Computing - Abstract
Distributed Artificial Intelligence (AI) model training over mobile edge networks encounters significant challenges due to the data and resource heterogeneity of edge devices. The former hampers the convergence rate of the global model, while the latter diminishes the devices' resource utilization efficiency. In this paper, we propose a generative AI-empowered federated learning to address these challenges by leveraging the idea of FIlling the MIssing (FIMI) portion of local data. Specifically, FIMI can be considered as a resource-aware data augmentation method that effectively mitigates the data heterogeneity while ensuring efficient FL training. We first quantify the relationship between the training data amount and the learning performance. We then study the FIMI optimization problem with the objective of minimizing the device-side overall energy consumption subject to required learning performance constraints. The decomposition-based analysis and the cross-entropy searching method are leveraged to derive the solution, where each device is assigned suitable AI-synthesized data and resource utilization policy. Experiment results demonstrate that FIMI can save up to 50% of the device-side energy to achieve the target global test accuracy in comparison with the existing methods. Meanwhile, FIMI can significantly enhance the converged global accuracy under the non-independently-and-identically distribution (non-IID) data., Comment: 13 pages, 5 figures. Submitted to IEEE for possible publication
- Published
- 2023
49. Residual Scheduling: A New Reinforcement Learning Approach to Solving Job Shop Scheduling Problem
- Author
-
Ho, Kuo-Hao, Jheng, Ruei-Yu, Wu, Ji-Han, Chiang, Fan, Chen, Yen-Chi, Wu, Yuan-Yu, and Wu, I-Chen
- Subjects
Computer Science - Artificial Intelligence - Abstract
Job-shop scheduling problem (JSP) is a mathematical optimization problem widely used in industries like manufacturing, and flexible JSP (FJSP) is also a common variant. Since they are NP-hard, it is intractable to find the optimal solution for all cases within reasonable times. Thus, it becomes important to develop efficient heuristics to solve JSP/FJSP. A kind of method of solving scheduling problems is construction heuristics, which constructs scheduling solutions via heuristics. Recently, many methods for construction heuristics leverage deep reinforcement learning (DRL) with graph neural networks (GNN). In this paper, we propose a new approach, named residual scheduling, to solving JSP/FJSP. In this new approach, we remove irrelevant machines and jobs such as those finished, such that the states include the remaining (or relevant) machines and jobs only. Our experiments show that our approach reaches state-of-the-art (SOTA) among all known construction heuristics on most well-known open JSP and FJSP benchmarks. In addition, we also observe that even though our model is trained for scheduling problems of smaller sizes, our method still performs well for scheduling problems of large sizes. Interestingly in our experiments, our approach even reaches zero gap for 49 among 50 JSP instances whose job numbers are more than 150 on 20 machines.
- Published
- 2023
50. Genomic correlation, shared loci, and causal relationship between insomnia and psoriasis: a large-scale genome-wide cross-trait analysis
- Author
-
Wang, Qing, Wu, Yuan, Wang, Xuehua, Zhang, Junhong, Li, Li, Wu, Jingjing, Lu, Yue, and Han, Ling
- Published
- 2024
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.