10,137,701 results on '"Liu, A.-A."'
Search Results
2. The Promises and Pitfalls of Using Language Models to Measure Instruction Quality in Education. EdWorkingPaper No. 24-948
- Author
-
Annenberg Institute for School Reform at Brown University, Paiheng Xu, Jing Liu, Nathan Jones, Julie Cohen, and Wei Ai
- Abstract
Assessing instruction quality is a fundamental component of any improvement efforts in the education system. However, traditional manual assessments are expensive, subjective, and heavily dependent on observers' expertise and idiosyncratic factors, preventing teachers from getting timely and frequent feedback. Different from prior research that focuses on low-inference instructional practices, this paper presents the first study that leverages Natural Language Processing (NLP) techniques to assess multiple high-inference instructional practices in two distinct educational settings: in-person K-12 classrooms and simulated performance tasks for pre-service teachers. This is also the first study that applies NLP to measure a teaching practice that has been demonstrated to be particularly effective for students with special needs. We confront two challenges inherent in NLP-based instructional analysis, including noisy and long input data and highly skewed distributions of human ratings. Our results suggest that pretrained Language Models (PLMs) demonstrate performances comparable to the agreement level of human raters for variables that are more discrete and require lower inference, but their efficacy diminishes with more complex teaching practices. Interestingly, using only teachers' utterances as input yields strong results for student-centered variables, alleviating common concerns over the difficulty of collecting and transcribing high-quality student speech data in in-person teaching settings. Our findings highlight both the potential and the limitations of current NLP techniques in the education domain, opening avenues for further exploration.
- Published
- 2024
3. Ultrafast measurement of field-particle energy transfer during chorus emissions in space
- Author
-
Liu, C. M., Zhao, B. N., Cao, J. B., Pollock, C. J., Russell, C. T., Liu, Y. Y., Xing, X. N., Linqvist, P. A., and Burch, J. L.
- Subjects
Physics - Space Physics - Abstract
Chorus is one of the strongest electromagnetic emissions naturally occurring in space, and can cause hazardous radiations to humans and satellites1-3. Although chorus has attracted extreme interest and been intensively studied for decades4-7, its generation and evolution remain highly debated, due to the complexity of the underlying physics and the limited capacity of previous spacecraft missions7. Chorus has also been believed to be governed by planetary magnetic dipolar fields5,7. Contrary to such conventional expectation, here we report unexpected observations of chorus in the terrestrial neutral sheet where magnetic dipolar effect is absent. Using unprecedentedly high-cadence data from the Magnetospheric Multiscale Mission, we present the first, ultrafast measurements of the wave dispersion relation and electron three-dimensional distributions within the waves, showing smoking-gun evidences for chorus-electron interactions and development of electron holes in the wave phase space. We estimate field-particle energy transfer inside the waves and find that the waves were extracting energy from local thermal electrons, in line with the wave positive growth rate derived from instability analysis. Our observations, opening new pathways for resolving long-standing controversies regarding the chorus emissions, are crucial for understanding nonlinear energy transport ubiquitously observed in space and astrophysical environments., Comment: under review; comments and suggestions are welcomed
- Published
- 2024
4. AEMLO: AutoEncoder-Guided Multi-Label Oversampling
- Author
-
Zhou, Ao, Liu, Bin, Wang, Jin, Sun, Kaiwei, and Liu, Kelin
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence - Abstract
Class imbalance significantly impacts the performance of multi-label classifiers. Oversampling is one of the most popular approaches, as it augments instances associated with less frequent labels to balance the class distribution. Existing oversampling methods generate feature vectors of synthetic samples through replication or linear interpolation and assign labels through neighborhood information. Linear interpolation typically generates new samples between existing data points, which may result in insufficient diversity of synthesized samples and further lead to the overfitting issue. Deep learning-based methods, such as AutoEncoders, have been proposed to generate more diverse and complex synthetic samples, achieving excellent performance on imbalanced binary or multi-class datasets. In this study, we introduce AEMLO, an AutoEncoder-guided Oversampling technique specifically designed for tackling imbalanced multi-label data. AEMLO is built upon two fundamental components. The first is an encoder-decoder architecture that enables the model to encode input data into a low-dimensional feature space, learn its latent representations, and then reconstruct it back to its original dimension, thus applying to the generation of new data. The second is an objective function tailored to optimize the sampling task for multi-label scenarios. We show that AEMLO outperforms the existing state-of-the-art methods with extensive empirical studies.
- Published
- 2024
5. Controllable Financial Market Generation with Diffusion Guided Meta Agent
- Author
-
Huang, Yu-Hao, Xu, Chang, Liu, Yang, Liu, Weiqing, Li, Wu-Jun, and Bian, Jiang
- Subjects
Computer Science - Computational Engineering, Finance, and Science ,Quantitative Finance - Trading and Market Microstructure - Abstract
Order flow modeling stands as the most fundamental and essential financial task, as orders embody the minimal unit within a financial market. However, current approaches often result in unsatisfactory fidelity in generating order flow, and their generation lacks controllability, thereby limiting their application scenario. In this paper, we advocate incorporating controllability into the market generation process, and propose a Diffusion Guided meta Agent(DiGA) model to address the problem. Specifically, we utilize a diffusion model to capture dynamics of market state represented by time-evolving distribution parameters about mid-price return rate and order arrival rate, and define a meta agent with financial economic priors to generate orders from the corresponding distributions. Extensive experimental results demonstrate that our method exhibits outstanding controllability and fidelity in generation. Furthermore, we validate DiGA's effectiveness as generative environment for downstream financial applications.
- Published
- 2024
6. QD-VMR: Query Debiasing with Contextual Understanding Enhancement for Video Moment Retrieval
- Author
-
Gao, Chenghua, Li, Min, Liu, Jianshuo, Ren, Junxing, Chen, Lin, Liu, Haoyu, Meng, Bo, Fu, Jitao, and Su, Wenwen
- Subjects
Computer Science - Artificial Intelligence - Abstract
Video Moment Retrieval (VMR) aims to retrieve relevant moments of an untrimmed video corresponding to the query. While cross-modal interaction approaches have shown progress in filtering out query-irrelevant information in videos, they assume the precise alignment between the query semantics and the corresponding video moments, potentially overlooking the misunderstanding of the natural language semantics. To address this challenge, we propose a novel model called \textit{QD-VMR}, a query debiasing model with enhanced contextual understanding. Firstly, we leverage a Global Partial Aligner module via video clip and query features alignment and video-query contrastive learning to enhance the cross-modal understanding capabilities of the model. Subsequently, we employ a Query Debiasing Module to obtain debiased query features efficiently, and a Visual Enhancement module to refine the video features related to the query. Finally, we adopt the DETR structure to predict the possible target video moments. Through extensive evaluations of three benchmark datasets, QD-VMR achieves state-of-the-art performance, proving its potential to improve the accuracy of VMR. Further analytical experiments demonstrate the effectiveness of our proposed module. Our code will be released to facilitate future research., Comment: 9 pages, 4 figures, 4 tables
- Published
- 2024
7. The Velocity Aberration Effect of the CSST Main Survey Camera
- Author
-
Feng, Hui-Mei, Cao, Zi-Huang, Lam, Man I, Li, Ran, Tian, Hao, Zhang, Xin, Wei, Peng, Li, Xin-Feng, Wang, Wei, Jones, Hugh R. A., Liu, Mao-Yuan, and Liu, Chao
- Subjects
Astrophysics - Instrumentation and Methods for Astrophysics - Abstract
In this study, we conducted simulations to find the geometric aberrations expected for images taken by the Main Survey Camera (MSC) of the Chinese Space Station Telescope (CSST) due to its motion. As anticipated by previous work, our findings indicate that the geometric distortion of light impacts the focal plane's apparent scale, with a more pronounced influence as the size of the focal plane increases. Our models suggest that the effect consistently influences the pixel scale in both the vertical and parallel directions. The apparent scale variation follows a sinusoidal distribution throughout one orbit period. Simulations reveal that the effect is particularly pronounced in the center of the Galaxy and gradually diminishes along the direction of ecliptic latitude. At low ecliptic latitudes, the total aberration leads to about 0.94 pixels offset (a 20-minute exposure) and 0.26 pixels offset (a 300-second exposure) at the edge of the field of view, respectively. Appropriate processings for the geometric effect during the CSST pre- and post-observation phases are presented., Comment: 13 pages, 8 figures; accepted by RAA
- Published
- 2024
- Full Text
- View/download PDF
8. Memory-Efficient LLM Training with Online Subspace Descent
- Author
-
Liang, Kaizhao, Liu, Bo, Chen, Lizhang, and Liu, Qiang
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Computer Science - Computation and Language - Abstract
Recently, a wide range of memory-efficient LLM training algorithms have gained substantial popularity. These methods leverage the low-rank structure of gradients to project optimizer states into a subspace using projection matrix found by singular value decomposition (SVD). However, convergence of these algorithms is highly dependent on the update rules of their projection matrix. In this work, we provide the \emph{first} convergence guarantee for arbitrary update rules of projection matrix. This guarantee is generally applicable to optimizers that can be analyzed with Hamiltonian Descent, including most common ones, such as LION, Adam. Inspired by our theoretical understanding, we propose Online Subspace Descent, a new family of subspace descent optimizer without SVD. Instead of updating the projection matrix with eigenvectors, Online Subspace Descent updates the projection matrix with online PCA. Online Subspace Descent is flexible and introduces only minimum overhead to training. We show that for the task of pretraining LLaMA models ranging from 60M to 7B parameters on the C4 dataset, Online Subspace Descent achieves lower perplexity and better downstream tasks performance than state-of-the-art low-rank training methods across different settings and narrows the gap with full-rank baselines., Comment: Code is available at https://github.com/kyleliang919/Online-Subspace-Descent
- Published
- 2024
9. Exploring the Role of Audio in Multimodal Misinformation Detection
- Author
-
Liu, Moyang, Liu, Yukun, Fu, Ruibo, Wen, Zhengqi, Tao, Jianhua, Liu, Xuefei, and Li, Guanjun
- Subjects
Computer Science - Multimedia - Abstract
With the rapid development of deepfake technology, especially the deep audio fake technology, misinformation detection on the social media scene meets a great challenge. Social media data often contains multimodal information which includes audio, video, text, and images. However, existing multimodal misinformation detection methods tend to focus only on some of these modalities, failing to comprehensively address information from all modalities. To comprehensively address the various modal information that may appear on social media, this paper constructs a comprehensive multimodal misinformation detection framework. By employing corresponding neural network encoders for each modality, the framework can fuse different modality information and support the multimodal misinformation detection task. Based on the constructed framework, this paper explores the importance of the audio modality in multimodal misinformation detection tasks on social media. By adjusting the architecture of the acoustic encoder, the effectiveness of different acoustic feature encoders in the multimodal misinformation detection tasks is investigated. Furthermore, this paper discovers that audio and video information must be carefully aligned, otherwise the misalignment across different audio and video modalities can severely impair the model performance.
- Published
- 2024
10. Commensurate and Incommensurate Chern Insulators in Magic-angle Bilayer Graphene
- Author
-
Zhang, Zaizhe, Yang, Jingxin, Xie, Bo, Feng, Zuo, Zhang, Shu, Watanabe, Kenji, Taniguchi, Takashi, Yang, Xiaoxia, Dai, Qing, Liu, Tao, Liu, Donghua, Liu, Kaihui, Song, Zhida, Liu, Jianpeng, and Lu, Xiaobo
- Subjects
Condensed Matter - Mesoscale and Nanoscale Physics ,Condensed Matter - Strongly Correlated Electrons - Abstract
The interplay between strong electron-electron interaction and symmetry breaking can have profound influence on the topological properties of materials. In magic angle twisted bilayer graphene (MATBG), the flat band with a single SU(4) flavor associated with the spin and valley degrees of freedom gains non-zero Chern number when C2z symmetry or C2zT symmetry is broken. Electron-electron interaction can further lift the SU(4) degeneracy, leading to the Chern insulator states. Here we report a complete sequence of zero-field Chern insulators at all odd integer fillings (v = +-1, +-3) with different chirality (C = 1 or -1) in hBN aligned MATBG which structurally breaks C2z symmetry. The Chern states at hole fillings (v = -1, -3), which are firstly observed in this work, host an opposite chirality compared with the electron filling scenario. By slightly doping the v = +-3 states, we have observed new correlated insulating states at incommensurate moir\'e fillings which is highly suggested to be intrinsic Wigner crystals according to our theoretical calculations. Remarkably, we have observed prominent Streda-formula violation around v = -3 state. By doping the Chern gap at v = -3 with notable number of electrons at finite magnetic field, the Hall resistance Ryx robustly quantizes to ~ h/e2 whereas longitudinal resistance Rxx vanishes, indicating that the chemical potential is pinned within a Chern gap, forming an incommensurate Chern insulator. By providing the first experimental observation of zero-field Chern insulators in the flat valence band, our work fills up the overall topological framework of MATBG with broken C2z symmetry. Our findings also demonstrate that doped topological flat band is an ideal platform to investigate exotic incommensurate correlated topological states., Comment: 28 pages. 5 figures
- Published
- 2024
11. Large Language Models as Foundations for Next-Gen Dense Retrieval: A Comprehensive Empirical Assessment
- Author
-
Luo, Kun, Qin, Minghao, Liu, Zheng, Xiao, Shitao, Zhao, Jun, and Liu, Kang
- Subjects
Computer Science - Computation and Language - Abstract
Pretrained language models like BERT and T5 serve as crucial backbone encoders for dense retrieval. However, these models often exhibit limited generalization capabilities and face challenges in improving in domain accuracy. Recent research has explored using large language models (LLMs) as retrievers, achieving SOTA performance across various tasks. Despite these advancements, the specific benefits of LLMs over traditional retrievers and the impact of different LLM configurations, such as parameter sizes, pretraining duration, and alignment processes on retrieval tasks remain unclear. In this work, we conduct a comprehensive empirical study on a wide range of retrieval tasks, including in domain accuracy, data efficiency, zero shot generalization, lengthy retrieval, instruction based retrieval, and multi task learning. We evaluate over 15 different backbone LLMs and non LLMs. Our findings reveal that larger models and extensive pretraining consistently enhance in domain accuracy and data efficiency. Additionally, larger models demonstrate significant potential in zero shot generalization, lengthy retrieval, instruction based retrieval, and multi task learning. These results underscore the advantages of LLMs as versatile and effective backbone encoders in dense retrieval, providing valuable insights for future research and development in this field., Comment: Submitted to EMNLP24
- Published
- 2024
12. Recent Advances on Machine Learning for Computational Fluid Dynamics: A Survey
- Author
-
Wang, Haixin, Cao, Yadi, Huang, Zijie, Liu, Yuxuan, Hu, Peiyan, Luo, Xiao, Song, Zezheng, Zhao, Wanjia, Liu, Jilin, Sun, Jinan, Zhang, Shikun, Wei, Long, Wang, Yue, Wu, Tailin, Ma, Zhi-Ming, and Sun, Yizhou
- Subjects
Computer Science - Machine Learning - Abstract
This paper explores the recent advancements in enhancing Computational Fluid Dynamics (CFD) tasks through Machine Learning (ML) techniques. We begin by introducing fundamental concepts, traditional methods, and benchmark datasets, then examine the various roles ML plays in improving CFD. The literature systematically reviews papers in recent five years and introduces a novel classification for forward modeling: Data-driven Surrogates, Physics-Informed Surrogates, and ML-assisted Numerical Solutions. Furthermore, we also review the latest ML methods in inverse design and control, offering a novel classification and providing an in-depth discussion. Then we highlight real-world applications of ML for CFD in critical scientific and engineering disciplines, including aerodynamics, combustion, atmosphere & ocean science, biology fluid, plasma, symbolic regression, and reduced order modeling. Besides, we identify key challenges and advocate for future research directions to address these challenges, such as multi-scale representation, physical knowledge encoding, scientific foundation model and automatic scientific discovery. This review serves as a guide for the rapidly expanding ML for CFD community, aiming to inspire insights for future advancements. We draw the conclusion that ML is poised to significantly transform CFD research by enhancing simulation accuracy, reducing computational time, and enabling more complex analyses of fluid dynamics. The paper resources can be viewed at https://github.com/WillDreamer/Awesome-AI4CFD., Comment: 22 pages, 6 figures
- Published
- 2024
13. ReorderBench: A Benchmark for Matrix Reordering
- Author
-
Zhu, Jiangning, Wang, Zheng, Shen, Zhiyang, Wei, Lai, Tian, Fengyuan, Liu, Mengchen, and Liu, Shixia
- Subjects
Computer Science - Human-Computer Interaction - Abstract
Matrix reordering permutes the rows and columns of a matrix to reveal meaningful visual patterns, such as blocks that represent clusters. A comprehensive collection of matrices, along with a scoring method for measuring the quality of visual patterns in these matrices, contributes to building a benchmark. This benchmark is essential for selecting or designing suitable reordering algorithms for specific tasks. In this paper, we build a matrix reordering benchmark, ReorderBench, with the goal of evaluating and improving matrix reordering techniques. This is achieved by generating a large set of representative and diverse matrices and scoring these matrices with a convolution- and entropy-based method. Our benchmark contains 2,835,000 binary matrices and 5,670,000 continuous matrices, each featuring one of four visual patterns: block, off-diagonal block, star, or band. We demonstrate the usefulness of ReorderBench through three main applications in matrix reordering: 1) evaluating different reordering algorithms, 2) creating a unified scoring model to measure the visual patterns in any matrix, and 3) developing a deep learning model for matrix reordering., Comment: Submitted to IEEE TVCG
- Published
- 2024
14. DimeRec: A Unified Framework for Enhanced Sequential Recommendation via Generative Diffusion Models
- Author
-
Li, Wuchao, Huang, Rui, Zhao, Haijun, Liu, Chi, Zheng, Kai, Liu, Qi, Mou, Na, Zhou, Guorui, Lian, Defu, Song, Yang, Bao, Wentian, Yu, Enyun, and Ou, Wenwu
- Subjects
Computer Science - Information Retrieval ,Computer Science - Machine Learning - Abstract
Sequential Recommendation (SR) plays a pivotal role in recommender systems by tailoring recommendations to user preferences based on their non-stationary historical interactions. Achieving high-quality performance in SR requires attention to both item representation and diversity. However, designing an SR method that simultaneously optimizes these merits remains a long-standing challenge. In this study, we address this issue by integrating recent generative Diffusion Models (DM) into SR. DM has demonstrated utility in representation learning and diverse image generation. Nevertheless, a straightforward combination of SR and DM leads to sub-optimal performance due to discrepancies in learning objectives (recommendation vs. noise reconstruction) and the respective learning spaces (non-stationary vs. stationary). To overcome this, we propose a novel framework called DimeRec (\textbf{Di}ffusion with \textbf{m}ulti-interest \textbf{e}nhanced \textbf{Rec}ommender). DimeRec synergistically combines a guidance extraction module (GEM) and a generative diffusion aggregation module (DAM). The GEM extracts crucial stationary guidance signals from the user's non-stationary interaction history, while the DAM employs a generative diffusion process conditioned on GEM's outputs to reconstruct and generate consistent recommendations. Our numerical experiments demonstrate that DimeRec significantly outperforms established baseline methods across three publicly available datasets. Furthermore, we have successfully deployed DimeRec on a large-scale short video recommendation platform, serving hundreds of millions of users. Live A/B testing confirms that our method improves both users' time spent and result diversification.
- Published
- 2024
15. Behavior Pattern Mining-based Multi-Behavior Recommendation
- Author
-
Li, Haojie, Cheng, Zhiyong, Yu, Xu, Liu, Jinhuan, Liu, Guanfeng, and Du, Junwei
- Subjects
Computer Science - Information Retrieval - Abstract
Multi-behavior recommendation systems enhance effectiveness by leveraging auxiliary behaviors (such as page views and favorites) to address the limitations of traditional models that depend solely on sparse target behaviors like purchases. Existing approaches to multi-behavior recommendations typically follow one of two strategies: some derive initial node representations from individual behavior subgraphs before integrating them for a comprehensive profile, while others interpret multi-behavior data as a heterogeneous graph, applying graph neural networks to achieve a unified node representation. However, these methods do not adequately explore the intricate patterns of behavior among users and items. To bridge this gap, we introduce a novel algorithm called Behavior Pattern mining-based Multi-behavior Recommendation (BPMR). Our method extensively investigates the diverse interaction patterns between users and items, utilizing these patterns as features for making recommendations. We employ a Bayesian approach to streamline the recommendation process, effectively circumventing the challenges posed by graph neural network algorithms, such as the inability to accurately capture user preferences due to over-smoothing. Our experimental evaluation on three real-world datasets demonstrates that BPMR significantly outperforms existing state-of-the-art algorithms, showing an average improvement of 268.29% in Recall@10 and 248.02% in NDCG@10 metrics. The code of our BPMR is openly accessible for use and further research at https://github.com/rookitkitlee/BPMR.
- Published
- 2024
16. Examining the Commitments and Difficulties Inherent in Multimodal Foundation Models for Street View Imagery
- Author
-
Yang, Zhenyuan, Lin, Xuhui, He, Qinyi, Huang, Ziye, Liu, Zhengliang, Jiang, Hanqi, Shu, Peng, Wu, Zihao, Li, Yiwei, Law, Stephen, Mai, Gengchen, Liu, Tianming, and Yang, Tao
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence - Abstract
The emergence of Large Language Models (LLMs) and multimodal foundation models (FMs) has generated heightened interest in their applications that integrate vision and language. This paper investigates the capabilities of ChatGPT-4V and Gemini Pro for Street View Imagery, Built Environment, and Interior by evaluating their performance across various tasks. The assessments include street furniture identification, pedestrian and car counts, and road width measurement in Street View Imagery; building function classification, building age analysis, building height analysis, and building structure classification in the Built Environment; and interior room classification, interior design style analysis, interior furniture counts, and interior length measurement in Interior. The results reveal proficiency in length measurement, style analysis, question answering, and basic image understanding, but highlight limitations in detailed recognition and counting tasks. While zero-shot learning shows potential, performance varies depending on the problem domains and image complexities. This study provides new insights into the strengths and weaknesses of multimodal foundation models for practical challenges in Street View Imagery, Built Environment, and Interior. Overall, the findings demonstrate foundational multimodal intelligence, emphasizing the potential of FMs to drive forward interdisciplinary applications at the intersection of computer vision and language.
- Published
- 2024
17. DUNE Phase II: Scientific Opportunities, Detector Concepts, Technological Solutions
- Author
-
DUNE Collaboration, Abud, A. Abed, Abi, B., Acciarri, R., Acero, M. A., Adames, M. R., Adamov, G., Adamowski, M., Adams, D., Adinolfi, M., Adriano, C., Aduszkiewicz, A., Aguilar, J., Akbar, F., Allison, K., Monsalve, S. Alonso, Alrashed, M., Alton, A., Alvarez, R., Alves, T., Amar, H., Amedo, P., Anderson, J., Andreopoulos, C., Andreotti, M., Andrews, M. P., Andrianala, F., Andringa, S., Anfimov, N., Ankowski, A., Antic, D., Antoniassi, M., Antonova, M., Antoshkin, A., Aranda-Fernandez, A., Arellano, L., Diaz, E. Arrieta, Arroyave, M. A., Asaadi, J., Ashkenazi, A., Asner, D. M., Asquith, L., Atkin, E., Auguste, D., Aurisano, A., Aushev, V., Autiero, D., Azam, M. B., Azfar, F., Back, A., Back, H., Back, J. J., Bagaturia, I., Bagby, L., Balashov, N., Balasubramanian, S., Baldi, P., Baldini, W., Baldonedo, J., Baller, B., Bambah, B., Banerjee, R., Barao, F., Barbu, D., Barenboim, G., Barham~Alzás, P., Barker, G. J., Barkhouse, W., Barr, G., Monarca, J. Barranco, Barros, A., Barros, N., Barrow, D., Barrow, J. L., Basharina-Freshville, A., Bashyal, A., Basque, V., Batchelor, C., Bathe-Peters, L., Battat, J. B. R., Battisti, F., Bay, F., Bazetto, M. C. Q., Alba, J. L. L. Bazo, Beacom, J. F., Bechetoille, E., Behera, B., Belchior, E., Bell, G., Bellantoni, L., Bellettini, G., Bellini, V., Beltramello, O., Benekos, N., Montiel, C. Benitez, Benjamin, D., Neves, F. Bento, Berger, J., Berkman, S., Bernal, J., Bernardini, P., Bersani, A., Bertolucci, S., Betancourt, M., Rodríguez, A. Betancur, Bevan, A., Bezawada, Y., Bezerra, A. T., Bezerra, T. J., Bhat, A., Bhatnagar, V., Bhatt, J., Bhattacharjee, M., Bhattacharya, M., Bhuller, S., Bhuyan, B., Biagi, S., Bian, J., Biery, K., Bilki, B., Bishai, M., Bitadze, A., Blake, A., Blaszczyk, F. D., Blazey, G. C., Blucher, E., Bodek, A., Bogenschuetz, J., Boissevain, J., Bolognesi, S., Bolton, T., Bomben, L., Bonesini, M., Bonilla-Diaz, C., Bonini, F., Booth, A., Boran, F., Bordoni, S., Merlo, R. Borges, Borkum, A., Bostan, N., Bouet, R., Boza, J., Bracinik, J., Brahma, B., Brailsford, D., Bramati, F., Branca, A., Brandt, A., Bremer, J., Brew, C., Brice, S. J., Brio, V., Brizzolari, C., Bromberg, C., Brooke, J., Bross, A., Brunetti, G., Brunetti, M., Buchanan, N., Budd, H., Buergi, J., Bundock, A., Burgardt, D., Butchart, S., V., G. Caceres, Cagnoli, I., Cai, T., Calabrese, R., Calcutt, J., Calivers, L., Calvo, E., Caminata, A., Camino, A. F., Campanelli, W., Campani, A., Benitez, A. Campos, Canci, N., Capó, J., Caracas, I., Caratelli, D., Carber, D., Carceller, J. M., Carini, G., Carlus, B., Carneiro, M. F., Carniti, P., Terrazas, I. Caro, Carranza, H., Carrara, N., Carroll, L., Carroll, T., Carter, A., Casarejos, E., Casazza, D., Forero, J. F. Castaño, Castaño, F. A., Castillo, A., Castromonte, C., Catano-Mur, E., Cattadori, C., Cavalier, F., Cavanna, F., Centro, S., Cerati, G., Cerna, C., Cervelli, A., Villanueva, A. Cervera, Chakraborty, K., Chakraborty, S., Chalifour, M., Chappell, A., Charitonidis, N., Chatterjee, A., Chen, H., Chen, M., Chen, W. C., Chen, Y., Chen-Wishart, Z., Cherdack, D., Chi, C., Chiapponi, F., Chirco, R., Chitirasreemadam, N., Cho, K., Choate, S., Chokheli, D., Chong, P. S., Chowdhury, B., Christian, D., Chukanov, A., Chung, M., Church, E., Cicala, M. F., Cicerchia, M., Cicero, V., Ciolini, R., Clarke, P., Cline, G., Coan, T. E., Cocco, A. G., Coelho, J. A. B., Cohen, A., Collazo, J., Collot, J., Conley, E., Conrad, J. M., Convery, M., Copello, S., Cortez, A. F. V., Cova, P., Cox, C., Cremaldi, L., Cremonesi, L., Crespo-Anadón, J. I., Crisler, M., Cristaldo, E., Crnkovic, J., Crone, G., Cross, R., Cudd, A., Cuesta, C., Cui, Y., Curciarello, F., Cussans, D., Dai, J., Dalager, O., Dallavalle, R., Dallaway, W., D'Amico, R., da Motta, H., Dar, Z. A., Darby, R., Peres, L. Da Silva, David, Q., Davies, G. S., Davini, S., Dawson, J., De Aguiar, R., De Almeida, P., Debbins, P., De Bonis, I., Decowski, M. P., de Gouvêa, A., De Holanda, P. C., Astiz, I. L. De Icaza, De Jong, P., Sanchez, P. Del Amo, De la Torre, A., De Lauretis, G., Delbart, A., Delepine, D., Delgado, M., Dell'Acqua, A., Monache, G. Delle, Delmonte, N., De Lurgio, P., Demario, R., De Matteis, G., Neto, J. R. T. de Mello, DeMuth, D. M., Dennis, S., Densham, C., Denton, P., Deptuch, G. W., De Roeck, A., De Romeri, V., Detje, J. P., Devine, J., Dharmapalan, R., Dias, M., Diaz, A., Díaz, J. S., Díaz, F., Di Capua, F., Di Domenico, A., Di Domizio, S., Di Falco, S., Di Giulio, L., Ding, P., Di Noto, L., Diociaiuti, E., Distefano, C., Diurba, R., Diwan, M., Djurcic, Z., Doering, D., Dolan, S., Dolek, F., Dolinski, M. J., Domenici, D., Domine, L., Donati, S., Donon, Y., Doran, S., Douglas, D., Doyle, T. A., Dragone, A., Drielsma, F., Duarte, L., Duchesneau, D., Duffy, K., Dugas, K., Dunne, P., Dutta, B., Duyang, H., Dwyer, D. A., Dyshkant, A. S., Dytman, S., Eads, M., Earle, A., Edayath, S., Edmunds, D., Eisch, J., Englezos, P., Ereditato, A., Erjavec, T., Escobar, C. O., Evans, J. J., Ewart, E., Ezeribe, A. C., Fahey, K., Fajt, L., Falcone, A., Fani', M., Farnese, C., Farrell, S., Farzan, Y., Fedoseev, D., Felix, J., Feng, Y., Fernandez-Martinez, E., Fernández-Posada, D., Ferry, G., Fialova, E., Fields, L., Filip, P., Filkins, A., Filthaut, F., Fine, R., Fiorillo, G., Fiorini, M., Fogarty, S., Foreman, W., Fowler, J., Franc, J., Francis, K., Franco, D., Franklin, J., Freeman, J., Fried, J., Friedland, A., Fuess, S., Furic, I. K., Furman, K., Furmanski, A. P., Gaba, R., Gabrielli, A., M~Gago, A., Galizzi, F., Gallagher, H., Gallice, N., Galymov, V., Gamberini, E., Gamble, T., Ganacim, F., Gandhi, R., Ganguly, S., Gao, F., Gao, S., Garcia-Gamez, D., García-Peris, M. Á., Gardim, F., Gardiner, S., Gastler, D., Gauch, A., Gauvreau, J., Gauzzi, P., Gazzana, S., Ge, G., Geffroy, N., Gelli, B., Gent, S., Gerlach, L., Ghorbani-Moghaddam, Z., Giammaria, T., Gibin, D., Gil-Botella, I., Gilligan, S., Gioiosa, A., Giovannella, S., Girerd, C., Giri, A. K., Giugliano, C., Giusti, V., Gnani, D., Gogota, O., Gollapinni, S., Gollwitzer, K., Gomes, R. A., Bermeo, L. V. Gomez, Fajardo, L. S. Gomez, Gonnella, F., Gonzalez-Diaz, D., Gonzalez-Lopez, M., Goodman, M. C., Goswami, S., Gotti, C., Goudeau, J., Goudzovski, E., Grace, C., Gramellini, E., Gran, R., Granados, E., Granger, P., Grant, C., Gratieri, D. R., Grauso, G., Green, P., Greenberg, S., Greer, J., Griffith, W. C., Groetschla, F. T., Grzelak, K., Gu, L., Gu, W., Guarino, V., Guarise, M., Guenette, R., Guerzoni, M., Guffanti, D., Guglielmi, A., Guo, B., Guo, F. Y., Gupta, A., Gupta, V., Gurung, G., Gutierrez, D., Guzowski, P., Guzzo, M. M., Gwon, S., Habig, A., Hadavand, H., Haegel, L., Haenni, R., Hagaman, L., Hahn, A., Haiston, J., Hakenmüller, J., Hamernik, T., Hamilton, P., Hancock, J., Happacher, F., Harris, D. A., Hart, A., Hartnell, J., Hartnett, T., Harton, J., Hasegawa, T., Hasnip, C. M., Hatcher, R., Hayrapetyan, K., Hays, J., Hazen, E., He, M., Heavey, A., Heeger, K. M., Heise, J., Hellmuth, P., Henry, S., Hernández-García, J., Herner, K., Hewes, V., Higuera, A., Hilgenberg, C., Hillier, S. J., Himmel, A., Hinkle, E., Hirsch, L. R., Ho, J., Hoff, J., Holin, A., Holvey, T., Hoppe, E., Horiuchi, S., Horton-Smith, G. A., Houdy, T., Howard, B., Howell, R., Hristova, I., Hronek, M. S., Huang, J., Huang, R. G., Hulcher, Z., Ibrahim, M., Iles, G., Ilic, N., Iliescu, A. M., Illingworth, R., Ingratta, G., Ioannisian, A., Irwin, B., Isenhower, L., Oliveira, M. Ismerio, Itay, R., Jackson, C. M., Jain, V., James, E., Jang, W., Jargowsky, B., Jena, D., Jentz, I., Ji, X., Jiang, C., Jiang, J., Jiang, L., Jipa, A., Jo, J. H., Joaquim, F. R., Johnson, W., Jollet, C., Jones, B., Jones, R., Jovancevic, N., Judah, M., Jung, C. K., Junk, T., Jwa, Y., Kabirnezhad, M., Kaboth, A. C., Kadenko, I., Kakorin, I., Kalitkina, A., Kalra, D., Kandemir, M., Kaplan, D. M., Karagiorgi, G., Karaman, G., Karcher, A., Karyotakis, Y., Kasai, S., Kasetti, S. P., Kashur, L., Katsioulas, I., Kauther, A., Kazaryan, N., Ke, L., Kearns, E., Keener, P. T., Kelly, K. J., Kemp, E., Kemularia, O., Kermaidic, Y., Ketchum, W., Kettell, S. H., Khabibullin, M., Khan, N., Khvedelidze, A., Kim, D., Kim, J., Kim, M. J., King, B., Kirby, B., Kirby, M., Kish, A., Klein, J., Kleykamp, J., Klustova, A., Kobilarcik, T., Koch, L., Koehler, K., Koerner, L. W., Koh, D. H., Kolupaeva, L., Korablev, D., Kordosky, M., Kosc, T., Kose, U., Kostelecký, V. A., Kothekar, K., Kotler, I., Kovalcuk, M., Kozhukalov, V., Krah, W., Kralik, R., Kramer, M., Kreczko, L., Krennrich, F., Kreslo, I., Kroupova, T., Kubota, S., Kubu, M., Kudenko, Y., Kudryavtsev, V. A., Kufatty, G., Kuhlmann, S., Kulagin, S., Kumar, J., Kumar, P., Kumaran, S., Kunzmann, J., Kuravi, R., Kurita, N., Kuruppu, C., Kus, V., Kutter, T., Kuźniak, M., Kvasnicka, J., Labree, T., Lackey, T., Lalău, I., Lambert, A., Land, B. J., Lane, C. E., Lane, N., Lang, K., Langford, T., Langstaff, M., Lanni, F., Lantwin, O., Larkin, J., Lasorak, P., Last, D., Laudrain, A., Laundrie, A., Laurenti, G., Lavaut, E., Laycock, P., Lazanu, I., LaZur, R., Lazzaroni, M., Le, T., Leardini, S., Learned, J., LeCompte, T., Legin, V., Miotto, G. Lehmann, Lehnert, R., de Oliveira, M. A. Leigui, Leitner, M., Silverio, D. Leon, Lepin, L. M., -Y~Li, J., Li, S. W., Li, Y., Liao, H., Lin, C. S., Lindebaum, D., Linden, S., Lineros, R. A., Lister, A., Littlejohn, B. R., Liu, H., Liu, J., Liu, Y., Lockwitz, S., Lokajicek, M., Lomidze, I., Long, K., Lopes, T. V., Lopez, J., de Rego, I. López, López-March, N., Lord, T., LoSecco, J. M., Louis, W. C., Sanchez, A. Lozano, Lu, X. -G., Luk, K. B., Lunday, B., Luo, X., Luppi, E., MacFarlane, D., Machado, A. A., Machado, P., Macias, C. T., Macier, J. R., MacMahon, M., Maddalena, A., Madera, A., Madigan, P., Magill, S., Magueur, C., Mahn, K., Maio, A., Major, A., Majumdar, K., Mameli, S., Man, M., Mandujano, R. C., Maneira, J., Manly, S., Mann, A., Manolopoulos, K., Plata, M. Manrique, Corchado, S. Manthey, Manyam, V. N., Marchan, M., Marchionni, A., Marciano, W., Marfatia, D., Mariani, C., Maricic, J., Marinho, F., Marino, A. D., Markiewicz, T., Marques, F. Das Chagas, Marquet, C., Marshak, M., Marshall, C. M., Marshall, J., Martina, L., Martín-Albo, J., Martinez, N., Caicedo, D. A. Martinez, López, F. Martínez, Miravé, P. Martínez, Martynenko, S., Mascagna, V., Massari, C., Mastbaum, A., Matichard, F., Matsuno, S., Matteucci, G., Matthews, J., Mauger, C., Mauri, N., Mavrokoridis, K., Mawby, I., Mazza, R., McAskill, T., McConkey, N., McFarland, K. S., McGrew, C., McNab, A., Meazza, L., Meddage, V. C. N., Mefodiev, A., Mehta, B., Mehta, P., Melas, P., Mena, O., Mendez, H., Mendez, P., Méndez, D. P., Menegolli, A., Meng, G., Mercuri, A. C. E. A., Meregaglia, A., Messier, M. D., Metallo, S., Metcalf, W., Mewes, M., Meyer, H., Miao, T., Micallef, J., Miccoli, A., Michna, G., Milincic, R., Miller, F., Miller, G., Miller, W., Mineev, O., Minotti, A., Miralles, L., Miranda, O. G., Mironov, C., Miryala, S., Miscetti, S., Mishra, C. S., Mishra, P., Mishra, S. R., Mislivec, A., Mitchell, M., Mladenov, D., Mocioiu, I., Mogan, A., Moggi, N., Mohanta, R., Mohayai, T. A., Mokhov, N., Molina, J., Bueno, L. Molina, Montagna, E., Montanari, A., Montanari, C., Montanari, D., Montanino, D., Zetina, L. M. Montaño, Mooney, M., Moor, A. F., Moore, Z., Moreno, D., Moreno-Palacios, O., Morescalchi, L., Moretti, D., Moretti, R., Morris, C., Mossey, C., Moura, C. A., Mouster, G., Mu, W., Mualem, L., Mueller, J., Muether, M., Muheim, F., Muir, A., Mulhearn, M., Munford, D., Munteanu, L. J., Muramatsu, H., Muraz, J., Murphy, M., Murphy, T., Muse, J., Mytilinaki, A., Nachtman, J., Nagai, Y., Nagu, S., Nandakumar, R., Naples, D., Narita, S., Navrer-Agasson, A., Nayak, N., Nebot-Guinot, M., Nehm, A., Nelson, J. K., Neogi, O., Nesbit, J., Nessi, M., Newbold, D., Newcomer, M., Nichol, R., Nicolas-Arnaldos, F., Nikolica, A., Nikolov, J., Niner, E., Nishimura, K., Norman, A., Norrick, A., Novella, P., Nowak, A., Nowak, J. A., Oberling, M., Ochoa-Ricoux, J. P., Oh, S., Oh, S. B., Olivier, A., Olshevskiy, A., Olson, T., Onel, Y., Onishchuk, Y., Oranday, A., Gann, G. D. Orebi, Osbiston, M., Vélez, J. A. Osorio, O'Sullivan, L., Ormachea, L. Otiniano, Ott, J., Pagani, L., Palacio, G., Palamara, O., Palestini, S., Paley, J. M., Pallavicini, M., Palomares, C., Pan, S., Panda, P., Vazquez, W. Panduro, Pantic, E., Paolone, V., Papaleo, R., Papanestis, A., Papoulias, D., Paramesvaran, S., Paris, A., Parke, S., Parozzi, E., Parsa, S., Parsa, Z., Parveen, S., Parvu, M., Pasciuto, D., Pascoli, S., Pasqualini, L., Pasternak, J., Patrick, C., Patrizii, L., Patterson, R. B., Patzak, T., Paudel, A., Paulucci, L., Pavlovic, Z., Pawloski, G., Payne, D., Pec, V., Pedreschi, E., Peeters, S. J. M., Pellico, W., Perez, A. Pena, Pennacchio, E., Penzo, A., Peres, O. L. G., Gonzalez, Y. F. Perez, Pérez-Molina, L., Pernas, C., Perry, J., Pershey, D., Pessina, G., Petrillo, G., Petta, C., Petti, R., Pfaff, M., Pia, V., Pickering, L., Pietropaolo, F., Pimentel, V. L., Pinaroli, G., Pincha, S., Pinchault, J., Pitts, K., Plows, K., Pollack, C., Pollman, T., Pompa, F., Pons, X., Poonthottathil, N., Popov, V., Poppi, F., Porter, J., Paix{ã}o, L. G. Porto, Potekhin, M., Potenza, R., Pozimski, J., Pozzato, M., Prakash, T., Pratt, C., Prest, M., Psihas, F., Pugnere, D., Qian, X., Queen, J., Raaf, J. L., Radeka, V., Rademacker, J., Radics, B., Raffaelli, F., Rafique, A., Raguzin, E., Rai, M., Rajagopalan, S., Rajaoalisoa, M., Rakhno, I., Rakotondravohitra, L., Ralte, L., Delgado, M. A. Ramirez, Ramson, B., Rappoldi, A., Raselli, G., Ratoff, P., Ray, R., Razafinime, H., Rea, E. M., Real, J. S., Rebel, B., Rechenmacher, R., Reichenbacher, J., Reitzner, S. D., Sfar, H. Rejeb, Renner, E., Renshaw, A., Rescia, S., Resnati, F., Diego~Restrepo, Reynolds, C., Ribas, M., Riboldi, S., Riccio, C., Riccobene, G., Ricol, J. S., Rigan, M., Rincón, E. V., Ritchie-Yates, A., Ritter, S., Rivera, D., Rivera, R., Robert, A., Rocha, J. L. Rocabado, Rochester, L., Roda, M., Rodrigues, P., Alonso, M. J. Rodriguez, Rondon, J. Rodriguez, Rosauro-Alcaraz, S., Rosier, P., Ross, D., Rossella, M., Rossi, M., Ross-Lonergan, M., Roy, N., Roy, P., Rubbia, C., Ruggeri, A., Ruiz, G., Russell, B., Ruterbories, D., Rybnikov, A., Sacerdoti, S., Saha, S., Sahoo, S. K., Sahu, N., Sala, P., Samios, N., Samoylov, O., Sanchez, M. C., Bravo, A. Sánchez, Sánchez-Castillo, A., Sanchez-Lucas, P., Sandberg, V., Sanders, D. A., Sanfilippo, S., Sankey, D., Santoro, D., Saoulidou, N., Sapienza, P., Sarasty, C., Sarcevic, I., Sarra, I., Savage, G., Savinov, V., Scanavini, G., Scaramelli, A., Scarff, A., Schefke, T., Schellman, H., Schifano, S., Schlabach, P., Schmitz, D., Schneider, A. W., Scholberg, K., Schukraft, A., Schuld, B., Segade, A., Segreto, E., Selyunin, A., Senadheera, D., Senise, C. R., Sensenig, J., Seo, S. H., Shaevitz, M. H., Shanahan, P., Sharma, P., Kumar, R., Poudel, S. Sharma, Shaw, K., Shaw, T., Shchablo, K., Shen, J., Shepherd-Themistocleous, C., Sheshukov, A., Shi, J., Shi, W., Shin, S., Shivakoti, S., Shoemaker, I., Shooltz, D., Shrock, R., Siddi, B., Siden, M., Silber, J., Simard, L., Sinclair, J., Sinev, G., Singh, J., Singh, L., Singh, P., Singh, V., Chauhan, S. Singh, Sipos, R., Sironneau, C., Sirri, G., Siyeon, K., Skarpaas, K., Smedley, J., Smith, E., Smith, J., Smith, P., Smolik, J., Smy, M., Snape, M., Snider, E. L., Snopok, P., Snowden-Ifft, D., Nunes, M. Soares, Sobel, H., Soderberg, M., Sokolov, S., Salinas, C. J. Solano, Söldner-Rembold, S., Solomey, N., Solovov, V., Sondheim, W. E., Sorel, M., Sotnikov, A., Soto-Oton, J., Sousa, A., Soustruznik, K., Spinella, F., Spitz, J., Spooner, N. J. C., Spurgeon, K., Stalder, D., Stancari, M., Stanco, L., Steenis, J., Stein, R., Steiner, H. M., Lisbôa, A. F. Steklain, Stepanova, A., Stewart, J., Stillwell, B., Stock, J., Stocker, F., Stokes, T., Strait, M., Strauss, T., Strigari, L., Stuart, A., Suarez, J. G., Subash, J., Surdo, A., Suter, L., Sutera, C. M., Sutton, K., Suvorov, Y., Svoboda, R., Swain, S. K., Szczerbinska, B., Szelc, A. M., Sztuc, A., Taffara, A., Talukdar, N., Tamara, J., Tanaka, H. A., Tang, S., Taniuchi, N., Casanova, A. M. Tapia, Oregui, B. Tapia, Tapper, A., Tariq, S., Tarpara, E., Tatar, E., Tayloe, R., Tedeschi, D., Teklu, A. M., Vidal, J. Tena, Tennessen, P., Tenti, M., Terao, K., Terranova, F., Testera, G., Thakore, T., Thea, A., Thomas, S., Thompson, A., Thorn, C., Timm, S. C., Tiras, E., Tishchenko, V., Todorović, N., Tomassetti, L., Tonazzo, A., Torbunov, D., Torti, M., Tortola, M., Tortorici, F., Tosi, N., Totani, D., Toups, M., Touramanis, C., Tran, D., Travaglini, R., Trevor, J., Triller, E., Trilov, S., Truchon, J., Truncali, D., Trzaska, W. H., Tsai, Y., Tsai, Y. -T., Tsamalaidze, Z., Tsang, K. V., Tsverava, N., Tu, S. Z., Tufanli, S., Tunnell, C., Turnberg, S., Turner, J., Tuzi, M., Tyler, J., Tyley, E., Tzanov, M., Uchida, M. A., González, J. Ureña, Urheim, J., Usher, T., Utaegbulam, H., Uzunyan, S., Vagins, M. R., Vahle, P., Valder, S., Valdiviesso, G. A., Valencia, E., Valentim, R., Vallari, Z., Vallazza, E., Valle, J. W. F., Van Berg, R., Van de Water, R. G., Forero, D. V., Vannozzi, A., Van Nuland-Troost, M., Varanini, F., Oliva, D. Vargas, Vasina, S., Vaughan, N., Vaziri, K., Vázquez-Ramos, A., Vega, J., Ventura, S., Verdugo, A., Vergani, S., Verzocchi, M., Vetter, K., Vicenzi, M., de Souza, H. Vieira, Vignoli, C., Vilela, C., Villa, E., Viola, S., Viren, B., Hernandez, A. P. Vizcaya, Vuong, Q., Waldron, A. V., Wallbank, M., Walsh, J., Walton, T., Wang, H., Wang, J., Wang, L., Wang, M. H. L. S., Wang, X., Wang, Y., Warburton, K., Warner, D., Warsame, L., Wascko, M. O., Waters, D., Watson, A., Wawrowska, K., Weber, A., Weber, C. M., Weber, M., Wei, H., Weinstein, A., Westerdale, S., Wetstein, M., Whalen, K., White, A., Whitehead, L. H., Whittington, D., Wilhlemi, J., Wilking, M. J., Wilkinson, A., Wilkinson, C., Wilson, F., Wilson, R. J., Winter, P., Wisniewski, W., Wolcott, J., Wolfs, J., Wongjirad, T., Wood, A., Wood, K., Worcester, E., Worcester, M., Wospakrik, M., Wresilo, K., Wret, C., Wu, S., Wu, W., Wurm, M., Wyenberg, J., Xiao, Y., Xiotidis, I., Yaeggy, B., Yahlali, N., Yandel, E., Yang, J., Yang, K., Yang, T., Yankelevich, A., Yershov, N., Yonehara, K., Young, T., Yu, B., Yu, H., Yu, J., Yu, Y., Yuan, W., Zaki, R., Zalesak, J., Zambelli, L., Zamorano, B., Zani, A., Zapata, O., Zazueta, L., Zeller, G. P., Zennamo, J., Zeug, K., Zhang, C., Zhang, S., Zhao, M., Zhivun, E., Zimmerman, E. D., Zucchelli, S., Zuklin, J., Zutshi, V., and Zwaska, R.
- Subjects
Physics - Instrumentation and Detectors ,High Energy Physics - Experiment - Abstract
The international collaboration designing and constructing the Deep Underground Neutrino Experiment (DUNE) at the Long-Baseline Neutrino Facility (LBNF) has developed a two-phase strategy toward the implementation of this leading-edge, large-scale science project. The 2023 report of the US Particle Physics Project Prioritization Panel (P5) reaffirmed this vision and strongly endorsed DUNE Phase I and Phase II, as did the European Strategy for Particle Physics. While the construction of the DUNE Phase I is well underway, this White Paper focuses on DUNE Phase II planning. DUNE Phase-II consists of a third and fourth far detector (FD) module, an upgraded near detector complex, and an enhanced 2.1 MW beam. The fourth FD module is conceived as a "Module of Opportunity", aimed at expanding the physics opportunities, in addition to supporting the core DUNE science program, with more advanced technologies. This document highlights the increased science opportunities offered by the DUNE Phase II near and far detectors, including long-baseline neutrino oscillation physics, neutrino astrophysics, and physics beyond the standard model. It describes the DUNE Phase II near and far detector technologies and detector design concepts that are currently under consideration. A summary of key R&D goals and prototyping phases needed to realize the Phase II detector technical designs is also provided. DUNE's Phase II detectors, along with the increased beam power, will complete the full scope of DUNE, enabling a multi-decadal program of groundbreaking science with neutrinos.
- Published
- 2024
18. SAM-REF: Rethinking Image-Prompt Synergy for Refinement in Segment Anything
- Author
-
Yu, Chongkai, Li, Anqi, Qu, Xiaochao, Liu, Luoqi, and Liu, Ting
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
The advent of the Segment Anything Model (SAM) marks a significant milestone for interactive segmentation using generalist models. As a late fusion model, SAM extracts image embeddings once and merges them with prompts in later interactions. This strategy limits the models ability to extract detailed information from the prompted target zone. Current specialist models utilize the early fusion strategy that encodes the combination of images and prompts to target the prompted objects, yet repetitive complex computations on the images result in high latency. The key to these issues is efficiently synergizing the images and prompts. We propose SAM-REF, a two-stage refinement framework that fully integrates images and prompts globally and locally while maintaining the accuracy of early fusion and the efficiency of late fusion. The first-stage GlobalDiff Refiner is a lightweight early fusion network that combines the whole image and prompts, focusing on capturing detailed information for the entire object. The second-stage PatchDiff Refiner locates the object detail window according to the mask and prompts, then refines the local details of the object. Experimentally, we demonstrated the high effectiveness and efficiency of our method in tackling complex cases with multiple interactions. Our SAM-REF model outperforms the current state-of-the-art method in most metrics on segmentation quality without compromising efficiency.
- Published
- 2024
19. Bidirectional Gated Mamba for Sequential Recommendation
- Author
-
Liu, Ziwei, Liu, Qidong, Wang, Yejing, Wang, Wanyu, Jia, Pengyue, Wang, Maolin, Liu, Zitao, Chang, Yi, and Zhao, Xiangyu
- Subjects
Computer Science - Artificial Intelligence - Abstract
In various domains, Sequential Recommender Systems (SRS) have become essential due to their superior capability to discern intricate user preferences. Typically, SRS utilize transformer-based architectures to forecast the subsequent item within a sequence. Nevertheless, the quadratic computational complexity inherent in these models often leads to inefficiencies, hindering the achievement of real-time recommendations. Mamba, a recent advancement, has exhibited exceptional performance in time series prediction, significantly enhancing both efficiency and accuracy. However, integrating Mamba directly into SRS poses several challenges. Its inherently unidirectional nature may constrain the model's capacity to capture the full context of user-item interactions, while its instability in state estimation can compromise its ability to detect short-term patterns within interaction sequences. To overcome these issues, we introduce a new framework named Selective Gated Mamba (SIGMA) for Sequential Recommendation. This framework leverages a Partially Flipped Mamba (PF-Mamba) to construct a bidirectional architecture specifically tailored to improve contextual modeling. Additionally, an input-sensitive Dense Selective Gate (DS Gate) is employed to optimize directional weights and enhance the processing of sequential information in PF-Mamba. For short sequence modeling, we have also developed a Feature Extract GRU (FE-GRU) to efficiently capture short-term dependencies. Empirical results indicate that SIGMA outperforms current models on five real-world datasets. Our implementation code is available at https://github.com/ziwliu-cityu/SIMGA to ease reproducibility.
- Published
- 2024
20. EAGLE: Elevating Geometric Reasoning through LLM-empowered Visual Instruction Tuning
- Author
-
Li, Zhihao, Du, Yao, Liu, Yang, Zhang, Yan, Liu, Yufang, Zhang, Mengdi, and Cai, Xunliang
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Multi-modal Large Language Models have recently experienced rapid developments and excel in various multi-modal tasks. However, they still struggle with mathematical geometric problem solving, which requires exceptional visual perception proficiency. Existing MLLMs mostly optimize the LLM backbone to acquire geometric reasoning capabilities, while rarely emphasizing improvements in visual comprehension. In this paper, we first investigate the visual perception performance of MLLMs when facing geometric diagrams. Our findings reveal that current MLLMs severely suffer from inaccurate geometric perception and hallucinations. To address these limitations, we propose EAGLE, a novel two-stage end-to-end visual enhancement MLLM framework designed to ElevAte Geometric reasoning through LLM-Empowered visual instruction tuning. Specifically, in the preliminary stage, we feed geometric image-caption pairs into our MLLM that contains a fully fine-tuning CLIP ViT and a frozen LLM, aiming to endow our model with basic geometric knowledge. In the subsequent advanced stage, we incorporate LoRA modules into the vision encoder and unfreeze the LLM backbone. This enables the model to leverage the inherent CoT rationales within question-answer pairs, guiding the MLLM to focus on nuanced visual cues and enhancing its overall perceptual capacity. Moreover, we optimize the cross-modal projector in both stages to foster adaptive visual-linguistic alignments. After the two-stage visual enhancement, we develop the geometry expert model EAGLE-7B. Extensive experiments on popular benchmarks demonstrate the effectiveness of our model. For example, on the GeoQA benchmark, EAGLE-7B not only surpasses the exemplary G-LLaVA 7B model by 2.9%, but also marginally outperforms the larger G-LLaVA 13B model. On the MathVista benchmark, EAGLE-7B achieves remarkable 3.8% improvements compared with the proprietary model GPT-4V.
- Published
- 2024
21. Automatic Dataset Construction (ADC): Sample Collection, Data Curation, and Beyond
- Author
-
Liu, Minghao, Di, Zonglin, Wei, Jiaheng, Wang, Zhongruo, Zhang, Hengxiang, Xiao, Ruixuan, Wang, Haoyu, Pang, Jinlong, Chen, Hao, Shah, Ankit, Wei, Hongxin, He, Xinlei, Zhao, Zhaowei, Wang, Haobo, Feng, Lei, Wang, Jindong, Davis, James, and Liu, Yang
- Subjects
Computer Science - Artificial Intelligence ,Computer Science - Machine Learning - Abstract
Large-scale data collection is essential for developing personalized training data, mitigating the shortage of training data, and fine-tuning specialized models. However, creating high-quality datasets quickly and accurately remains a challenge due to annotation errors, the substantial time and costs associated with human labor. To address these issues, we propose Automatic Dataset Construction (ADC), an innovative methodology that automates dataset creation with negligible cost and high efficiency. Taking the image classification task as a starting point, ADC leverages LLMs for the detailed class design and code generation to collect relevant samples via search engines, significantly reducing the need for manual annotation and speeding up the data generation process. Despite these advantages, ADC also encounters real-world challenges such as label errors (label noise) and imbalanced data distributions (label bias). We provide open-source software that incorporates existing methods for label error detection, robust learning under noisy and biased data, ensuring a higher-quality training data and more robust model training procedure. Furthermore, we design three benchmark datasets focused on label noise detection, label noise learning, and class-imbalanced learning. These datasets are vital because there are few existing datasets specifically for label noise detection, despite its importance. Finally, we evaluate the performance of existing popular methods on these datasets, thereby facilitating further research in the field.
- Published
- 2024
22. Minute-Cadence Observations of the LAMOST Fields with the TMTS: IV -- Catalog of Cataclysmic Variables from the First 3-yr Survey
- Author
-
Liu, Qichun, Lin, Jie, Wang, Xiaofeng, Dai, Zhibin, Sun, Yongkang, Xi, Gaobo, Mo, Jun, Liu, Jialian, Yan, Shengyu, Filippenko, Alexei V., Brink, Thomas G., Yang, Yi, Patra, Kishore C., Cai, Yongzhi, Chen, Zhihao, Chen, Liyang, Guo, Fangzhou, Jiang, Xiaojun, Li, Gaici, Li, Wenxiong, Lin, Weili, Miao, Cheng, Ma, Xiaoran, Peng, Haowei, Xia, Qiqi, Xiang, Danfeng, and Zhang, Jicheng
- Subjects
Astrophysics - Solar and Stellar Astrophysics - Abstract
The Tsinghua University--Ma Huateng Telescopes for Survey (TMTS) started to monitor the LAMOST plates in 2020, leading to the discovery of numerous short-period eclipsing binaries, peculiar pulsators, flare stars, and other variable objects. Here, we present the uninterrupted light curves for a sample of 64 cataclysmic variables (CVs) observed/discovered using the TMTS during its first three-year observations, and we introduce new CVs and new light-variation periods (from known CVs) revealed through the TMTS observations. Thanks to the high-cadence observations of TMTS, diverse light variations, including superhumps, quasi-periodic oscillations, large-amplitude orbital modulations, and rotational modulations, are able to be detected in our CV samples, providing key observational clues for understanding the fast-developing physical processes in various CVs. All of these short-timescale light-curve features help further classify the subtypes of CV systems. We highlight the light-curve features observed in our CV sample and discuss further implications of minute-cadence light curves for CV identifications and classifications. Moreover, we examine the H$\alpha$ emission lines in the spectra from our nonmagnetic CV samples (i.e., dwarf novae and nova-like subclasses) and find that the distribution of H$\alpha$ emission strength shows significant differences between the sources with orbital periods above and below the period gap, which agrees with the trend seen from the SDSS nonmagnetic CV sample., Comment: 27 pages, 12 figures in main text, accepted for the publication in Universe
- Published
- 2024
23. uMedSum: A Unified Framework for Advancing Medical Abstractive Summarization
- Author
-
Nagar, Aishik, Liu, Yutong, Liu, Andy T., Schlegel, Viktor, Dwivedi, Vijay Prakash, Kaliya-Perumal, Arun-Kumar, Kalanchiam, Guna Pratheep, Tang, Yili, and Tan, Robby T.
- Subjects
Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computer Science - Machine Learning - Abstract
Medical abstractive summarization faces the challenge of balancing faithfulness and informativeness. Current methods often sacrifice key information for faithfulness or introduce confabulations when prioritizing informativeness. While recent advancements in techniques like in-context learning (ICL) and fine-tuning have improved medical summarization, they often overlook crucial aspects such as faithfulness and informativeness without considering advanced methods like model reasoning and self-improvement. Moreover, the field lacks a unified benchmark, hindering systematic evaluation due to varied metrics and datasets. This paper addresses these gaps by presenting a comprehensive benchmark of six advanced abstractive summarization methods across three diverse datasets using five standardized metrics. Building on these findings, we propose uMedSum, a modular hybrid summarization framework that introduces novel approaches for sequential confabulation removal followed by key missing information addition, ensuring both faithfulness and informativeness. Our work improves upon previous GPT-4-based state-of-the-art (SOTA) medical summarization methods, significantly outperforming them in both quantitative metrics and qualitative domain expert evaluations. Notably, we achieve an average relative performance improvement of 11.8% in reference-free metrics over the previous SOTA. Doctors prefer uMedSum's summaries 6 times more than previous SOTA in difficult cases where there are chances of confabulations or missing information. These results highlight uMedSum's effectiveness and generalizability across various datasets and metrics, marking a significant advancement in medical summarization., Comment: 12 pages
- Published
- 2024
24. Mental-Perceiver: Audio-Textual Multimodal Learning for Mental Health Assessment
- Author
-
Qin, Jinghui, Liu, Changsong, Tang, Tianchi, Liu, Dahuang, Wang, Minghao, Huang, Qianying, Xu, Yang, and Zhang, Rumin
- Subjects
Computer Science - Computers and Society - Abstract
Mental disorders, such as anxiety and depression, have become a global issue that affects the regular lives of people across different ages. Without proper detection and treatment, anxiety and depression can hinder the sufferer's study, work, and daily life. Fortunately, recent advancements of digital and AI technologies provide new opportunities for better mental health care and many efforts have been made in developing automatic anxiety and depression assessment techniques. However, this field still lacks a publicly available large-scale dataset that can facilitate the development and evaluation of AI-based techniques. To address this limitation, we have constructed a new large-scale \textbf{M}ulti-\textbf{M}odal \textbf{Psy}chological assessment corpus (MMPsy) on anxiety and depression assessment of Mandarin-speaking adolescents. The MMPsy contains audios and extracted transcripts of responses from automated anxiety or depression assessment interviews along with the self-reported anxiety or depression evaluations of the participants using standard mental health assessment questionnaires. Our dataset contains over 7,700 post-processed recordings of interviews for anxiety assessment and over 4,200 recordings for depression assessment. Using this dataset, we have developed a novel deep-learning based mental disorder estimation model, named \textbf{Mental-Perceiver}, to detect anxious/depressive mental states from recorded audio and transcript data. Extensive experiments on our MMPsy and the commonly-used DAIC-WOZ datasets have shown the effectiveness and superiority of our proposed Mental-Perceiver model in anxiety and depression detection. The MMPsy dataset will be made publicly available later to facilitate the research and development of AI-based techniques in the mental health care field.
- Published
- 2024
25. A Deconfounding Approach to Climate Model Bias Correction
- Author
-
Gao, Wentao, Li, Jiuyong, Cheng, Debo, Liu, Lin, Liu, Jixue, Le, Thuc Duy, Du, Xiaojing, Chen, Xiongren, Zhao, Yanchang, and Chen, Yun
- Subjects
Statistics - Machine Learning ,Computer Science - Artificial Intelligence ,Computer Science - Machine Learning ,Physics - Atmospheric and Oceanic Physics - Abstract
Global Climate Models (GCMs) are crucial for predicting future climate changes by simulating the Earth systems. However, GCM outputs exhibit systematic biases due to model uncertainties, parameterization simplifications, and inadequate representation of complex climate phenomena. Traditional bias correction methods, which rely on historical observation data and statistical techniques, often neglect unobserved confounders, leading to biased results. This paper proposes a novel bias correction approach to utilize both GCM and observational data to learn a factor model that captures multi-cause latent confounders. Inspired by recent advances in causality based time series deconfounding, our method first constructs a factor model to learn latent confounders from historical data and then applies them to enhance the bias correction process using advanced time series forecasting models. The experimental results demonstrate significant improvements in the accuracy of precipitation outputs. By addressing unobserved confounders, our approach offers a robust and theoretically grounded solution for climate model bias correction.
- Published
- 2024
26. Efficient Detection of Toxic Prompts in Large Language Models
- Author
-
Liu, Yi, Yu, Junzhe, Sun, Huijia, Shi, Ling, Deng, Gelei, Chen, Yuqi, and Liu, Yang
- Subjects
Computer Science - Cryptography and Security ,Computer Science - Artificial Intelligence ,Computer Science - Computation and Language ,Computer Science - Software Engineering - Abstract
Large language models (LLMs) like ChatGPT and Gemini have significantly advanced natural language processing, enabling various applications such as chatbots and automated content generation. However, these models can be exploited by malicious individuals who craft toxic prompts to elicit harmful or unethical responses. These individuals often employ jailbreaking techniques to bypass safety mechanisms, highlighting the need for robust toxic prompt detection methods. Existing detection techniques, both blackbox and whitebox, face challenges related to the diversity of toxic prompts, scalability, and computational efficiency. In response, we propose ToxicDetector, a lightweight greybox method designed to efficiently detect toxic prompts in LLMs. ToxicDetector leverages LLMs to create toxic concept prompts, uses embedding vectors to form feature vectors, and employs a Multi-Layer Perceptron (MLP) classifier for prompt classification. Our evaluation on various versions of the LLama models, Gemma-2, and multiple datasets demonstrates that ToxicDetector achieves a high accuracy of 96.39\% and a low false positive rate of 2.00\%, outperforming state-of-the-art methods. Additionally, ToxicDetector's processing time of 0.0780 seconds per prompt makes it highly suitable for real-time applications. ToxicDetector achieves high accuracy, efficiency, and scalability, making it a practical method for toxic prompt detection in LLMs., Comment: Accepted by the 39th IEEE/ACM International Conference on Automated Software Engineering (ASE 2024)
- Published
- 2024
27. In situ mixer calibration for superconducting quantum circuits
- Author
-
Wu, Nan, Lin, Jing, Xie, Changrong, Guo, Zechen, Huang, Wenhui, Zhang, Libo, Zhou, Yuxuan, Sun, Xuandong, Zhang, Jiawei, Guo, Weijie, Linpeng, Xiayu, Liu, Song, Liu, Yang, Ren, Wenhui, Tao, Ziyu, Jiang, Ji, Chu, Ji, Niu, Jingjing, Zhong, Youpeng, and Yu, Dapeng
- Subjects
Quantum Physics - Abstract
Mixers play a crucial role in superconducting quantum computing, primarily by facilitating frequency conversion of signals to enable precise control and readout of quantum states. However, imperfections, particularly carrier leakage and unwanted sideband signal, can significantly compromise control fidelity. To mitigate these defects, regular and precise mixer calibrations are indispensable, yet they pose a formidable challenge in large-scale quantum control. Here, we introduce an in situ calibration technique and outcome-focused mixer calibration scheme using superconducting qubits. Our method leverages the qubit's response to imperfect signals, allowing for calibration without modifying the wiring configuration. We experimentally validate the efficacy of this technique by benchmarking single-qubit gate fidelity and qubit coherence time., Comment: 9 pages, 7 figures
- Published
- 2024
28. Optimizing Federated Graph Learning with Inherent Structural Knowledge and Dual-Densely Connected GNNs
- Author
-
Wang, Longwen, Liu, Jianchun, Liu, Zhi, and Huang, Jinyang
- Subjects
Computer Science - Machine Learning - Abstract
Federated Graph Learning (FGL) is an emerging technology that enables clients to collaboratively train powerful Graph Neural Networks (GNNs) in a distributed manner without exposing their private data. Nevertheless, FGL still faces the challenge of the severe non-Independent and Identically Distributed (non-IID) nature of graphs, which possess diverse node and edge structures, especially across varied domains. Thus, exploring the knowledge inherent in these structures becomes significantly crucial. Existing methods, however, either overlook the inherent structural knowledge in graph data or capture it at the cost of significantly increased resource demands (e.g., FLOPs and communication bandwidth), which can be detrimental to distributed paradigms. Inspired by this, we propose FedDense, a novel FGL framework that optimizes the utilization efficiency of inherent structural knowledge. To better acquire knowledge of diverse and underexploited structures, FedDense first explicitly encodes the structural knowledge inherent within graph data itself alongside node features. Besides, FedDense introduces a Dual-Densely Connected (DDC) GNN architecture that exploits the multi-scale (i.e., one-hop to multi-hop) feature and structure insights embedded in the aggregated feature maps at each layer. In addition to the exploitation of inherent structures, we consider resource limitations in FGL, devising exceedingly narrow layers atop the DDC architecture and adopting a selective parameter sharing strategy to reduce resource costs substantially. We conduct extensive experiments using 15 datasets across 4 different domains, demonstrating that FedDense consistently surpasses baselines by a large margin in training performance, while demanding minimal resources.
- Published
- 2024
29. Vision Calorimeter for Anti-neutron Reconstruction: A Baseline
- Author
-
Yu, Hongtian, Li, Yangu, Wu, Mingrui, Shen, Letian, Liu, Yue, Song, Yunxuan, Ye, Qixiang, Lyu, Xiaorui, Mao, Yajun, Zheng, Yangheng, and Liu, Yunfan
- Subjects
High Energy Physics - Experiment ,Computer Science - Computer Vision and Pattern Recognition - Abstract
In high-energy physics, anti-neutrons ($\bar{n}$) are fundamental particles that frequently appear as final-state particles, and the reconstruction of their kinematic properties provides an important probe for understanding the governing principles. However, this confronts significant challenges instrumentally with the electromagnetic calorimeter (EMC), a typical experimental sensor but recovering the information of incident $\bar{n}$ insufficiently. In this study, we introduce Vision Calorimeter (ViC), a baseline method for anti-neutron reconstruction that leverages deep learning detectors to analyze the implicit relationships between EMC responses and incident $\bar{n}$ characteristics. Our motivation lies in that energy distributions of $\bar{n}$ samples deposited in the EMC cell arrays embody rich contextual information. Converted to 2-D images, such contextual energy distributions can be used to predict the status of $\bar{n}$ ($i.e.$, incident position and momentum) through a deep learning detector along with pseudo bounding boxes and a specified training objective. Experimental results demonstrate that ViC substantially outperforms the conventional reconstruction approach, reducing the prediction error of incident position by 42.81% (from 17.31$^{\circ}$ to 9.90$^{\circ}$). More importantly, this study for the first time realizes the measurement of incident $\bar{n}$ momentum, underscoring the potential of deep learning detectors for particle reconstruction. Code is available at https://github.com/yuhongtian17/ViC.
- Published
- 2024
30. Microsatellite-based real-time quantum key distribution
- Author
-
Li, Yang, Cai, Wen-Qi, Ren, Ji-Gang, Wang, Chao-Ze, Yang, Meng, Zhang, Liang, Wu, Hui-Ying, Chang, Liang, Wu, Jin-Cai, Jin, Biao, Xue, Hua-Jian, Li, Xue-Jiao, Liu, Hui, Yu, Guang-Wen, Tao, Xue-Ying, Chen, Ting, Liu, Chong-Fei, Luo, Wen-Bin, Zhou, Jie, Yong, Hai-Lin, Li, Yu-Huai, Li, Feng-Zhi, Jiang, Cong, Chen, Hao-Ze, Wu, Chao, Tong, Xin-Hai, Xie, Si-Jiang, Zhou, Fei, Liu, Wei-Yue, Liu, Nai-Le, Li, Li, Xu, Feihu, Cao, Yuan, Yin, Juan, Shu, Rong, Wang, Xiang-Bin, Zhang, Qiang, Wang, Jian-Yu, Liao, Sheng-Kai, Peng, Cheng-Zhi, and Pan, Jian-Wei
- Subjects
Quantum Physics - Abstract
A quantum network provides an infrastructure connecting quantum devices with revolutionary computing, sensing, and communication capabilities. As the best-known application of a quantum network, quantum key distribution (QKD) shares secure keys guaranteed by the laws of quantum mechanics. A quantum satellite constellation offers a solution to facilitate the quantum network on a global scale. The Micius satellite has verified the feasibility of satellite quantum communications, however, scaling up quantum satellite constellations is challenging, requiring small lightweight satellites, portable ground stations and real-time secure key exchange. Here we tackle these challenges and report the development of a quantum microsatellite capable of performing space-to-ground QKD using portable ground stations. The quantum microsatellite features a payload weighing approximately 23 kg, while the portable ground station weighs about 100 kg. These weights represent reductions by more than an order and two orders of magnitude, respectively, compared to the Micius satellite. Additionally, we multiplex bidirectional satellite-ground optical communication with quantum communication, enabling key distillation and secure communication in real-time. Using the microsatellite and the portable ground stations, we demonstrate satellite-based QKD with multiple ground stations and achieve the sharing of up to 0.59 million bits of secure keys during a single satellite pass. The compact quantum payload can be readily assembled on existing space stations or small satellites, paving the way for a satellite-constellation-based quantum and classical network for widespread real-life applications., Comment: 40 pages, 8 figures
- Published
- 2024
31. Open-FinLLMs: Open Multimodal Large Language Models for Financial Applications
- Author
-
Xie, Qianqian, Li, Dong, Xiao, Mengxi, Jiang, Zihao, Xiang, Ruoyu, Zhang, Xiao, Chen, Zhengyu, He, Yueru, Han, Weiguang, Yang, Yuzhe, Chen, Shunian, Zhang, Yifei, Shen, Lihang, Kim, Daniel, Liu, Zhiwei, Luo, Zheheng, Yu, Yangyang, Cao, Yupeng, Deng, Zhiyang, Yao, Zhiyuan, Li, Haohang, Feng, Duanyu, Dai, Yongfu, Somasundaram, VijayaSai, Lu, Peng, Zhao, Yilun, Long, Yitao, Xiong, Guojun, Smith, Kaleb, Yu, Honghai, Lai, Yanzhao, Peng, Min, Nie, Jianyun, Suchow, Jordan W., Liu, Xiao-Yang, Wang, Benyou, Lopez-Lira, Alejandro, Huang, Jimin, and Ananiadou, Sophia
- Subjects
Computer Science - Computation and Language ,Computer Science - Computational Engineering, Finance, and Science ,Quantitative Finance - Computational Finance - Abstract
Large language models (LLMs) have advanced financial applications, yet they often lack sufficient financial knowledge and struggle with tasks involving multi-modal inputs like tables and time series data. To address these limitations, we introduce \textit{Open-FinLLMs}, a series of Financial LLMs. We begin with FinLLaMA, pre-trained on a 52 billion token financial corpus, incorporating text, tables, and time-series data to embed comprehensive financial knowledge. FinLLaMA is then instruction fine-tuned with 573K financial instructions, resulting in FinLLaMA-instruct, which enhances task performance. Finally, we present FinLLaVA, a multimodal LLM trained with 1.43M image-text instructions to handle complex financial data types. Extensive evaluations demonstrate FinLLaMA's superior performance over LLaMA3-8B, LLaMA3.1-8B, and BloombergGPT in both zero-shot and few-shot settings across 19 and 4 datasets, respectively. FinLLaMA-instruct outperforms GPT-4 and other Financial LLMs on 15 datasets. FinLLaVA excels in understanding tables and charts across 4 multimodal tasks. Additionally, FinLLaMA achieves impressive Sharpe Ratios in trading simulations, highlighting its robust financial application capabilities. We will continually maintain and improve our models and benchmarks to support ongoing innovation in academia and industry., Comment: 33 pages, 13 figures
- Published
- 2024
32. GRANDlib: A simulation pipeline for the Giant Radio Array for Neutrino Detection (GRAND)
- Author
-
GRAND Collaboration, Batista, Rafael Alves, Benoit-Lévy, Aurélien, Bister, Teresa, Bohacova, Martina, Bustamante, Mauricio, Carvalho, Washington, Chen, Yiren, Cheng, LingMei, Chiche, Simon, Colley, Jean-Marc, Correa, Pablo, Laurenciu, Nicoleta Cucu, Dai, Zigao, de Almeida, Rogerio M., de Errico, Beatriz, de Jong, Sijbrand, Neto, João R. T. de Mello, de Vries, Krijn D., Decoene, Valentin, Denton, Peter B., Duan, Bohao, Duan, Kaikai, Engel, Ralph, Erba, William, Fan, Yizhong, Ferrière, Arsène, Gou, QuanBu, Gu, Junhua, Guelfand, Marion, Guo, Jianhua, Guo, Yiqing, Guépin, Claire, Gülzow, Lukas, Haungs, Andreas, Havelka, Matej, He, Haoning, Hivon, Eric, Hu, Hongbo, Huang, Xiaoyuan, Huang, Yan, Huege, Tim, Jiang, Wen, Koirala, Ramesh, Kong, ChuiZheng, Kotera, Kumiko, Köhler, Jelena, Lago, Bruno L., Lai, Zhisen, Coz, Sandra Le, Legrand, François, Leisos, Antonios, Li, Rui, Li, Xingyu, Li, YiFei, Liu, Cheng, Liu, Ruoyu, Liu, Wei, Ma, Pengxiong, Macias, Oscar, Magnard, Frédéric, Marcowith, Alexandre, Martineau-Huynh, Olivier, McKinley, Thomas, Minodier, Paul, Mitra, Pragati, Mostafá, Miguel, Murase, Kohta, Niess, Valentin, Nonis, Stavros, Ogio, Shoichi, Oikonomou, Foteini, Pan, Hongwei, Papageorgiou, Konstantinos, Pierog, Tanguy, Piotrowski, Lech Wiktor, Prunet, Simon, Qian, Xiangli, Roth, Markus, Sako, Takashi, Schoorlemmer, Harm, Szálas-Motesiczky, Dániel, Sławiński, Szymon, Tian, Xishui, Timmermans, Anne, Timmermans, Charles, Tobiska, Petr, Tsirigotis, Apostolos, Tueros, Matías, Vittakis, George, Wang, Hanrui, Wang, Jiale, Wang, Shen, Wang, Xiangyu, Wang, Xu, Wei, Daming, Wei, Feng, Wu, Xiangping, Wu, Xuefeng, Xu, Xin, Xu, Xing, Yang, Fufu, Yang, Lili, Yang, Xuan, Yuan, Qiang, Zarka, Philippe, Zeng, Houdun, Zhang, Chao, Zhang, Jianli, Zhang, Kewen, Zhang, Pengfei, Zhang, Qingchi, Zhang, Songbo, Zhang, Yi, and Zhou, Hao
- Subjects
Astrophysics - Instrumentation and Methods for Astrophysics ,High Energy Physics - Experiment ,High Energy Physics - Phenomenology - Abstract
The operation of upcoming ultra-high-energy cosmic-ray, gamma-ray, and neutrino radio-detection experiments, like the Giant Radio Array for Neutrino Detection (GRAND), poses significant computational challenges involving the production of numerous simulations of particle showers and their detection, and a high data throughput. GRANDlib is an open-source software tool designed to meet these challenges. Its primary goal is to perform end-to-end simulations of the detector operation, from the interaction of ultra-high-energy particles, through -- by interfacing with external air-shower simulations -- the ensuing particle shower development and its radio emission, to its detection by antenna arrays and its processing by data-acquisition systems. Additionally, GRANDlib manages the visualization, storage, and retrieval of experimental and simulated data. We present an overview of GRANDlib to serve as the basis of future GRAND analyses., Comment: 11 pages, 9 figures, plus appendices
- Published
- 2024
33. Revisiting the measurements and interpretations of DLVO forces
- Author
-
Feng, Bo, Liu, Xiantang, Liu, Xinmin, Li, Yingli, and Li, Hang
- Subjects
Physics - Chemical Physics - Abstract
The DLVO theory and electrical double layer (EDL) theory are the foundation of colloid and interface science. With the invention and development of surface forces apparatus (SFA) and atomic force microscope (AFM), the measurements and interpretations of DLVO forces (i.e., mainly measuring the EDL force (electrostatic force) FEDL and van der Waals force FvdW, and interpreting the potential {\psi}, charge density {\sigma}, and Hamaker constant H) can be greatly facilitated by various surface force measurement techniques, and would have been very promising in advancing the DLVO theory, EDL theory, and colloid and interface science. However, although numerous studies have been conducted, pervasive anomalous results can be identified throughout the literature, main including: (1) the fitted {\psi}/{\sigma} is normally extremely small ({\psi} can be close to or (much) smaller than {\psi}{\zeta} (zeta potential)) and varies greatly; (2) the fitted {\psi}/{\sigma} can exceed the allowable range of calculation; and (3) the measured FvdW and the fitted H vary greatly. Based on rigorous and comprehensive arguments, we have reasonably explained the pervasive anomalous results in the literature and further speculated that, the pervasive anomalous results are existing but not noticed and questioned owing to the two important aspects: (1) the pervasive unreasonable understandings of EDL theory and (2) the commonly neglected systematic errors. Consequently, we believe that the related studies have been seriously hampered. We therefore call for re-examination and re-analysis of related experimental results and theoretical understandings by careful consideration of the EDL theory and systematic errors. On these bases, we can interpret the experimental results properly and promote the development of EDL theory, colloid and interface science, and many related fields., Comment: 71 pages, 18 figures
- Published
- 2024
34. EELE: Exploring Efficient and Extensible LoRA Integration in Emotional Text-to-Speech
- Author
-
Qi, Xin, Fu, Ruibo, Wen, Zhengqi, Tao, Jianhua, Shi, Shuchen, Lu, Yi, Wang, Zhiyong, Wang, Xiaopeng, Xie, Yuankun, Liu, Yukun, Li, Guanjun, Liu, Xuefei, and Li, Yongwei
- Subjects
Computer Science - Sound ,Electrical Engineering and Systems Science - Audio and Speech Processing - Abstract
In the current era of Artificial Intelligence Generated Content (AIGC), a Low-Rank Adaptation (LoRA) method has emerged. It uses a plugin-based approach to learn new knowledge with lower parameter quantities and computational costs, and it can be plugged in and out based on the specific sub-tasks, offering high flexibility. However, the current application schemes primarily incorporate LoRA into the pre-introduced conditional parts of the speech models. This fixes the position of LoRA, limiting the flexibility and scalability of its application. Therefore, we propose the Exploring Efficient and Extensible LoRA Integration in Emotional Text-to-Speech (EELE) method. Starting from a general neutral speech model, we do not pre-introduce emotional information but instead use the LoRA plugin to design a flexible adaptive scheme that endows the model with emotional generation capabilities. Specifically, we initially train the model using only neutral speech data. After training is complete, we insert LoRA into different modules and fine-tune the model with emotional speech data to find the optimal insertion scheme. Through experiments, we compare and test the effects of inserting LoRA at different positions within the model and assess LoRA's ability to learn various emotions, effectively proving the validity of our method. Additionally, we explore the impact of the rank size of LoRA and the difference compared to directly fine-tuning the entire model.
- Published
- 2024
35. A Noval Feature via Color Quantisation for Fake Audio Detection
- Author
-
Wang, Zhiyong, Wang, Xiaopeng, Xie, Yuankun, Fu, Ruibo, Wen, Zhengqi, Tao, Jianhua, Liu, Yukun, Li, Guanjun, Qi, Xin, Lu, Yi, Liu, Xuefei, and Li, Yongwei
- Subjects
Computer Science - Sound ,Electrical Engineering and Systems Science - Audio and Speech Processing - Abstract
In the field of deepfake detection, previous studies focus on using reconstruction or mask and prediction methods to train pre-trained models, which are then transferred to fake audio detection training where the encoder is used to extract features, such as wav2vec2.0 and Masked Auto Encoder. These methods have proven that using real audio for reconstruction pre-training can better help the model distinguish fake audio. However, the disadvantage lies in poor interpretability, meaning it is hard to intuitively present the differences between deepfake and real audio. This paper proposes a noval feature extraction method via color quantisation which constrains the reconstruction to use a limited number of colors for the spectral image-like input. The proposed method ensures reconstructed input differs from the original, which allows for intuitive observation of the focus areas in the spectral reconstruction. Experiments conducted on the ASVspoof2019 dataset demonstrate that the proposed method achieves better classification performance compared to using the original spectral as input and pretraining the recolor network can also benefit the fake audio detection., Comment: accepted by ISCSLP2024
- Published
- 2024
36. ETGuard: Malicious Encrypted Traffic Detection in Blockchain-based Power Grid Systems
- Author
-
Zhou, Peng, Liu, Yongdong, Ma, Lixun, Zhang, Weiye, Tan, Haohan, Liu, Zhenguang, and Huang, Butian
- Subjects
Computer Science - Cryptography and Security ,Computer Science - Artificial Intelligence - Abstract
The escalating prevalence of encryption protocols has led to a concomitant surge in the number of malicious attacks that hide in encrypted traffic. Power grid systems, as fundamental infrastructure, are becoming prime targets for such attacks. Conventional methods for detecting malicious encrypted packets typically use a static pre-trained model. We observe that these methods are not well-suited for blockchain-based power grid systems. More critically, they fall short in dynamic environments where new types of encrypted attacks continuously emerge. Motivated by this, in this paper we try to tackle these challenges from two aspects: (1) We present a novel framework that is able to automatically detect malicious encrypted traffic in blockchain-based power grid systems and incrementally learn from new malicious traffic. (2) We mathematically derive incremental learning losses to resist the forgetting of old attack patterns while ensuring the model is capable of handling new encrypted attack patterns. Empirically, our method achieves state-of-the-art performance on three different benchmark datasets. We also constructed the first malicious encrypted traffic dataset for blockchain-based power grid scenario. Our code and dataset are available at https://github.com/PPPmzt/ETGuard, hoping to inspire future research.
- Published
- 2024
37. CoRA: Collaborative Information Perception by Large Language Model's Weights for Recommendation
- Author
-
Liu, Yuting, Zhang, Jinghao, Dang, Yizhou, Liang, Yuliang, Liu, Qiang, Guo, Guibing, Zhao, Jianzhe, and Wang, Xingwei
- Subjects
Computer Science - Information Retrieval ,Computer Science - Machine Learning - Abstract
Involving collaborative information in Large Language Models (LLMs) is a promising technique for adapting LLMs for recommendation. Existing methods achieve this by concatenating collaborative features with text tokens into a unified sequence input and then fine-tuning to align these features with LLM's input space. Although effective, in this work, we identify two limitations when adapting LLMs to recommendation tasks, which hinder the integration of general knowledge and collaborative information, resulting in sub-optimal recommendation performance. (1) Fine-tuning LLM with recommendation data can undermine its inherent world knowledge and fundamental competencies, which are crucial for interpreting and inferring recommendation text. (2) Incorporating collaborative features into textual prompts disrupts the semantics of the original prompts, preventing LLM from generating appropriate outputs. In this paper, we propose a new paradigm, CoRA (an acronym for Collaborative LoRA), with a collaborative weights generator. Rather than input space alignment, this method aligns collaborative information with LLM's parameter space, representing them as incremental weights to update LLM's output. This way, LLM perceives collaborative information without altering its general knowledge and text inference capabilities. Specifically, we employ a collaborative filtering model to extract user and item embeddings, converting them into collaborative weights with low-rank properties through the collaborative weights generator. We then merge the collaborative weights into LLM's weights, enabling LLM to perceive the collaborative signals and generate personalized recommendations without fine-tuning or extra collaborative tokens in prompts. Extensive experiments confirm that CoRA effectively integrates collaborative information into LLM, enhancing recommendation performance.
- Published
- 2024
38. Rethinking Video Segmentation with Masked Video Consistency: Did the Model Learn as Intended?
- Author
-
Liang, Chen, Guo, Qiang, Qu, Xiaochao, Liu, Luoqi, and Liu, Ting
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Video segmentation aims at partitioning video sequences into meaningful segments based on objects or regions of interest within frames. Current video segmentation models are often derived from image segmentation techniques, which struggle to cope with small-scale or class-imbalanced video datasets. This leads to inconsistent segmentation results across frames. To address these issues, we propose a training strategy Masked Video Consistency, which enhances spatial and temporal feature aggregation. MVC introduces a training strategy that randomly masks image patches, compelling the network to predict the entire semantic segmentation, thus improving contextual information integration. Additionally, we introduce Object Masked Attention (OMA) to optimize the cross-attention mechanism by reducing the impact of irrelevant queries, thereby enhancing temporal modeling capabilities. Our approach, integrated into the latest decoupled universal video segmentation framework, achieves state-of-the-art performance across five datasets for three video segmentation tasks, demonstrating significant improvements over previous methods without increasing model parameters.
- Published
- 2024
39. LSVOS Challenge 3rd Place Report: SAM2 and Cutie based VOS
- Author
-
Liu, Xinyu, Zhang, Jing, Zhang, Kexin, Liu, Xu, and Li, Lingling
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Information Retrieval - Abstract
Video Object Segmentation (VOS) presents several challenges, including object occlusion and fragmentation, the dis-appearance and re-appearance of objects, and tracking specific objects within crowded scenes. In this work, we combine the strengths of the state-of-the-art (SOTA) models SAM2 and Cutie to address these challenges. Additionally, we explore the impact of various hyperparameters on video instance segmentation performance. Our approach achieves a J\&F score of 0.7952 in the testing phase of LSVOS challenge VOS track, ranking third overall., Comment: arXiv admin note: text overlap with arXiv:2406.03668
- Published
- 2024
40. MeshFormer: High-Quality Mesh Generation with 3D-Guided Reconstruction Model
- Author
-
Liu, Minghua, Zeng, Chong, Wei, Xinyue, Shi, Ruoxi, Chen, Linghao, Xu, Chao, Zhang, Mengqi, Wang, Zhaoning, Zhang, Xiaoshuai, Liu, Isabella, Wu, Hongzhi, and Su, Hao
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Graphics - Abstract
Open-world 3D reconstruction models have recently garnered significant attention. However, without sufficient 3D inductive bias, existing methods typically entail expensive training costs and struggle to extract high-quality 3D meshes. In this work, we introduce MeshFormer, a sparse-view reconstruction model that explicitly leverages 3D native structure, input guidance, and training supervision. Specifically, instead of using a triplane representation, we store features in 3D sparse voxels and combine transformers with 3D convolutions to leverage an explicit 3D structure and projective bias. In addition to sparse-view RGB input, we require the network to take input and generate corresponding normal maps. The input normal maps can be predicted by 2D diffusion models, significantly aiding in the guidance and refinement of the geometry's learning. Moreover, by combining Signed Distance Function (SDF) supervision with surface rendering, we directly learn to generate high-quality meshes without the need for complex multi-stage training processes. By incorporating these explicit 3D biases, MeshFormer can be trained efficiently and deliver high-quality textured meshes with fine-grained geometric details. It can also be integrated with 2D diffusion models to enable fast single-image-to-3D and text-to-3D tasks. Project page: https://meshformer3d.github.io, Comment: 20 pages, 9 figures
- Published
- 2024
41. SpaRP: Fast 3D Object Reconstruction and Pose Estimation from Sparse Views
- Author
-
Xu, Chao, Li, Ang, Chen, Linghao, Liu, Yulin, Shi, Ruoxi, Su, Hao, and Liu, Minghua
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence ,Computer Science - Graphics - Abstract
Open-world 3D generation has recently attracted considerable attention. While many single-image-to-3D methods have yielded visually appealing outcomes, they often lack sufficient controllability and tend to produce hallucinated regions that may not align with users' expectations. In this paper, we explore an important scenario in which the input consists of one or a few unposed 2D images of a single object, with little or no overlap. We propose a novel method, SpaRP, to reconstruct a 3D textured mesh and estimate the relative camera poses for these sparse-view images. SpaRP distills knowledge from 2D diffusion models and finetunes them to implicitly deduce the 3D spatial relationships between the sparse views. The diffusion model is trained to jointly predict surrogate representations for camera poses and multi-view images of the object under known poses, integrating all information from the input sparse views. These predictions are then leveraged to accomplish 3D reconstruction and pose estimation, and the reconstructed 3D model can be used to further refine the camera poses of input views. Through extensive experiments on three datasets, we demonstrate that our method not only significantly outperforms baseline methods in terms of 3D reconstruction quality and pose prediction accuracy but also exhibits strong efficiency. It requires only about 20 seconds to produce a textured mesh and camera poses for the input views. Project page: https://chaoxu.xyz/sparp., Comment: ECCV 2024
- Published
- 2024
42. Physics-Aware Combinatorial Assembly Planning using Deep Reinforcement Learning
- Author
-
Liu, Ruixuan, Chen, Alan, Zhao, Weiye, and Liu, Changliu
- Subjects
Computer Science - Robotics ,Computer Science - Machine Learning - Abstract
Combinatorial assembly uses standardized unit primitives to build objects that satisfy user specifications. Lego is a widely used platform for combinatorial assembly, in which people use unit primitives (ie Lego bricks) to build highly customizable 3D objects. This paper studies sequence planning for physical combinatorial assembly using Lego. Given the shape of the desired object, we want to find a sequence of actions for placing Lego bricks to build the target object. In particular, we aim to ensure the planned assembly sequence is physically executable. However, assembly sequence planning (ASP) for combinatorial assembly is particularly challenging due to its combinatorial nature, ie the vast number of possible combinations and complex constraints. To address the challenges, we employ deep reinforcement learning to learn a construction policy for placing unit primitives sequentially to build the desired object. Specifically, we design an online physics-aware action mask that efficiently filters out invalid actions and guides policy learning. In the end, we demonstrate that the proposed method successfully plans physically valid assembly sequences for constructing different Lego structures. The generated construction plan can be executed in real.
- Published
- 2024
43. GLIMMER: Incorporating Graph and Lexical Features in Unsupervised Multi-Document Summarization
- Author
-
Liu, Ran, Liu, Ming, Yu, Min, Jiang, Jianguo, Li, Gang, Zhang, Dan, Li, Jingyuan, Meng, Xiang, and Huang, Weiqing
- Subjects
Computer Science - Computation and Language - Abstract
Pre-trained language models are increasingly being used in multi-document summarization tasks. However, these models need large-scale corpora for pre-training and are domain-dependent. Other non-neural unsupervised summarization approaches mostly rely on key sentence extraction, which can lead to information loss. To address these challenges, we propose a lightweight yet effective unsupervised approach called GLIMMER: a Graph and LexIcal features based unsupervised Multi-docuMEnt summaRization approach. It first constructs a sentence graph from the source documents, then automatically identifies semantic clusters by mining low-level features from raw texts, thereby improving intra-cluster correlation and the fluency of generated sentences. Finally, it summarizes clusters into natural sentences. Experiments conducted on Multi-News, Multi-XScience and DUC-2004 demonstrate that our approach outperforms existing unsupervised approaches. Furthermore, it surpasses state-of-the-art pre-trained multi-document summarization models (e.g. PEGASUS and PRIMERA) under zero-shot settings in terms of ROUGE scores. Additionally, human evaluations indicate that summaries generated by GLIMMER achieve high readability and informativeness scores. Our code is available at https://github.com/Oswald1997/GLIMMER., Comment: 19 pages, 7 figures. Accepted by ECAI 2024
- Published
- 2024
44. Edge detection imaging by quasi-bound states in the continuum
- Author
-
Liu, Tingting, Qiu, Jumin, Xu, Lei, Qin, Meibao, Wan, Lipeng, Yu, Tianbao, Liu, Qiegen, Huang, Lujun, and Xiao, Shuyuan
- Subjects
Physics - Optics ,Physics - Applied Physics - Abstract
Optical metasurfaces have revolutionized analog computing and image processing at sub-wavelength scales with faster speed and lower power consumption. They typically involve spatial differentiation with engineered angular dispersion. Quasi-bound states in the continuum (quasi-BICs) have recently emerged as a powerful tool for tailoring properties of optical resonances. While quasi-BICs have been explored in various applications that require high $Q$-factors and enhanced field confinement, their full potential in image processing remains unexplored. Here, we demonstrate edge detection imaging by leveraging a quasi-BIC in an all-dielectric metasurface. This metasurface, composed of four nanodisks per unit cell, supports a polarization-independent quasi-BIC through structural perturbations, allowing simultaneously engineering $Q$-factor and angular dispersion. Importantly, we find that with suitable parameters, this quasi-BIC metasurface can perform isotropic two-dimensional spatial differentiation, which is the core element for realizing edge detection. Following the theoretical design, we fabricate the metasurfaces on the silicon-on-insulator platform and experimentally validate their capability of high-quality, efficient, and uniform edge detection imaging under different incident polarizations. Our results illuminate the mechanisms of edge detection with quasi-BIC metasurfaces and highlight new opportunities for their application in ultra-compact, low-power optical computing devices., Comment: 17 pages, 5 figures
- Published
- 2024
45. The influence of magnon renormalization and interband coupling on the spin Seebeck effect in YIG
- Author
-
Yin, Yuling, Liu, Yang, Liu, Yiqun, and Wan, Xiangang
- Subjects
Condensed Matter - Materials Science - Abstract
With exceptionally low magnetic damping, YIG has been extensively applied in the realm of magnetism, encompassing the researches into the spin Seebeck effect. YIG has 20 magnon bands, including 8 higher-energy bands denoted as $\alpha_{1\sim8}$, and 12 lower-energy bands denoted as $\beta_{1\sim12}$. Here, we study the impact of the complex intraband and interband magnon couplings on the transport coefficients of YIG. Four-magnon processes in YIG are considered, and a self-consistent mean-field approximation is made for these interaction terms. We find that the $\beta$ bands exhibit minimal variation with increasing temperature, whereas the $\alpha$ bands undergo a noticeable decline as the temperature rises. These counterintuitive results agree well with the observation of earlier inelastic neutron scattering experiments and the results of the theoretical calculations in recent years. We also find it sufficient to include only the contribution of magnons on the acoustic band $\beta_1$ when studying the spin conductivity ($\sigma_m$). However, when calculating the spin Seebeck coefficient ($S_m$) and the magnon thermal conductivity ($\kappa_m$), the results calculated using only $\beta_{1}$ show noticeable deviations over a large temperature range compared to the full band calculations. These deviations are well mitigated when the $\beta_{2}$ and $\beta_{3}$ bands are considered.
- Published
- 2024
46. Exploiting Fine-Grained Prototype Distribution for Boosting Unsupervised Class Incremental Learning
- Author
-
Liu, Jiaming, Liu, Hongyuan, Qin, Zhili, Han, Wei, Fan, Yulu, Yang, Qinli, and Shao, Junming
- Subjects
Computer Science - Machine Learning ,Computer Science - Computer Vision and Pattern Recognition - Abstract
The dynamic nature of open-world scenarios has attracted more attention to class incremental learning (CIL). However, existing CIL methods typically presume the availability of complete ground-truth labels throughout the training process, an assumption rarely met in practical applications. Consequently, this paper explores a more challenging problem of unsupervised class incremental learning (UCIL). The essence of addressing this problem lies in effectively capturing comprehensive feature representations and discovering unknown novel classes. To achieve this, we first model the knowledge of class distribution by exploiting fine-grained prototypes. Subsequently, a granularity alignment technique is introduced to enhance the unsupervised class discovery. Additionally, we proposed a strategy to minimize overlap between novel and existing classes, thereby preserving historical knowledge and mitigating the phenomenon of catastrophic forgetting. Extensive experiments on the five datasets demonstrate that our approach significantly outperforms current state-of-the-art methods, indicating the effectiveness of the proposed method.
- Published
- 2024
47. Two points are enough
- Author
-
Liu, Hao, Zhao, Yanbin, Zheng, Huarong, Fan, Xiulin, Deng, Zhihua, Chen, Mengchi, Wang, Xingkai, Liu, Zhiyang, Lu, Jianguo, and Chen, Jian
- Subjects
Condensed Matter - Materials Science ,Physics - Data Analysis, Statistics and Probability - Abstract
Prognosis and diagnosis play an important role in accelerating the development of lithium-ion batteries, as well as reliable and long-life operation. In this work, we answer an important question: What is the minimum amount of data required to extract features for accurate battery prognosis and diagnosis? Based on the first principle, we successfully extracted the best two-point feature (BTPF) for accurate battery prognosis and diagnosis using the fewest data points (only two) and the simplest feature selection method (Pearson correlation coefficient). The BTPF extraction method is tested on 820 cells from 6 open-source datasets (covering five different chemistry types, seven manufacturers, and three data types). It achieves comparable accuracy to state-of-the-art features in both prognosis and diagnosis tasks. This work challenges the cognition of existing studies on the difficulty of battery prognosis and diagnosis tasks, subverts the fixed pattern of establishing prognosis and diagnosis methods for complex dynamic systems through deliberate feature engineering, highlights the promise of data-driven methods for field battery prognosis and diagnosis applications, and provides a new benchmark for future studies.
- Published
- 2024
48. Tunable interfacial Rashba spin-orbit coupling in asymmetric Al$_x$In$_{1-x}$Sb/InSb/CdTe quantum well heterostructures
- Author
-
Ruan, Hanzhi, Zhi, Zhenghang, Wu, Yuyang, Liu, Jiuming, Huang, Puyang, Yao, Shan, Liu, Xinqi, Tang, Chenjia, Yao, Qi, Sun, Lu, Zhang, Yifan, Xiao, Yujie, Che, Renchao, and Kou, Xufeng
- Subjects
Condensed Matter - Materials Science ,Condensed Matter - Mesoscale and Nanoscale Physics - Abstract
The manipulation of Rashba-type spin-orbit coupling (SOC) in molecular beam epitaxy-grown Al$_x$In$_{1-x}$Sb/InSb/CdTe quantum well heterostructures is reported. The effective band bending provides robust two-dimensional quantum confinement, while the unidirectional built-in electric field from the asymmetric hetero-interfaces results in pronounced Rashba SOC strength. By tuning the Al concentration in the top Al$_x$In$_{1-x}$Sb barrier layer, the optimal structure with $x = 0.15$ shows the largest Rashba coefficient of 0.23 eV-Angstrom. and the highest low-temperature electron mobility of 4400 cm$^2$/Vs . Quantitative investigations of the weak anti-localization effect further confirm the dominant D'yakonov-Perel (DP) spin relaxation mechanism during charge-to-spin conversion. These findings highlight the significance of quantum well engineering in shaping magneto-resistance responses, and narrow bandgap semiconductor-based heterostructures may offer a reliable platform for energy-efficient spintronic applications.
- Published
- 2024
49. Optimal Few-GHW Linear Codes and Their Subcode Support Weight Distributions
- Author
-
Pan, Xu, Chen, Hao, Liu, Hongwei, and Liu, Shengwei
- Subjects
Computer Science - Information Theory - Abstract
Few-weight codes have been constructed and studied for many years, since their fascinating relations to finite geometries, strongly regular graphs and Boolean functions. Simplex codes are one-weight Griesmer $[\frac{q^k-1}{q-1},k ,q^{k-1}]_q$-linear codes and they meet all Griesmer bounds of the generalized Hamming weights of linear codes. All the subcodes with dimension $r$ of a $[\frac{q^k-1}{q-1},k ,q^{k-1}]_q$-simplex code have the same subcode support weight $\frac{q^{k-r}(q^r-1)}{q-1}$ for $1\leq r\leq k$. In this paper, we construct linear codes meeting the Griesmer bound of the $r$-generalized Hamming weight, such codes do not meet the Griesmer bound of the $j$-generalized Hamming weight for $1\leq j
- Published
- 2024
50. The curse of random quantum data
- Author
-
Zhang, Kaining, Liu, Junyu, Liu, Liu, Jiang, Liang, Hsieh, Min-Hsiu, and Tao, Dacheng
- Subjects
Quantum Physics ,Computer Science - Machine Learning - Abstract
Quantum machine learning, which involves running machine learning algorithms on quantum devices, may be one of the most significant flagship applications for these devices. Unlike its classical counterparts, the role of data in quantum machine learning has not been fully understood. In this work, we quantify the performances of quantum machine learning in the landscape of quantum data. Provided that the encoding of quantum data is sufficiently random, the performance, we find that the training efficiency and generalization capabilities in quantum machine learning will be exponentially suppressed with the increase in the number of qubits, which we call "the curse of random quantum data". Our findings apply to both the quantum kernel method and the large-width limit of quantum neural networks. Conversely, we highlight that through meticulous design of quantum datasets, it is possible to avoid these curses, thereby achieving efficient convergence and robust generalization. Our conclusions are corroborated by extensive numerical simulations., Comment: 40 pages, 8 figures
- Published
- 2024
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.