94 results on '"Huang, Heyan"'
Search Results
2. Quantization Method Integrated with Progressive Quantization and Distillation Learning
- Author
-
Huang, Heyan, Pan, Bocheng, Wang, Liwei, and Jiang, Cheng
- Published
- 2023
- Full Text
- View/download PDF
3. Incorporating history and future into non-autoregressive machine translation
- Author
-
Wang, Shuheng, Huang, Heyan, and Shi, Shumin
- Published
- 2023
- Full Text
- View/download PDF
4. Chinese sentence semantic matching based on multi-level relevance extraction and aggregation for intelligent human–robot interaction
- Author
-
Lu, Wenpeng, Zhao, Pengyu, Li, Yifeng, Wang, Shoujin, Huang, Heyan, Shi, Shumin, and Wu, Hao
- Published
- 2022
- Full Text
- View/download PDF
5. A pattern-aware self-attention network for distant supervised relation extraction
- Author
-
Shang, Yu-Ming, Huang, Heyan, Sun, Xin, Wei, Wei, and Mao, Xian-Ling
- Published
- 2022
- Full Text
- View/download PDF
6. Food recommendation with graph convolutional network
- Author
-
Gao, Xiaoyan, Feng, Fuli, Huang, Heyan, Mao, Xian-Ling, Lan, Tian, and Chi, Zewen
- Published
- 2022
- Full Text
- View/download PDF
7. Immobilization of Cs and Sr within perovskite-type Ba0.7-ySry(La, Cs)0.3ZrO3 glass/ceramic composite waste forms
- Author
-
Liu, Haifeng, Wang, Hua, Zhao, Jian, Li, Jie, Zhang, Xingquan, Yang, Jiacheng, Zhu, Yongchang, Xie, Ruishi, Zheng, Kui, Huang, Heyan, and Huo, Jichuan
- Published
- 2022
- Full Text
- View/download PDF
8. Domain structure-based transfer learning for cross-domain word representation
- Author
-
Huang, Heyan and Liu, Qian
- Published
- 2021
- Full Text
- View/download PDF
9. Classifier-adaptation knowledge distillation framework for relation extraction and event detection with imbalanced data
- Author
-
Song, Dandan, Xu, Jing, Pang, Jinhui, and Huang, Heyan
- Published
- 2021
- Full Text
- View/download PDF
10. Deep kernel supervised hashing for node classification in structural networks
- Author
-
Guo, Jia-Nan, Mao, Xian-Ling, Lin, Shu-Yang, Wei, Wei, and Huang, Heyan
- Published
- 2021
- Full Text
- View/download PDF
11. Document-level relation extraction with Entity-Selection Attention
- Author
-
Yuan, Changsen, Huang, Heyan, Feng, Chong, Shi, Ge, and Wei, Xiaochi
- Published
- 2021
- Full Text
- View/download PDF
12. Extracting salient features from convolutional discriminative filters
- Author
-
Zhou, Yuxiang, Liao, Lejian, Gao, Yang, and Huang, Heyan
- Published
- 2021
- Full Text
- View/download PDF
13. Domain-specific meta-embedding with latent semantic structures
- Author
-
Liu, Qian, Lu, Jie, Zhang, Guangquan, Shen, Tao, Zhang, Zhihan, and Huang, Heyan
- Published
- 2021
- Full Text
- View/download PDF
14. Rapid and high sensing response to acetone based on La1-xBaxMnO3+δ nanopowders
- Author
-
Liu, Haifeng, Li, Chengbo, Xie, Ruishi, Li, Jie, Ma, Guohua, Zheng, Kui, Zhang, Xingquan, Huang, Heyan, Peng, Tongjiang, and Huo, Jichuan
- Published
- 2021
- Full Text
- View/download PDF
15. Meteorological Characteristics of a Continuous Ice-Covered Event on Ultra-High Voltage Transmission Lines in Yunnan Region in 2021.
- Author
-
He, Sen, Song, Yunhai, Huang, Heyan, He, Yuhao, Zhou, Shaohui, and Gao, Zhiqiu
- Subjects
ELECTRIC lines ,AIR masses ,METEOROLOGICAL observations ,ELECTRIC power distribution grids ,VOLTAGE - Abstract
Yunnan plays a pivotal role in transmitting electricity from west to east within China's Southern Power Grid. During 7–13 January 2021, a large-scale continuous ice-covering event of ultra-high voltage (UHV) transmission lines occurred in the Qujing area of eastern Yunnan Province. Based on ERA5 reanalysis data and meteorological observation data of UHV transmission line icing in China's Southern Power Grid, the synoptic causes of the icing are comprehensively analyzed from various perspectives, including weather situations, vertical stratification of temperature and humidity, local meteorological elements, and atmospheric circulation indices. The results indicate a strong East Asian trough and a blocking high directing northern airflow southward ahead of the ridge. Cold air enters the Qujing area and combines with warm and moist air from the subtropical high pressure of 50–110° E. As warm and cold air masses form a quasi-stationary front over the northern mountainous area of Qujing due to topographic uplift, the mechanism of "supercooling and warm rain" caused by the "warm–cold" temperature profile structure leads to freezing rain events. Large-scale circulation indices in the Siberian High, East Asian Trough, and 50–110° E Subtropical High regions provided clear precursor signals within 0–2 days before the icing events. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
16. Synthesis and comparative studies of coumarin-based nonlinear optical chromophores with different conjugated electron bridge
- Author
-
Deng, Guowei, Chen, Yuan, Zheng, Dongmei, Zhu, Jiang, Huang, Heyan, Sun, Kang, Zhang, Xiaoling, Li, Zhonghui, and Liu, Jialei
- Published
- 2020
- Full Text
- View/download PDF
17. Fabrication, structure, electrochemical properties, and enhanced pseudocapacitive performance of cobalt oxyhydroxide films via a simple strategy
- Author
-
Guo, Baogang, Li, Tongcai, Liu, Haifeng, Xie, Ruishi, Zhang, Xingquan, Huang, Heyan, Li, Yuanli, Ma, Yongjun, and Huo, Jichuan
- Published
- 2020
- Full Text
- View/download PDF
18. A simple, effective and ligand-free route toward cadmium sulfide nanocrystals on titanium plate: Insights into their crystal structure, local structure and optical properties
- Author
-
Xie, Ruishi, Li, Yuanli, Huang, Heyan, Pan, Xiaoqin, Guo, Baogang, Liu, Haifeng, Hu, Hailong, Zhang, Xingquan, and Ma, Yongjun
- Published
- 2019
- Full Text
- View/download PDF
19. Social Influence Analysis: Models, Methods, and Evaluation
- Author
-
Li, Kan, Zhang, Lin, and Huang, Heyan
- Published
- 2018
- Full Text
- View/download PDF
20. AttenWalker: Unsupervised Long-Document Question Answering via Attention-based Graph Walking
- Author
-
Nie, Yuxiang, Huang, Heyan, Wei, Wei, and Mao, Xian-Ling
- Subjects
FOS: Computer and information sciences ,Computer Science - Computation and Language ,Computation and Language (cs.CL) - Abstract
Annotating long-document question answering (long-document QA) pairs is time-consuming and expensive. To alleviate the problem, it might be possible to generate long-document QA pairs via unsupervised question answering (UQA) methods. However, existing UQA tasks are based on short documents, and can hardly incorporate long-range information. To tackle the problem, we propose a new task, named unsupervised long-document question answering (ULQA), aiming to generate high-quality long-document QA instances in an unsupervised manner. Besides, we propose AttenWalker, a novel unsupervised method to aggregate and generate answers with long-range dependency so as to construct long-document QA pairs. Specifically, AttenWalker is composed of three modules, i.e., span collector, span linker and answer aggregator. Firstly, the span collector takes advantage of constituent parsing and reconstruction loss to select informative candidate spans for constructing answers. Secondly, by going through the attention graph of a pre-trained long-document model, potentially interrelated text spans (that might be far apart) could be linked together via an attention-walking algorithm. Thirdly, in the answer aggregator, linked spans are aggregated into the final answer via the mask-filling ability of a pre-trained model. Extensive experiments show that AttenWalker outperforms previous methods on Qasper and NarrativeQA. In addition, AttenWalker also shows strong performance in the few-shot learning setting., Accepted to the Findings of ACL 2023
- Published
- 2023
21. Extending Embedding Representation by Incorporating Latent Relations
- Author
-
Gao Yang, Wang Wenbo, Liu Qian, Huang Heyan, and Yuefeng Li
- Subjects
Word embedding ,text mining ,natural language processing ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
The semantic representation of words is a fundamental task in natural language processing and text mining. Learning word embedding has shown its power on various tasks. Most studies are aimed at generating embedding representation of a word based on encoding its context information. However, many latent relations, such as co-occurring associative patterns and semantic conceptual relations, are not well considered. In this paper, we propose an extensible model to incorporate these kinds of valuable latent relations to increase the semantic relatedness of word pairs by learning word embeddings. To assess the effectiveness of our model, we conduct experiments on both information retrieval and text classification tasks. The results indicate the effectiveness of our model as well as its flexibility on different tasks.
- Published
- 2018
- Full Text
- View/download PDF
22. Labeled Phrase Latent Dirichlet Allocation and its online learning algorithm
- Author
-
Tang, Yi-Kun, Mao, Xian-Ling, and Huang, Heyan
- Published
- 2018
- Full Text
- View/download PDF
23. Bridging The Gap: Entailment Fused-T5 for Open-retrieval Conversational Machine Reading Comprehension
- Author
-
Zhang, Xiao, Huang, Heyan, Chi, Zewen, and Mao, Xian-Ling
- Subjects
FOS: Computer and information sciences ,Computer Science - Computation and Language ,Computation and Language (cs.CL) - Abstract
Open-retrieval conversational machine reading comprehension (OCMRC) simulates real-life conversational interaction scenes. Machines are required to make a decision of "Yes/No/Inquire" or generate a follow-up question when the decision is "Inquire" based on retrieved rule texts, user scenario, user question, and dialogue history. Recent studies explored the methods to reduce the information gap between decision-making and question generation and thus improve the performance of generation. However, the information gap still exists because these pipeline structures are still limited in decision-making, span extraction, and question rephrasing three stages. Decision-making and generation are reasoning separately, and the entailment reasoning utilized in decision-making is hard to share through all stages. To tackle the above problem, we proposed a novel one-stage end-to-end framework, called Entailment Fused-T5 (EFT), to bridge the information gap between decision-making and generation in a global understanding manner. The extensive experimental results demonstrate that our proposed framework achieves new state-of-the-art performance on the OR-ShARC benchmark.
- Published
- 2022
24. ConsPrompt: Easily Exploiting Contrastive Samples for Few-shot Prompt Learning
- Author
-
Weng, Jinta, Hu, Yue, Tian, Zhihong, and Huang, Heyan
- Subjects
FOS: Computer and information sciences ,Artificial Intelligence (cs.AI) ,Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computation and Language (cs.CL) - Abstract
Prompt learning recently become an effective linguistic tool to motivate the PLMs' knowledge on few-shot-setting tasks. However, studies have shown the lack of robustness still exists in prompt learning, since suitable initialization of continuous prompt and expert-first manual prompt are essential in fine-tuning process. What is more, human also utilize their comparative ability to motivate their existing knowledge for distinguishing different examples. Motivated by this, we explore how to use contrastive samples to strengthen prompt learning. In detail, we first propose our model ConsPrompt combining with prompt encoding network, contrastive sampling module, and contrastive scoring module. Subsequently, two sampling strategies, similarity-based and label-based strategies, are introduced to realize differential contrastive learning. The effectiveness of proposed ConsPrompt is demonstrated in five different few-shot learning tasks and shown the similarity-based sampling strategy is more effective than label-based in combining contrastive learning. Our results also exhibits the state-of-the-art performance and robustness in different few-shot settings, which proves that the ConsPrompt could be assumed as a better knowledge probe to motivate PLMs.
- Published
- 2022
25. Revisiting Grammatical Error Correction Evaluation and Beyond
- Author
-
Gong, Peiyuan, Liu, Xuebo, Huang, Heyan, and Zhang, Min
- Subjects
FOS: Computer and information sciences ,Artificial Intelligence (cs.AI) ,Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computation and Language (cs.CL) - Abstract
Pretraining-based (PT-based) automatic evaluation metrics (e.g., BERTScore and BARTScore) have been widely used in several sentence generation tasks (e.g., machine translation and text summarization) due to their better correlation with human judgments over traditional overlap-based methods. Although PT-based methods have become the de facto standard for training grammatical error correction (GEC) systems, GEC evaluation still does not benefit from pretrained knowledge. This paper takes the first step towards understanding and improving GEC evaluation with pretraining. We first find that arbitrarily applying PT-based metrics to GEC evaluation brings unsatisfactory correlation results because of the excessive attention to inessential systems outputs (e.g., unchanged parts). To alleviate the limitation, we propose a novel GEC evaluation metric to achieve the best of both worlds, namely PT-M2 which only uses PT-based metrics to score those corrected parts. Experimental results on the CoNLL14 evaluation task show that PT-M2 significantly outperforms existing methods, achieving a new state-of-the-art result of 0.949 Pearson correlation. Further analysis reveals that PT-M2 is robust to evaluate competitive GEC systems. Source code and scripts are freely available at https://github.com/pygongnlp/PT-M2., Accepted to EMNLP 2022
- Published
- 2022
26. Unsupervised Hashing with Semantic Concept Mining
- Author
-
Tu, Rong-Cheng, Mao, Xian-Ling, Lin, Kevin Qinghong, Cai, Chengfei, Qin, Weize, Wang, Hongfa, Wei, Wei, and Huang, Heyan
- Subjects
FOS: Computer and information sciences ,Computer Vision and Pattern Recognition (cs.CV) ,Computer Science - Computer Vision and Pattern Recognition ,Information Retrieval (cs.IR) ,Computer Science - Information Retrieval - Abstract
Recently, to improve the unsupervised image retrieval performance, plenty of unsupervised hashing methods have been proposed by designing a semantic similarity matrix, which is based on the similarities between image features extracted by a pre-trained CNN model. However, most of these methods tend to ignore high-level abstract semantic concepts contained in images. Intuitively, concepts play an important role in calculating the similarity among images. In real-world scenarios, each image is associated with some concepts, and the similarity between two images will be larger if they share more identical concepts. Inspired by the above intuition, in this work, we propose a novel Unsupervised Hashing with Semantic Concept Mining, called UHSCM, which leverages a VLP model to construct a high-quality similarity matrix. Specifically, a set of randomly chosen concepts is first collected. Then, by employing a vision-language pretraining (VLP) model with the prompt engineering which has shown strong power in visual representation learning, the set of concepts is denoised according to the training images. Next, the proposed method UHSCM applies the VLP model with prompting again to mine the concept distribution of each image and construct a high-quality semantic similarity matrix based on the mined concept distributions. Finally, with the semantic similarity matrix as guiding information, a novel hashing loss with a modified contrastive loss based regularization item is proposed to optimize the hashing network. Extensive experiments on three benchmark datasets show that the proposed method outperforms the state-of-the-art baselines in the image retrieval task.
- Published
- 2022
27. Knowledge‐enriched joint‐learning model for implicit emotion cause extraction.
- Author
-
Wu, Chenghao, Shi, Shumin, Hu, Jiaxing, and Huang, Heyan
- Abstract
Emotion cause extraction (ECE) task that aims at extracting potential trigger events of certain emotions has attracted extensive attention recently. However, current work neglects the implicit emotion expressed without any explicit emotional keywords, which appears more frequently in application scenarios. The lack of explicit emotion information makes it extremely hard to extract emotion causes only with the local context. Moreover, an entire event is usually across multiple clauses, while existing work merely extracts cause events at clause level and cannot effectively capture complete cause event information. To address these issues, the events are first redefined at the tuple level and a span‐based tuple‐level algorithm is proposed to extract events from different clauses. Based on it, a corpus for implicit emotion cause extraction that tries to extract causes of implicit emotions is constructed. The authors propose a knowledge‐enriched joint‐learning model of implicit emotion recognition and implicit emotion cause extraction tasks (KJ‐IECE), which leverages commonsense knowledge from ConceptNet and NRC_VAD to better capture connections between emotion and corresponding cause events. Experiments on both implicit and explicit emotion cause extraction datasets demonstrate the effectiveness of the proposed model. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
28. Case-Sensitive Neural Machine Translation
- Author
-
Shi, Xuewen, Huang, Heyan, Jian, Ping, and Tang, Yi-Kun
- Subjects
Case-sensitive ,Neural machine translation ,Natural language processing ,Article - Abstract
Even as an important lexical information for Latin languages, word case is often ignored in machine translation. According to observations, the translation performance drops significantly when we introduce case-sensitive evaluation metrics. In this paper, we introduce two types of case-sensitive neural machine translation (NMT) approaches to alleviate the above problems: i) adding case tokens into the decoding sequence, and ii) adopting case prediction to the conventional NMT. Our proposed approaches incorporate case information to the NMT decoder by jointly learning target word generation and word case prediction. We compare our approaches with multiple kinds of baselines including NMT with naive case-restoration methods and analyze the impacts of various setups on our approaches. Experimental results on three typical translation tasks (Zh-En, En-Fr, En-De) show that our proposed methods lead to the improvements up to 2.5, 1.0 and 0.5 in case-sensitive BLEU scores respectively. Further analyses also illustrate the inherent reasons why our approaches lead to different improvements on different translation tasks.
- Published
- 2020
29. Exploring Dense Retrieval for Dialogue Response Selection
- Author
-
Lan, Tian, Cai, Deng, Wang, Yan, Su, Yixuan, Huang, Heyan, and Mao, Xian-Ling
- Subjects
FOS: Computer and information sciences ,Artificial Intelligence (cs.AI) ,Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computation and Language (cs.CL) - Abstract
Recent progress in deep learning has continuously improved the accuracy of dialogue response selection. In particular, sophisticated neural network architectures are leveraged to capture the rich interactions between dialogue context and response candidates. While remarkably effective, these models also bring in a steep increase in computational cost. Consequently, such models can only be used as a re-rank module in practice. In this study, we present a solution to directly select proper responses from a large corpus or even a nonparallel corpus that only consists of unpaired sentences, using a dense retrieval model. To push the limits of dense retrieval, we design an interaction layer upon the dense retrieval models and apply a set of tailor-designed learning strategies. Our model shows superiority over strong baselines on the conventional re-rank evaluation setting, which is remarkable given its efficiency. To verify the effectiveness of our approach in realistic scenarios, we also conduct full-rank evaluation, where the target is to select proper responses from a full candidate pool that may contain millions of candidates and evaluate them fairly through human annotations. Our proposed model notably outperforms pipeline baselines that integrate fast recall and expressive re-rank modules. Human evaluation results show that enlarging the candidate pool with nonparallel corpora improves response quality further., 11 pages, 4 figures, 6 tables
- Published
- 2021
30. XLM-E: Cross-lingual Language Model Pre-training via ELECTRA
- Author
-
Chi, Zewen, Huang, Shaohan, Dong, Li, Ma, Shuming, Zheng, Bo, Singhal, Saksham, Bajaj, Payal, Song, Xia, Mao, Xian-Ling, Huang, Heyan, and Wei, Furu
- Subjects
FOS: Computer and information sciences ,Computer Science - Computation and Language ,InformationSystems_INFORMATIONSTORAGEANDRETRIEVAL ,ComputingMethodologies_DOCUMENTANDTEXTPROCESSING ,Computation and Language (cs.CL) - Abstract
In this paper, we introduce ELECTRA-style tasks to cross-lingual language model pre-training. Specifically, we present two pre-training tasks, namely multilingual replaced token detection, and translation replaced token detection. Besides, we pretrain the model, named as XLM-E, on both multilingual and parallel corpora. Our model outperforms the baseline models on various cross-lingual understanding tasks with much less computation cost. Moreover, analysis shows that XLM-E tends to obtain better cross-lingual transferability., ACL-2022
- Published
- 2021
31. Bone Tissue Engineering in the Treatment of Bone Defects.
- Author
-
Xue, Nannan, Ding, Xiaofeng, Huang, Rizhong, Jiang, Ruihan, Huang, Heyan, Pan, Xin, Min, Wen, Chen, Jun, Duan, Jin-Ao, Liu, Pei, and Wang, Yiwei
- Subjects
TISSUE scaffolds ,TISSUE engineering ,THREE-dimensional printing ,OLDER people ,COMPUTER engineering ,BONE grafting ,DISEASE incidence - Abstract
Bones play an important role in maintaining exercise and protecting organs. Bone defect, as a common orthopedic disease in clinics, can cause tremendous damage with long treatment cycles. Therefore, the treatment of bone defect remains as one of the main challenges in clinical practice. Today, with increased incidence of bone disease in the aging population, demand for bone repair material is high. At present, the method of clinical treatment for bone defects including non-invasive therapy and invasive therapy. Surgical treatment is the most effective way to treat bone defects, such as using bone grafts, Masquelet technique, Ilizarov technique etc. In recent years, the rapid development of tissue engineering technology provides a new treatment strategy for bone repair. This review paper introduces the current situation and challenges of clinical treatment of bone defect repair in detail. The advantages and disadvantages of bone tissue engineering scaffolds are comprehensively discussed from the aspect of material, preparation technology, and function of bone tissue engineering scaffolds. This paper also summarizes the 3D printing technology based on computer technology, aiming at designing personalized artificial scaffolds that can accurately fit bone defects. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
32. Improving Non-Autoregressive Machine Translation Using Sentence-Level Semantic Agreement.
- Author
-
Wang, Shuheng, Huang, Heyan, and Shi, Shumin
- Subjects
MACHINE translating - Abstract
Theinference stage can be accelerated significantly using a Non-Autoregressive Transformer (NAT). However, the training objective used in the NAT model also aims to minimize the loss between the generated words and the golden words in the reference. Since the dependencies between the target words are lacking, this training objective computed at word level can easily cause semantic inconsistency between the generated and source sentences. To alleviate this issue, we propose a new method, Sentence-Level Semantic Agreement (SLSA), to obtain consistency between the source and generated sentences. Specifically, we utilize contrastive learning to pull the sentence representations of the source and generated sentences closer together. In addition, to strengthen the capability of the encoder, we also integrate an agreement module into the encoder to obtain a better representation of the source sentence. The experiments are conducted on three translation datasets: the WMT 2014 EN → DE task, the WMT 2016 EN → RO task, and the IWSLT 2014 DE → DE task, and the improvement in the NAT model's performance shows the effect of our proposed method. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
33. Self-attention Comparison Module for Boosting Performance on Retrieval-based Open-Domain Dialog Systems
- Author
-
Lan, Tian, Mao, Xian-Ling, Zhao, Zhipeng, Wei, Wei, and Huang, Heyan
- Subjects
FOS: Computer and information sciences ,Computer Science - Computation and Language ,Computation and Language (cs.CL) - Abstract
Since the pre-trained language models are widely used, retrieval-based open-domain dialog systems, have attracted considerable attention from researchers recently. Most of the previous works select a suitable response only according to the matching degree between the query and each individual candidate response. Although good performance has been achieved, these recent works ignore the comparison among the candidate responses, which could provide rich information for selecting the most appropriate response. Intuitively, better decisions could be made when the models can get access to the comparison information among all the candidate responses. In order to leverage the comparison information among the candidate responses, in this paper, we propose a novel and plug-in Self-attention Comparison Module for retrieval-based open-domain dialog systems, called SCM. Extensive experiment results demonstrate that our proposed self-attention comparison module effectively boosts the performance of the existing retrieval-based open-domain dialog systems. Besides, we have publicly released our source codes for future research.
- Published
- 2020
34. Deep Cross-modal Hashing via Margin-dynamic-softmax Loss
- Author
-
Tu, Rong-Cheng, Mao, Xian-Ling, Tu, Rongxin, Bian, Binbin, Wei, Wei, and Huang, Heyan
- Subjects
FOS: Computer and information sciences ,Computer Vision and Pattern Recognition (cs.CV) ,Computer Science - Computer Vision and Pattern Recognition ,Information Retrieval (cs.IR) ,Computer Science - Multimedia ,Multimedia (cs.MM) ,Computer Science - Information Retrieval - Abstract
Due to their high retrieval efficiency and low storage cost for cross-modal search task, cross-modal hashing methods have attracted considerable attention. For the supervised cross-modal hashing methods, how to make the learned hash codes preserve semantic information sufficiently contained in the label of datapoints is the key to further enhance the retrieval performance. Hence, almost all supervised cross-modal hashing methods usually depends on defining a similarity between datapoints with the label information to guide the hashing model learning fully or partly. However, the defined similarity between datapoints can only capture the label information of datapoints partially and misses abundant semantic information, then hinders the further improvement of retrieval performance. Thus, in this paper, different from previous works, we propose a novel cross-modal hashing method without defining the similarity between datapoints, called Deep Cross-modal Hashing via \textit{Margin-dynamic-softmax Loss} (DCHML). Specifically, DCHML first trains a proxy hashing network to transform each category information of a dataset into a semantic discriminative hash code, called proxy hash code. Each proxy hash code can preserve the semantic information of its corresponding category well. Next, without defining the similarity between datapoints to supervise the training process of the modality-specific hashing networks , we propose a novel \textit{margin-dynamic-softmax loss} to directly utilize the proxy hashing codes as supervised information. Finally, by minimizing the novel \textit{margin-dynamic-softmax loss}, the modality-specific hashing networks can be trained to generate hash codes which can simultaneously preserve the cross-modal similarity and abundant semantic information well.
- Published
- 2020
35. Which Kind Is Better in Open-domain Multi-turn Dialog,Hierarchical or Non-hierarchical Models? An Empirical Study
- Author
-
Lan, Tian, Mao, Xian-Ling, Wei, Wei, and Huang, Heyan
- Subjects
FOS: Computer and information sciences ,Computer Science - Machine Learning ,Computer Science - Computation and Language ,Statistics - Machine Learning ,Machine Learning (stat.ML) ,Computation and Language (cs.CL) ,Machine Learning (cs.LG) - Abstract
Currently, open-domain generative dialog systems have attracted considerable attention in academia and industry. Despite the success of single-turn dialog generation, multi-turn dialog generation is still a big challenge. So far, there are two kinds of models for open-domain multi-turn dialog generation: hierarchical and non-hierarchical models. Recently, some works have shown that the hierarchical models are better than non-hierarchical models under their experimental settings; meanwhile, some works also demonstrate the opposite conclusion. Due to the lack of adequate comparisons, it's not clear which kind of models are better in open-domain multi-turn dialog generation. Thus, in this paper, we will measure systematically nearly all representative hierarchical and non-hierarchical models over the same experimental settings to check which kind is better. Through extensive experiments, we have the following three important conclusions: (1) Nearly all hierarchical models are worse than non-hierarchical models in open-domain multi-turn dialog generation, except for the HRAN model. Through further analysis, the excellent performance of HRAN mainly depends on its word-level attention mechanism; (2) The performance of other hierarchical models will also obtain a great improvement if integrating the word-level attention mechanism into these models. The modified hierarchical models even significantly outperform the non-hierarchical models; (3) The reason why the word-level attention mechanism is so powerful for hierarchical models is because it can leverage context information more effectively, especially the fine-grained information. Besides, we have implemented all of the models and already released the codes.
- Published
- 2020
36. VSRN: Visual-Semantic Relation Network for Video Visual Relation Inference.
- Author
-
Cao, Qianwen and Huang, Heyan
- Subjects
- *
DEEP learning , *FEATURE extraction - Abstract
Video visual relation inference refers to the task of automatically detecting the relation triplets between the observed objects in videos with the form of ${ < subject, predicate, object>}$ , which requires correctly labeling each detected object and their interaction predicates. Despite the recent advances in image visual relation detection using deep learning techniques, relation inference in videos remains a challenging topic. On one hand, since the introduction of temporal information, it needs to model the rich spatio-temporal visual information for objects and videos. On the other hand, wild videos are often annotated with incomplete relation triplet tags and a few of them are semantically overlapped. However, previous methods adopt hand-crafted visual features extracted from the trajectories, describing local appearance characteristics of isolated objects. And they treat the problem as a multi-class classification task, which makes the relation tags mutually exclusive. To address the above issues, we propose a novel model, termed Visual-Semantic Relation Network (VSRN). In this network, we leverage three-dimensional convolution kernel to capture spatio-temporal features, and encode global visual features in videos through pooling operation on each time slice. Moreover, the semantic collocations between objects are also incorporated so as to obtain comprehensive representations of the relationships. For relation classification, we treat the problem as a multi-label classification task and regard each tag to be independent to predict various relationships. Additionally, we modify commonly used evaluation metric, video-wise recall, to a pair-wise metric (Roop) for testing the performance of models in predicting multiple relationships for the object pairs, Extensive experimental results on two large-scale datasets demonstrate the effectiveness of our proposed model which significantly outperforms the previous works. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
37. Deep Cross-Modal Hashing With Hashing Functions and Unified Hash Codes Jointly Learning.
- Author
-
Tu, Rong-Cheng, Mao, Xian-Ling, Ma, Bing, Hu, Yong, Yan, Tan, Wei, Wei, and Huang, Heyan
- Subjects
COMPUTER programming education ,MATHEMATICAL optimization ,BINARY codes - Abstract
Due to their high retrieval efficiency and low storage cost, cross-modal hashing methods have attracted considerable attention. Generally, compared with shallow cross-modal hashing methods, deep cross-modal hashing methods can achieve a more satisfactory performance by integrating feature learning and hash codes optimizing into a same framework. However, most existing deep cross-modal hashing methods either cannot learn a unified hash code for the two correlated data-points of different modalities in a database instance or cannot guide the learning of unified hash codes by the feedback of hashing function learning procedure, to enhance the retrieval accuracy. To address the issues above, in this paper, we propose a novel end-to-end Deep Cross-Modal Hashing with Hashing Functions and Unified Hash Codes Jointly Learning (DCHUC). Specifically, by an iterative optimization algorithm, DCHUC jointly learns unified hash codes for image-text pairs in a database and a pair of hash functions for unseen query image-text pairs. With the iterative optimization algorithm, the learned unified hash codes can be used to guide the hashing function learning procedure; Meanwhile, the learned hashing functions can feedback to guide the unified hash codes optimizing procedure. Extensive experiments on three public datasets demonstrate that the proposed method outperforms the state-of-the-art cross-modal hashing methods. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
38. Generating Informative Dialogue Responses with Keywords-Guided Networks
- Author
-
Xu, Heng-Da, Mao, Xian-Ling, Chi, Zewen, Zhu, Jing-Jing, Sun, Fanshu, and Huang, Heyan
- Subjects
FOS: Computer and information sciences ,Computer Science - Computation and Language ,Computation and Language (cs.CL) - Abstract
Recently, open-domain dialogue systems have attracted growing attention. Most of them use the sequence-to-sequence (Seq2Seq) architecture to generate responses. However, traditional Seq2Seq-based open-domain dialogue models tend to generate generic and safe responses, which are less informative, unlike human responses. In this paper, we propose a simple but effective keywords-guided Sequence-to-Sequence model (KW-Seq2Seq) which uses keywords information as guidance to generate open-domain dialogue responses. Specifically, KW-Seq2Seq first uses a keywords decoder to predict some topic keywords, and then generates the final response under the guidance of them. Extensive experiments demonstrate that the KW-Seq2Seq model produces more informative, coherent and fluent responses, yielding substantive gain in both automatic and human evaluation metrics.
- Published
- 2020
39. Context-Sensitive Generation Network for Handing Unknown Slot Values in Dialogue State Tracking
- Author
-
Yang, Puhai, Huang, Heyan, and Mao, Xian-Ling
- Subjects
FOS: Computer and information sciences ,Computer Science - Computation and Language ,Computation and Language (cs.CL) - Abstract
As a key component in a dialogue system, dialogue state tracking plays an important role. It is very important for dialogue state tracking to deal with the problem of unknown slot values. As far as we known, almost all existing approaches depend on pointer network to solve the unknown slot value problem. These pointer network-based methods usually have a hidden assumption that there is at most one out-of-vocabulary word in an unknown slot value because of the character of a pointer network. However, often, there are multiple out-of-vocabulary words in an unknown slot value, and it makes the existing methods perform bad. To tackle the problem, in this paper, we propose a novel Context-Sensitive Generation network (CSG) which can facilitate the representation of out-of-vocabulary words when generating the unknown slot value. Extensive experiments show that our proposed method performs better than the state-of-the-art baselines.
- Published
- 2020
40. Learning Relation Ties with a Force-Directed Graph in Distant Supervised Relation Extraction
- Author
-
Shang, Yuming, Huang, Heyan, Sun, Xin, and Mao, Xianling
- Subjects
FOS: Computer and information sciences ,Computer Science - Computation and Language ,Computation and Language (cs.CL) - Abstract
Relation ties, defined as the correlation and mutual exclusion between different relations, are critical for distant supervised relation extraction. Existing approaches model this property by greedily learning local dependencies. However, they are essentially limited by failing to capture the global topology structure of relation ties. As a result, they may easily fall into a locally optimal solution. To solve this problem, in this paper, we propose a novel force-directed graph based relation extraction model to comprehensively learn relation ties. Specifically, we first build a graph according to the global co-occurrence of relations. Then, we borrow the idea of Coulomb's Law from physics and introduce the concept of attractive force and repulsive force to this graph to learn correlation and mutual exclusion between relations. Finally, the obtained relation representations are applied as an inter-dependent relation classifier. Experimental results on a large scale benchmark dataset demonstrate that our model is capable of modeling global relation ties and significantly outperforms other baselines. Furthermore, the proposed force-directed graph can be used as a module to augment existing relation extraction systems and improve their performance., Learning Relation Ties
- Published
- 2020
41. Evaluating Low-Resource Machine Translation between Chinese and Vietnamese with Back-Translation
- Author
-
Li, Hongzheng and Huang, Heyan
- Subjects
FOS: Computer and information sciences ,Computer Science - Computation and Language ,Computation and Language (cs.CL) - Abstract
Back translation (BT) has been widely used and become one of standard techniques for data augmentation in Neural Machine Translation (NMT), BT has proven to be helpful for improving the performance of translation effectively, especially for low-resource scenarios. While most works related to BT mainly focus on European languages, few of them study languages in other areas around the world. In this paper, we investigate the impacts of BT on Asia language translations between the extremely low-resource Chinese and Vietnamese language pair. We evaluate and compare the effects of different sizes of synthetic data on both NMT and Statistical Machine Translation (SMT) models for Chinese to Vietnamese and Vietnamese to Chinese, with character-based and word-based settings. Some conclusions from previous works are partially confirmed and we also draw some other interesting findings and conclusions, which are beneficial to understand BT further.
- Published
- 2020
42. Neural Chinese Word Segmentation as Sequence to Sequence Translation
- Author
-
Shi, Xuewen, Huang, Heyan, Jian, Ping, Guo, Yuhang, Wei, Xiaochi, and Tang, Yi-Kun
- Subjects
FOS: Computer and information sciences ,Computer Science - Computation and Language ,Computation and Language (cs.CL) - Abstract
Recently, Chinese word segmentation (CWS) methods using neural networks have made impressive progress. Most of them regard the CWS as a sequence labeling problem which construct models based on local features rather than considering global information of input sequence. In this paper, we cast the CWS as a sequence translation problem and propose a novel sequence-to-sequence CWS model with an attention-based encoder-decoder framework. The model captures the global information from the input and directly outputs the segmented sequence. It can also tackle other NLP tasks with CWS jointly in an end-to-end mode. Experiments on Weibo, PKU and MSRA benchmark datasets show that our approach has achieved competitive performances compared with state-of-the-art methods. Meanwhile, we successfully applied our proposed model to jointly learning CWS and Chinese spelling correction, which demonstrates its applicability of multi-task fusion., In proceedings of SMP 2017 (Chinese National Conference on Social Media Processing)
- Published
- 2019
43. Tag Recommendation by Word-Level Tag Sequence Modeling
- Author
-
Shi, Xuewen, Huang, Heyan, Zhao, Shuyang, Jian, Ping, and Tang, Yi-Kun
- Subjects
FOS: Computer and information sciences ,Computer Science - Computation and Language ,Computation and Language (cs.CL) - Abstract
In this paper, we transform tag recommendation into a word-based text generation problem and introduce a sequence-to-sequence model. The model inherits the advantages of LSTM-based encoder for sequential modeling and attention-based decoder with local positional encodings for learning relations globally. Experimental results on Zhihu datasets illustrate the proposed model outperforms other state-of-the-art text classification based methods., This is a full length version of the paper in DASFAA 2019
- Published
- 2019
44. Can Monolingual Pretrained Models Help Cross-Lingual Classification?
- Author
-
Chi, Zewen, Dong, Li, Wei, Furu, Mao, Xian-Ling, and Huang, Heyan
- Subjects
FOS: Computer and information sciences ,Computer Science - Computation and Language ,Computation and Language (cs.CL) - Abstract
Multilingual pretrained language models (such as multilingual BERT) have achieved impressive results for cross-lingual transfer. However, due to the constant model capacity, multilingual pre-training usually lags behind the monolingual competitors. In this work, we present two approaches to improve zero-shot cross-lingual classification, by transferring the knowledge from monolingual pretrained models to multilingual ones. Experimental results on two cross-lingual classification benchmarks show that our methods outperform vanilla multilingual fine-tuning., 5 pages
- Published
- 2019
45. SEPT: Improving Scientific Named Entity Recognition with Span Representation
- Author
-
Yan, Tan, Huang, Heyan, and Mao, Xian-Ling
- Subjects
FOS: Computer and information sciences ,Computer Science - Computation and Language ,Computation and Language (cs.CL) ,Information Retrieval (cs.IR) ,Computer Science - Information Retrieval - Abstract
We introduce a new scientific named entity recognizer called SEPT, which stands for Span Extractor with Pre-trained Transformers. In recent papers, span extractors have been demonstrated to be a powerful model compared with sequence labeling models. However, we discover that with the development of pre-trained language models, the performance of span extractors appears to become similar to sequence labeling models. To keep the advantages of span representation, we modified the model by under-sampling to balance the positive and negative samples and reduce the search space. Furthermore, we simplify the origin network architecture to combine the span extractor with BERT. Experiments demonstrate that even simplified architecture achieves the same performance and SEPT achieves a new state of the art result in scientific named entity recognition even without relation information involved., This work is outdated. The result should not be trusted
- Published
- 2019
46. Hadamard 'Pipeline' Coding Computational Ghost Imaging
- Author
-
Zhou, Cheng, Zhao, Xiwei, Huang, Heyan, Wang, Gangcheng, Wang, Xue, Song, Lijun, and Xue, Kang
- Subjects
TheoryofComputation_ANALYSISOFALGORITHMSANDPROBLEMCOMPLEXITY ,Image and Video Processing (eess.IV) ,FOS: Electrical engineering, electronic engineering, information engineering ,Mathematics::Classical Analysis and ODEs ,FOS: Physical sciences ,Data_CODINGANDINFORMATIONTHEORY ,Electrical Engineering and Systems Science - Image and Video Processing ,Optics (physics.optics) ,MathematicsofComputing_DISCRETEMATHEMATICS ,Physics - Optics - Abstract
The Hadamard matrix with orthogonality is a more important modulation matrix for computational ghost imaging (CGI), especially its optimized Hadamard matrix. However, as far as we know, little mention has been paid to efficient and convenient Hadamard matrix generation for CGI. The existing methods are to reconstruct any row of Hadamard matrix into two-dimensional matrix and then optimize it. In this work, we propose a Hadamard `pipeline' coding computational ghost imaging approach, which can directly generate two-dimensional Hadamard derived pattern and Hadamard optimization sequence, whereby both the memory consumption and the complexity of coding implementation for CGI can be significantly reduced. The optimization method of commonly used hadamard optimization sequence implementation is also discussed. This method provides a new approach for Hadamard sequence optimization and ghost imaging applications.
- Published
- 2019
47. Generative Dialog Policy for Task-oriented Dialog Systems
- Author
-
Lan, Tian, Mao, Xianling, and Huang, Heyan
- Subjects
FOS: Computer and information sciences ,Artificial Intelligence (cs.AI) ,InformationSystems_MODELSANDPRINCIPLES ,Computer Science - Computation and Language ,Computer Science - Artificial Intelligence ,Computation and Language (cs.CL) - Abstract
There is an increasing demand for task-oriented dialogue systems which can assist users in various activities such as booking tickets and restaurant reservations. In order to complete dialogues effectively, dialogue policy plays a key role in task-oriented dialogue systems. As far as we know, the existing task-oriented dialogue systems obtain the dialogue policy through classification, which can assign either a dialogue act and its corresponding parameters or multiple dialogue acts without their corresponding parameters for a dialogue action. In fact, a good dialogue policy should construct multiple dialogue acts and their corresponding parameters at the same time. However, it's hard for existing classification-based methods to achieve this goal. Thus, to address the issue above, we propose a novel generative dialogue policy learning method. Specifically, the proposed method uses attention mechanism to find relevant segments of given dialogue context and input utterance and then constructs the dialogue policy by a seq2seq way for task-oriented dialogue systems. Extensive experiments on two benchmark datasets show that the proposed model significantly outperforms the state-of-the-art baselines. In addition, we have publicly released our codes.
- Published
- 2019
48. Multi-task Learning for Low-resource Second Language Acquisition Modeling
- Author
-
Hu, Yong, Huang, Heyan, Lan, Tian, Wei, Xiaochi, Nie, Yuxiang, Qi, Jiarui, Yang, Liner, and Mao, Xian-Ling
- Subjects
FOS: Computer and information sciences ,Computer Science - Computation and Language ,Computation and Language (cs.CL) - Abstract
Second language acquisition (SLA) modeling is to predict whether second language learners could correctly answer the questions according to what they have learned. It is a fundamental building block of the personalized learning system and has attracted more and more attention recently. However, as far as we know, almost all existing methods cannot work well in low-resource scenarios due to lacking of training data. Fortunately, there are some latent common patterns among different language-learning tasks, which gives us an opportunity to solve the low-resource SLA modeling problem. Inspired by this idea, in this paper, we propose a novel SLA modeling method, which learns the latent common patterns among different language-learning datasets by multi-task learning and are further applied to improving the prediction performance in low-resource scenarios. Extensive experiments show that the proposed method performs much better than the state-of-the-art baselines in the low-resource scenario. Meanwhile, it also obtains improvement slightly in the non-low-resource scenario.
- Published
- 2019
49. Complicated Table Structure Recognition
- Author
-
Chi, Zewen, Huang, Heyan, Xu, Heng-Da, Yu, Houjin, Yin, Wanxuan, and Mao, Xian-Ling
- Subjects
FOS: Computer and information sciences ,Computer Science - Machine Learning ,Information Retrieval (cs.IR) ,Machine Learning (cs.LG) ,Computer Science - Information Retrieval - Abstract
The task of table structure recognition aims to recognize the internal structure of a table, which is a key step to make machines understand tables. Currently, there are lots of studies on this task for different file formats such as ASCII text and HTML. It also attracts lots of attention to recognize the table structures in PDF files. However, it is hard for the existing methods to accurately recognize the structure of complicated tables in PDF files. The complicated tables contain spanning cells which occupy at least two columns or rows. To address the issue, we propose a novel graph neural network for recognizing the table structure in PDF files, named GraphTSR. Specifically, it takes table cells as input, and then recognizes the table structures by predicting relations among cells. Moreover, to evaluate the task better, we construct a large-scale table structure recognition dataset from scientific papers, named SciTSR, which contains 15,000 tables from PDF files and their corresponding structure labels. Extensive experiments demonstrate that our proposed model is highly effective for complicated tables and outperforms state-of-the-art baselines over a benchmark dataset and our new constructed dataset.
- Published
- 2019
50. A Parallel Recurrent Neural Network for Language Modeling with POS Tags
- Author
-
Su, Chao, Huang, Heyan, Shi, Shumin, Guo, Yuhang, and Wu, Hao
- Published
- 2017
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.