12 results on '"Pengcheng Yang"'
Search Results
2. Inductively Representing Out-of-Knowledge-Graph Entities by Optimal Estimation Under Translational Assumptions
- Author
-
Damai Dai, Baobao Chang, Fuli Luo, Pengcheng Yang, Zhifang Sui, Tianyu Liu, and Hua Zheng
- Subjects
Theoretical computer science ,Optimal estimation ,Knowledge graph ,Simple (abstract algebra) ,Computer science ,Margin (machine learning) ,Embedding ,Effective method - Abstract
Conventional Knowledge Graph Completion (KGC) assumes that all test entities appear during training. However, in real-world scenarios, Knowledge Graphs (KG) evolve fast with out-of-knowledge-graph (OOKG) entities added frequently, and we need to efficiently represent these entities. Most existing Knowledge Graph Embedding (KGE) methods cannot represent OOKG entities without costly retraining on the whole KG. To enhance efficiency, we propose a simple and effective method that inductively represents OOKG entities by their optimal estimation under translational assumptions. Moreover, given pretrained embeddings of the in-knowledge-graph (IKG) entities, our method even needs no additional learning. Experimental results on two KGC tasks with OOKG entities show that our method outperforms the previous methods by a large margin with higher efficiency.
- Published
- 2021
3. Multi-Granularity Contrasting for Cross-Lingual Pre-Training
- Author
-
Shicheng Li, Jun Xie, Pengcheng Yang, and Fuli Luo
- Subjects
Cross lingual ,business.industry ,Computer science ,Training (meteorology) ,Artificial intelligence ,Granularity ,computer.software_genre ,business ,computer ,Natural language processing - Published
- 2021
4. Context-Interactive Pre-Training for Document Machine Translation
- Author
-
Pengcheng Yang, Jun Xie, Pei Zhang, Boxing Chen, and Weihua Luo
- Subjects
Dependency (UML) ,Machine translation ,business.industry ,Computer science ,Context (language use) ,Translation (geometry) ,computer.software_genre ,Variety (cybernetics) ,Benchmark (computing) ,Contextual information ,Artificial intelligence ,business ,computer ,Sentence ,Natural language processing - Abstract
Document machine translation aims to translate the source sentence into the target language in the presence of additional contextual information. However, it typically suffers from a lack of doc-level bilingual data. To remedy this, here we propose a simple yet effective context-interactive pre-training approach, which targets benefiting from external large-scale corpora. The proposed model performs inter sentence generation to capture the cross-sentence dependency within the target document, and cross sentence translation to make better use of valuable contextual information. Comprehensive experiments illustrate that our approach can achieve state-of-the-art performance on three benchmark datasets, which significantly outperforms a variety of baselines.
- Published
- 2021
5. MAAM: A Morphology-Aware Alignment Model for Unsupervised Bilingual Lexicon Induction
- Author
-
Pengcheng Yang, Fuli Luo, Xu Sun, Tianyu Liu, and Peng Chen
- Subjects
Morphology (linguistics) ,Computer science ,business.industry ,02 engineering and technology ,Translation (geometry) ,computer.software_genre ,Task (project management) ,Bilingual lexicon ,03 medical and health sciences ,ComputingMethodologies_PATTERNRECOGNITION ,0302 clinical medicine ,030221 ophthalmology & optometry ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Language model ,Artificial intelligence ,business ,computer ,Word (computer architecture) ,Natural language processing - Abstract
The task of unsupervised bilingual lexicon induction (UBLI) aims to induce word translations from monolingual corpora in two languages. Previous work has shown that morphological variation is an intractable challenge for the UBLI task, where the induced translation in failure case is usually morphologically related to the correct translation. To tackle this challenge, we propose a morphology-aware alignment model for the UBLI task. The proposed model aims to alleviate the adverse effect of morphological variation by introducing grammatical information learned by the pre-trained denoising language model. Results show that our approach can substantially outperform several state-of-the-art unsupervised systems, and even achieves competitive performance compared to supervised methods.
- Published
- 2019
6. Towards Comprehensive Description Generation from Factual Attribute-value Tables
- Author
-
Pengcheng Yang, Zhifang Sui, Baobao Chang, Wei Wu, Tianyu Liu, and Fuli Luo
- Subjects
Information retrieval ,Data model ,Computer science ,020204 information systems ,Benchmark (surveying) ,0202 electrical engineering, electronic engineering, information engineering ,Key (cryptography) ,Reinforcement learning ,020201 artificial intelligence & image processing ,02 engineering and technology ,Value (mathematics) ,Generator (mathematics) - Abstract
The comprehensive descriptions for factual attribute-value tables, which should be accurate, informative and loyal, can be very helpful for end users to understand the structured data in this form. However previous neural generators might suffer from key attributes missing, less informative and groundless information problems, which impede the generation of high-quality comprehensive descriptions for tables. To relieve these problems, we first propose force attention (FA) method to encourage the generator to pay more attention to the uncovered attributes to avoid potential key attributes missing. Furthermore, we propose reinforcement learning for information richness to generate more informative as well as more loyal descriptions for tables. In our experiments, we utilize the widely used WIKIBIO dataset as a benchmark. Besides, we create WB-filter based on WIKIBIO to test our model in the simulated user-oriented scenarios, in which the generated descriptions should accord with particular user interests. Experimental results show that our model outperforms the state-of-the-art baselines on both automatic and human evaluation.
- Published
- 2019
7. Enhancing Topic-to-Essay Generation with External Commonsense Knowledge
- Author
-
Tianyu Liu, Fuli Luo, Xu Sun, Pengcheng Yang, and Lei Li
- Subjects
Commonsense knowledge ,business.industry ,Computer science ,Mechanism (biology) ,media_common.quotation_subject ,02 engineering and technology ,010501 environmental sciences ,01 natural sciences ,Data science ,Task (project management) ,Knowledge base ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Quality (business) ,business ,Set (psychology) ,0105 earth and related environmental sciences ,Generator (mathematics) ,media_common - Abstract
Automatic topic-to-essay generation is a challenging task since it requires generating novel, diverse, and topic-consistent paragraph-level text with a set of topics as input. Previous work tends to perform essay generation based solely on the given topics while ignoring massive commonsense knowledge. However, this commonsense knowledge provides additional background information, which can help to generate essays that are more novel and diverse. Towards filling this gap, we propose to integrate commonsense from the external knowledge base into the generator through dynamic memory mechanism. Besides, the adversarial training based on a multi-label discriminator is employed to further improve topic-consistency. We also develop a series of automatic evaluation metrics to comprehensively assess the quality of the generated essay. Experiments show that with external commonsense knowledge and adversarial training, the generated essays are more novel, diverse, and topic-consistent than existing methods in terms of both automatic and human evaluation.
- Published
- 2019
8. Cross-Modal Commentator: Automatic Machine Commenting Based on Cross-Modal Information
- Author
-
Pengcheng Yang, Chengyang Huang, Lei Li, Zhihan Zhang, Xu Sun, and Fuli Luo
- Subjects
Information retrieval ,Dependency (UML) ,Computer science ,business.industry ,02 engineering and technology ,Construct (python library) ,010501 environmental sciences ,01 natural sciences ,Task (project management) ,Modal ,User experience design ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Social media ,business ,0105 earth and related environmental sciences - Abstract
Automatic commenting of online articles can provide additional opinions and facts to the reader, which improves user experience and engagement on social media platforms. Previous work focuses on automatic commenting based solely on textual content. However, in real-scenarios, online articles usually contain multiple modal contents. For instance, graphic news contains plenty of images in addition to text. Contents other than text are also vital because they are not only more attractive to the reader but also may provide critical information. To remedy this, we propose a new task: cross-model automatic commenting (CMAC), which aims to make comments by integrating multiple modal contents. We construct a large-scale dataset for this task and explore several representative methods. Going a step further, an effective co-attention model is presented to capture the dependency between textual and visual information. Evaluation results show that our proposed model can achieve better performance than competitive baselines.
- Published
- 2019
9. A Deep Reinforced Sequence-to-Set Model for Multi-Label Classification
- Author
-
Qi Su, Yi Zhang, Xu Sun, Pengcheng Yang, Shuming Ma, and Junyang Lin
- Subjects
Sequence ,business.industry ,Computer science ,Maximum likelihood ,Pattern recognition ,02 engineering and technology ,Task (project management) ,Set (abstract data type) ,ComputingMethodologies_PATTERNRECOGNITION ,Simple (abstract algebra) ,Order (business) ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Reinforcement learning ,020201 artificial intelligence & image processing ,Artificial intelligence ,Sensitivity (control systems) ,business - Abstract
Multi-label classification (MLC) aims to predict a set of labels for a given instance. Based on a pre-defined label order, the sequence-to-sequence (Seq2Seq) model trained via maximum likelihood estimation method has been successfully applied to the MLC task and shows powerful ability to capture high-order correlations between labels. However, the output labels are essentially an unordered set rather than an ordered sequence. This inconsistency tends to result in some intractable problems, e.g., sensitivity to the label order. To remedy this, we propose a simple but effective sequence-to-set model. The proposed model is trained via reinforcement learning, where reward feedback is designed to be independent of the label order. In this way, we can reduce the dependence of the model on the label order, as well as capture high-order correlations between labels. Extensive experiments show that our approach can substantially outperform competitive baselines, as well as effectively reduce the sensitivity to the label order.
- Published
- 2019
10. Learning to Control the Fine-grained Sentiment for Story Ending Generation
- Author
-
Damai Dai, Xu Sun, Baobao Chang, Tianyu Liu, Pengcheng Yang, Zhifang Sui, and Fuli Luo
- Subjects
0209 industrial biotechnology ,Focus (computing) ,business.industry ,Computer science ,InformationSystems_INFORMATIONSTORAGEANDRETRIEVAL ,Natural language generation ,02 engineering and technology ,computer.software_genre ,ComputingMethodologies_ARTIFICIALINTELLIGENCE ,Task (project management) ,020901 industrial engineering & automation ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Artificial intelligence ,InformationSystems_MISCELLANEOUS ,Layer (object-oriented design) ,Control (linguistics) ,business ,computer ,Natural language processing ,Generator (mathematics) - Abstract
Automatic story ending generation is an interesting and challenging task in natural language generation. Previous studies are mainly limited to generate coherent, reasonable and diversified story endings, and few works focus on controlling the sentiment of story endings. This paper focuses on generating a story ending which meets the given fine-grained sentiment intensity. There are two major challenges to this task. First is the lack of story corpus which has fine-grained sentiment labels. Second is the difficulty of explicitly controlling sentiment intensity when generating endings. Therefore, we propose a generic and novel framework which consists of a sentiment analyzer and a sentimental generator, respectively addressing the two challenges. The sentiment analyzer adopts a series of methods to acquire sentiment intensities of the story dataset. The sentimental generator introduces the sentiment intensity into decoder via a Gaussian Kernel Layer to control the sentiment of the output. To the best of our knowledge, this is the first endeavor to control the fine-grained sentiment for story ending generation without manually annotating sentiment labels. Experiments show that our proposed framework can generate story endings which are not only more coherent and fluent but also able to meet the given sentiment intensity better.
- Published
- 2019
11. Towards Fine-grained Text Sentiment Transfer
- Author
-
Xu Sun, Baobao Chang, Yutong Tan, Jie Zhou, Peng Li, Fuli Luo, Pengcheng Yang, and Zhifang Sui
- Subjects
Sequence ,business.industry ,Computer science ,02 engineering and technology ,010501 environmental sciences ,Machine learning ,computer.software_genre ,01 natural sciences ,Task (project management) ,symbols.namesake ,Transformation (function) ,Margin (machine learning) ,0202 electrical engineering, electronic engineering, information engineering ,Gaussian function ,symbols ,020201 artificial intelligence & image processing ,Artificial intelligence ,Layer (object-oriented design) ,business ,Focus (optics) ,computer ,0105 earth and related environmental sciences - Abstract
In this paper, we focus on the task of fine-grained text sentiment transfer (FGST). This task aims to revise an input sequence to satisfy a given sentiment intensity, while preserving the original semantic content. Different from the conventional sentiment transfer task that only reverses the sentiment polarity (positive/negative) of text, the FTST task requires more nuanced and fine-grained control of sentiment. To remedy this, we propose a novel Seq2SentiSeq model. Specifically, the numeric sentiment intensity value is incorporated into the decoder via a Gaussian kernel layer to finely control the sentiment intensity of the output. Moreover, to tackle the problem of lacking parallel data, we propose a cycle reinforcement learning algorithm to guide the model training. In this framework, the elaborately designed rewards can balance both sentiment transformation and content preservation, while not requiring any ground truth output. Experimental results show that our approach can outperform existing methods by a large margin in both automatic evaluation and human evaluation.
- Published
- 2019
12. Learning Sentiment Memories for Sentiment Modification without Parallel Data
- Author
-
Yi Zhang, Xu Sun, Jingjing Xu, and Pengcheng Yang
- Subjects
FOS: Computer and information sciences ,Computer Science - Computation and Language ,Computer science ,business.industry ,InformationSystems_INFORMATIONSTORAGEANDRETRIEVAL ,Context (language use) ,02 engineering and technology ,010501 environmental sciences ,computer.software_genre ,ComputingMethodologies_ARTIFICIALINTELLIGENCE ,01 natural sciences ,Task (project management) ,Transformation (function) ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Artificial intelligence ,InformationSystems_MISCELLANEOUS ,business ,Computation and Language (cs.CL) ,computer ,Natural language processing ,0105 earth and related environmental sciences - Abstract
The task of sentiment modification requires reversing the sentiment of the input and preserving the sentiment-independent content. However, aligned sentences with the same content but different sentiments are usually unavailable. Due to the lack of such parallel data, it is hard to extract sentiment independent content and reverse the sentiment in an unsupervised way. Previous work usually can not reconcile sentiment transformation and content preservation. In this paper, motivated by the fact the non-emotional context (e.g., "staff") provides strong cues for the occurrence of emotional words (e.g., "friendly"), we propose a novel method that automatically extracts appropriate sentiment information from learned sentiment memories according to specific context. Experiments show that our method substantially improves the content preservation degree and achieves the state-of-the-art performance., Comment: Accepted by EMNLP 2018
- Published
- 2018
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.