17 results on '"Haotian Zhang"'
Search Results
2. Learning to balance exploration and exploitation in pareto local search for multi-objective combinatorial optimization
- Author
-
Haotian Zhang, Jialong Shi, Jianyong Sun, and Zongben Xu
- Published
- 2022
3. One size does not fit all: security hardening of MIPS embedded systems via static binary debloating for shared libraries
- Author
-
Haotian Zhang, Mengfei Ren, Yu Lei, and Jiang Ming
- Published
- 2022
4. Personalized Multi-modal Video Retrieval on Mobile Devices
- Author
-
Iqbal Mohomed, Konstantinos G. Derpanis, Afsaneh Fazly, Ran Zhang, Allan D. Jepson, and Haotian Zhang
- Subjects
Matching (statistics) ,Information retrieval ,Computer science ,Face (geometry) ,Process (computing) ,Proper noun ,Mobile device ,Natural language ,Personalization ,Ranking (information retrieval) - Abstract
Current video retrieval systems on mobile devices cannot process complex natural language queries, especially if they contain personalized concepts, such as proper names. To address these shortcomings, we propose an efficient and privacy-preserving video retrieval system that works well with personalized queries containing proper names, without re-training using personalized labelled data from users. Our system first computes an initial ranking of a video collection by using a generic attention-based video-text matching model (i.e., a model designed for non-personalized queries), and then uses a face detector to conduct personalized adjustments to these initial rankings. These adjustments are done by reasoning over the face information from the detector and the attention information provided by the generic model. We show that our system significantly outperforms existing keyword-based retrieval systems, and achieves comparable performance to the generic matching model fine-tuned on plenty of labelled data. Our results suggest that the proposed system can effectively capture both semantic context and personalized information in queries.
- Published
- 2021
5. ROD2021 Challenge
- Author
-
Hung-Min Hsu, Renshu Gu, Zhongyu Jiang, Kwang-Ju Kim, Hui Liu, Gaoang Wang, Jiarui Cai, Yizhou Wang, Jenq-Neng Hwang, and Haotian Zhang
- Subjects
Continuous-wave radar ,Artificial neural network ,Computer science ,law ,Human–computer interaction ,Benchmark (computing) ,Radar ,Object perception ,Object detection ,Task (project management) ,law.invention - Abstract
The Radar Object Detection 2021 (ROD2021) Challenge, held in the ACM International Conference on Multimedia Retrieval (ICMR) 2021, has been introduced to detect and classify objects purely using an FMCW radar for autonomous driving applications. As a robust sensor to all-weather conditions, radar has rich information hidden in the radio frequencies, which can potentially achieve object detection and classification. This insight will provide a new object perception solution for an autonomous vehicle even in adverse driving scenarios. The ROD2021 Challenge is the first public benchmark focusing on this topic, which attracts great attention and participation. There are more than 260 participants among 37 teams from more than 10 countries with different academic and industrial affiliations, contributing about 300 submissions in the first phase and 400 submissions in the second phase. The final performance is evaluated by average precision (AP). Results add strong value and a better understanding of the radar object detection task for the autonomous vehicle community.
- Published
- 2021
6. Analysis of Faces in a Decade of US Cable TV News
- Author
-
Jeremy Barenholtz, Jacob Ritchie, Xinwei Yao, Kayvon Fatahalian, Michaela Murray, Will Crichton, Daniel Y. Fu, Geraldine Moriba, Maneesh Agrawala, Haotian Zhang, James Hong, and Ben Hannel
- Subjects
World Wide Web ,Quantitative analysis (finance) ,Computer science ,Scale (social sciences) ,Media monitoring ,Public interface ,Face (sociological concept) ,Set (psychology) ,Analysis method - Abstract
Cable (TV) news reaches millions of US households each day. News stakeholders such as communications researchers, journalists, and media monitoring organizations are interested in the visual content of cable news, especially who is on-screen. Manual analysis, however, is labor intensive and limits the size of prior studies. We conduct a large-scale, quantitative analysis of the faces in a decade of cable news video from the top three US cable news networks (CNN, FOX, and MSNBC), totaling 244,038 hours between January 2010 and July 2019. Our work uses technologies such as automatic face and gender recognition to measure the "screen time" of faces and to enable visual analysis and exploration at scale. Our analysis method gives insight into a broad set of socially relevant topics. For instance, male-presenting faces receive much more screen time than female-presenting faces (2.4x in 2010, 1.9x in 2019). To make our dataset and annotations accessible, we release a public interface at https://tvnews.stanford.edu that allows the general public to write queries and to perform their own analyses.
- Published
- 2021
7. Deep just-in-time defect prediction: how far are we?
- Author
-
Lingming Zhang, Zhengran Zeng, Yuqun Zhang, and Haotian Zhang
- Subjects
Continuous testing ,Computer science ,business.industry ,Deep learning ,Construct (python library) ,Commit ,Machine learning ,computer.software_genre ,Classifier (linguistics) ,Code (cryptography) ,Feature (machine learning) ,Embedding ,Artificial intelligence ,business ,computer - Abstract
Defect prediction aims to automatically identify potential defective code with minimal human intervention and has been widely studied in the literature. Just-in-Time (JIT) defect prediction focuses on program changes rather than whole programs, and has been widely adopted in continuous testing. CC2Vec, state-of-the-art JIT defect prediction tool, first constructs a hierarchical attention network (HAN) to learn distributed vector representations of both code additions and deletions, and then concatenates them with two other embedding vectors representing commit messages and overall code changes extracted by the existing DeepJIT approach to train a model for predicting whether a given commit is defective. Although CC2Vec has been shown to be the state of the art for JIT defect prediction, it was only evaluated on a limited dataset and not compared with all representative baselines. Therefore, to further investigate the efficacy and limitations of CC2Vec, this paper performs an extensive study of CC2Vec on a large-scale dataset with over 310,370 changes (8.3 X larger than the original CC2Vec dataset). More specifically, we also empirically compare CC2Vec against DeepJIT and representative traditional JIT defect prediction techniques. The experimental results show that CC2Vec cannot consistently outperform DeepJIT, and neither of them can consistently outperform traditional JIT defect prediction. We also investigate the impact of individual traditional defect prediction features and find that the added-line-number feature outperforms other traditional features. Inspired by this finding, we construct a simplistic JIT defect prediction approach which simply adopts the added-line-number feature with the logistic regression classifier. Surprisingly, such a simplistic approach can outperform CC2Vec and DeepJIT in defect prediction, and can be 81k X/120k X faster in training/testing. Furthermore, the paper also provides various practical guidelines for advancing JIT defect prediction in the near future.
- Published
- 2021
8. PatchScope: Memory Object Centric Patch Diffing
- Author
-
Yuncong Zhu, Heng Yin, Lei Zhao, Jiang Ming, Haotian Zhang, and Yichen Zhang
- Subjects
business.industry ,Computer science ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,020207 software engineering ,Context (language use) ,Memory corruption ,02 engineering and technology ,Data structure ,computer.software_genre ,Software ,Software security assurance ,0202 electrical engineering, electronic engineering, information engineering ,Code (cryptography) ,Key (cryptography) ,020201 artificial intelligence & image processing ,Data mining ,business ,Representation (mathematics) ,computer ,ComputingMethodologies_COMPUTERGRAPHICS - Abstract
Software patching is one of the most significant mechanisms to combat vulnerabilities. To demystify underlying patch details, the techniques of patch differential analysis (a.k.a. patch diffing) are proposed to find differences between patched and unpatched programs' binary code. Considering the sophisticated security patches, patch diffing is expected to not only correctly locate patch changes but also provide sufficient explanation for understanding patch details and the fixed vulnerabilities. Unfortunately, none of the existing patch diffing techniques can meet these requirements. In this study, we first perform a large-scale study on code changes of security patches for better understanding their patterns. We then point out several challenges and design principles for patch diffing. To address the above challenges, we design a dynamic patch diffing technique PatchScope. Our technique is motivated by two key observations: 1) the way that a program processes its input reveals a wealth of semantic information, and 2) most memory corruption patches regulate the handling of malformed inputs via updating the manipulations of input-related data structures. The core of PatchScope is a new semantics-aware program representation, memory object access sequence, which characterizes how a program references data structures to manipulate inputs. The representation can not only deliver succinct patch differences but also offer rich patch context information such as input-patch correlations. Such information can interpret patch differences and further help security analysts understand patch details, locate vulnerability root causes, and even detect buggy patches.
- Published
- 2020
9. Can automated program repair refine fault localization? a unified debugging approach
- Author
-
Lingming Zhang, Yiling Lou, Haotian Zhang, Ali Ghanbari, Lu Zhang, Xia Li, and Dan Hao
- Subjects
Computer science ,media_common.quotation_subject ,Distributed computing ,Supervised learning ,020207 software engineering ,02 engineering and technology ,Fault (power engineering) ,Debugging ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Benchmark (computing) ,Leverage (statistics) ,Software system ,Dimension (data warehouse) ,Scope (computer science) ,media_common - Abstract
A large body of research efforts have been dedicated to automated software debugging, including both automated fault localization and program repair. However, existing fault localization techniques have limited effectiveness on real-world software systems while even the most advanced program repair techniques can only fix a small ratio of real-world bugs. Although fault localization and program repair are inherently connected, their only existing connection in the literature is that program repair techniques usually use off-the-shelf fault localization techniques (e.g., Ochiai) to determine the potential candidate statements/elements for patching. In this work, we propose the unified debugging approach to unify the two areas in the other direction for the first time, i.e., can program repair in turn help with fault localization? In this way, we not only open a new dimension for more powerful fault localization, but also extend the application scope of program repair to all possible bugs (not only the bugs that can be directly automatically fixed). We have designed ProFL to leverage patch-execution results (from program repair) as the feedback information for fault localization. The experimental results on the widely used Defects4J benchmark show that the basic ProFL can already at least localize 37.61% more bugs within Top-1 than state-of-the-art spectrum and mutation based fault localization. Furthermore, ProFL can boost state-of-the-art fault localization via both unsupervised and supervised learning. Meanwhile, we have demonstrated ProFL's effectiveness under different settings and through a case study within Alipay, a popular online payment system with over 1 billion global users.
- Published
- 2020
10. Exploit the Connectivity
- Author
-
Yizhou Wang, Jenq-Neng Hwang, Haotian Zhang, Gaoang Wang, and Renshu Gu
- Subjects
FOS: Computer and information sciences ,Vertex (computer graphics) ,Similarity (geometry) ,business.industry ,Computer science ,Computer Vision and Pattern Recognition (cs.CV) ,Epipolar geometry ,Association (object-oriented programming) ,Frame (networking) ,Computer Science - Computer Vision and Pattern Recognition ,020206 networking & telecommunications ,02 engineering and technology ,Tracking (particle physics) ,Video tracking ,0202 electrical engineering, electronic engineering, information engineering ,Benchmark (computing) ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,business - Abstract
Multi-object tracking (MOT) is an important topic and critical task related to both static and moving camera applications, such as traffic flow analysis, autonomous driving and robotic vision. However, due to unreliable detection, occlusion and fast camera motion, tracked targets can be easily lost, which makes MOT very challenging. Most recent works exploit spatial and temporal information for MOT, but how to combine appearance and temporal features is still not well addressed. In this paper, we propose an innovative and effective tracking method called TrackletNet Tracker (TNT) that combines temporal and appearance information together as a unified framework. First, we define a graph model which treats each tracklet as a vertex. The tracklets are generated by associating detection results frame by frame with the help of the appearance similarity and the spatial consistency. To compensate camera movement, epipolar constraints are taken into consideration in the association. Then, for every pair of two tracklets, the similarity, called the connectivity in the paper, is measured by our designed multi-scale TrackletNet. Afterwards, the tracklets are clustered into groups and each group represents a unique object ID. Our proposed TNT has the ability to handle most of the challenges in MOT, and achieves promising results on MOT16 and MOT17 benchmark datasets compared with other state-of-the-art methods.
- Published
- 2019
11. Eye in the Sky
- Author
-
Gaoang Wang, Haotian Zhang, Jenq-Neng Hwang, and Zhichao Lei
- Subjects
FOS: Computer and information sciences ,Computer science ,business.industry ,Computer Vision and Pattern Recognition (cs.CV) ,Track (disk drive) ,Deep learning ,Image and Video Processing (eess.IV) ,Computer Science - Computer Vision and Pattern Recognition ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,02 engineering and technology ,Electrical Engineering and Systems Science - Image and Video Processing ,010501 environmental sciences ,Tracking (particle physics) ,01 natural sciences ,Object detection ,Drone ,Aerial photography ,Video tracking ,FOS: Electrical engineering, electronic engineering, information engineering ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,business ,0105 earth and related environmental sciences - Abstract
Drones, or general UAVs, equipped with a single camera have been widely deployed to a broad range of applications, such as aerial photography, fast goods delivery and most importantly, surveillance. Despite the great progress achieved in computer vision algorithms, these algorithms are not usually optimized for dealing with images or video sequences acquired by drones, due to various challenges such as occlusion, fast camera motion and pose variation. In this paper, a drone-based multi-object tracking and 3D localization scheme is proposed based on the deep learning based object detection. We first combine a multi-object tracking method called TrackletNet Tracker (TNT) which utilizes temporal and appearance information to track detected objects located on the ground for UAV applications. Then, we are also able to localize the tracked ground objects based on the group plane estimated from the Multi-View Stereo technique. The system deployed on the drone can not only detect and track the objects in a scene, but can also localize their 3D coordinates in meters with respect to the drone camera. The experiments have proved our tracker can reliably handle most of the detected objects captured by drones and achieve favorable 3D localization performance when compared with the state-of-the-art methods., Accepted to ACMMM2019
- Published
- 2019
12. Dynamic Sampling Meets Pooling
- Author
-
Mark D. Smucker, Nimesh Ghelani, Haotian Zhang, Amira Ghenai, Gordon V. Cormack, Shahin Rahbariasl, Mustafa Abualsaud, and Maura R. Grossman
- Subjects
Set (abstract data type) ,Computer science ,Statistics ,Pooling ,Rank (computer programming) ,Sampling (statistics) ,Relevance (information retrieval) - Abstract
A team of six assessors used Dynamic Sampling (Cormack and Grossman 2018) and one hour of assessment effort per topic to form, without pooling, a test collection for the TREC 2018 Common Core Track. Later, official relevance assessments were rendered by NIST for documents selected by depth-10 pooling augmented by move-to-front (MTF) pooling (Cormack et al. 1998), as well as the documents selected by our Dynamic Sampling effort. MAP estimates rendered from dynamically sampled assessments using the xinfAP statistical evaluator are comparable to those rendered from the complete set of official assessments using the standard trec_eval tool. MAP estimates rendered using only documents selected by pooling, on the other hand, differ substantially. The results suggest that the use of Dynamic Sampling without pooling can, for an order of magnitude less assessment effort, yield information-retrieval effectiveness estimates that exhibit lower bias, lower error, and comparable ability to rank system effectiveness.
- Published
- 2019
13. Effective User Interaction for High-Recall Retrieval
- Author
-
Mustafa Abualsaud, Maura R. Grossman, Nimesh Ghelani, Haotian Zhang, Mark D. Smucker, and Gordon V. Cormack
- Subjects
010104 statistics & probability ,Search engine ,Information retrieval ,Recall ,Computer science ,05 social sciences ,Active learning ,Relevance (information retrieval) ,0509 other social sciences ,0101 mathematics ,050904 information & library sciences ,01 natural sciences ,Test (assessment) - Abstract
High-recall retrieval --- finding all or nearly all relevant documents --- is critical to applications such as electronic discovery, systematic review, and the construction of test collections for information retrieval tasks. The effectiveness of current methods for high-recall information retrieval is limited by their reliance on human input, either to generate queries, or to assess the relevance of documents. Past research has shown that humans can assess the relevance of documents faster and with little loss in accuracy by judging shorter document surrogates, e.g.\ extractive summaries, in place of full documents. To test the hypothesis that short document surrogates can reduce assessment time and effort for high-recall retrieval, we conducted a 50-person, controlled, user study. We designed a high-recall retrieval system using continuous active learning (CAL) that could display either full documents or short document excerpts for relevance assessment. In addition, we tested the value of integrating a search engine with CAL. In the experiment, we asked participants to try to find as many relevant documents as possible within one hour. We observed that our study participants were able to find significantly more relevant documents when they used the system with document excerpts as opposed to full documents. We also found that allowing participants to compose and execute their own search queries did not improve their ability to find relevant documents and, by some measures, impaired performance. These results suggest that for high-recall systems to maximize performance, system designers should think carefully about the amount and nature of user interaction incorporated into the system.
- Published
- 2018
14. A System for Efficient High-Recall Retrieval
- Author
-
Nimesh Ghelani, Mark D. Smucker, Mustafa Abualsaud, Gordon V. Cormack, Haotian Zhang, and Maura R. Grossman
- Subjects
Flexibility (engineering) ,Search engine ,Information retrieval ,Recall ,Computer science ,Active learning (machine learning) ,020204 information systems ,Interface (computing) ,Active learning ,0202 electrical engineering, electronic engineering, information engineering ,Relevance feedback ,020201 artificial intelligence & image processing ,02 engineering and technology - Abstract
The goal of high-recall information retrieval (HRIR) is to find all or nearly all relevant documents for a search topic. In this paper, we present the design of our system that affords efficient high-recall retrieval. HRIR systems commonly rely on iterative relevance feedback. Our system uses a state-of-the-art implementation of continuous active learning (CAL), and is designed to allow other feedback systems to be attached with little work. Our system allows users to judge documents as fast as possible with no perceptible interface lag. We also support the integration of a search engine for users who would like to interactively search and judge documents. In addition to detailing the design of our system, we report on user feedback collected as part of a 50 participants user study. While we have found that users find the most relevant documents when we restrict user interaction, a majority of participants prefer having flexibility in user interaction. Our work has implications on how to build effective assessment systems and what features of the system are believed to be useful by users.
- Published
- 2018
15. Automatically Extracting High-Quality Negative Examples for Answer Selection in Question Answering
- Author
-
Haotian Zhang, Jimmy Lin, Jinfeng Rao, and Mark D. Smucker
- Subjects
Information retrieval ,Computer science ,business.industry ,Heuristic ,media_common.quotation_subject ,Deep learning ,02 engineering and technology ,Margin (machine learning) ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Selection (linguistics) ,Question answering ,020201 artificial intelligence & image processing ,Quality (business) ,Source document ,Artificial intelligence ,business ,media_common - Abstract
We propose a heuristic called "one answer per document" for automatically extracting high-quality negative examples for answer selection in question answering. Starting with a collection of question-answer pairs from the popular TrecQA dataset, we identify the original documents from which the answers were drawn. Sentences from these source documents that contain query terms (aside from the answers) are selected as negative examples. Training on the original data plus these negative examples yields improvements in effectiveness by a margin that is comparable to successive recent publications on this dataset. Our technique is completely unsupervised, which means that the gains come essentially for free. We confirm that the improvements can be directly attributed to our heuristic, as other approaches to extracting comparable amounts of training data are not effective. Beyond the empirical validation of this heuristic, we also share our improved TrecQA dataset with the community to support further work in answer selection.
- Published
- 2017
16. Optimizing Nugget Annotations with Active Learning
- Author
-
Haotian Zhang, Olga Vechtomova, Rakesh Guttikonda, Gaurav Baruah, Jimmy Lin, and Mark D. Smucker
- Subjects
Annotation ,Information retrieval ,Active learning (machine learning) ,Computer science ,020204 information systems ,InformationSystems_INFORMATIONSTORAGEANDRETRIEVAL ,0202 electrical engineering, electronic engineering, information engineering ,Process (computing) ,Question answering ,020201 artificial intelligence & image processing ,02 engineering and technology ,Automatic summarization ,Sentence - Abstract
Nugget-based evaluations, such as those deployed in the TREC Temporal Summarization and Question Answering tracks, require human assessors to determine whether a nugget is present in a given piece of text. This process, known as nugget annotation, is labor-intensive. In this paper, we present two active learning techniques that prioritize the sequence in which candidate nugget/sentence pairs are presented to an assessor, based on the likelihood that the sentence contains a nugget. Our approach builds on the recognition that nugget annotation is similar to high-recall retrieval, and we adapt proven existing solutions. Simulation experiments with four existing TREC test collections show that our techniques yield far more matches for a given level of effort than baselines that are typically deployed in previous nugget-based evaluations.
- Published
- 2016
17. Sampling Strategies and Active Learning for Volume Estimation
- Author
-
Mark D. Smucker, Jimmy Lin, Gordon V. Cormack, and Haotian Zhang
- Subjects
Information retrieval ,Computer science ,business.industry ,Sampling (statistics) ,020207 software engineering ,02 engineering and technology ,Volume estimation ,Residual ,Machine learning ,computer.software_genre ,Popularity ,020204 information systems ,Active learning ,0202 electrical engineering, electronic engineering, information engineering ,Leverage (statistics) ,Social media ,Artificial intelligence ,business ,computer - Abstract
This paper tackles the challenge of accurately and efficiently estimating the number of relevant documents in a collection for a particular topic. One real-world application is estimating the volume of social media posts (e.g., tweets) pertaining to a topic, which is fundamental to tracking the popularity of politicians and brands, the potential sales of a product, etc. Our insight is to leverage active learning techniques to find all the "easy" documents, and then to use sampling techniques to infer the number of relevant documents in the residual collection. We propose a simple yet effective technique for determining this "switchover" point, which intuitively can be understood as the "knee" in an effort vs. recall gain curve, as well as alternative sampling strategies beyond the knee. We show on several TREC datasets and a collection of tweets that our best technique yields more accurate estimates (with the same effort) than several alternatives.
- Published
- 2016
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.