56 results on '"Razvan Bunescu"'
Search Results
2. LSTMs and Deep Residual Networks for Carbohydrate and Bolus Recommendations in Type 1 Diabetes Management
- Author
-
Jeremy Beauchamp, Razvan Bunescu, Cindy Marling, Zhongen Li, and Chang Liu
- Subjects
diabetes management ,deep learning ,artificial intelligence ,Chemical technology ,TP1-1185 - Abstract
To avoid serious diabetic complications, people with type 1 diabetes must keep their blood glucose levels (BGLs) as close to normal as possible. Insulin dosages and carbohydrate consumption are important considerations in managing BGLs. Since the 1960s, models have been developed to forecast blood glucose levels based on the history of BGLs, insulin dosages, carbohydrate intake, and other physiological and lifestyle factors. Such predictions can be used to alert people of impending unsafe BGLs or to control insulin flow in an artificial pancreas. In past work, we have introduced an LSTM-based approach to blood glucose level prediction aimed at “what-if” scenarios, in which people could enter foods they might eat or insulin amounts they might take and then see the effect on future BGLs. In this work, we invert the “what-if” scenario and introduce a similar architecture based on chaining two LSTMs that can be trained to make either insulin or carbohydrate recommendations aimed at reaching a desired BG level in the future. Leveraging a recent state-of-the-art model for time series forecasting, we then derive a novel architecture for the same recommendation task, in which the two LSTM chain is used as a repeating block inside a deep residual architecture. Experimental evaluations using real patient data from the OhioT1DM dataset show that the new integrated architecture compares favorably with the previous LSTM-based approach, substantially outperforming the baselines. The promising results suggest that this novel approach could potentially be of practical use to people with type 1 diabetes for self-management of BGLs.
- Published
- 2021
- Full Text
- View/download PDF
3. Galaxy morphology prediction using Capsule Networks
- Author
-
Reza Katebi, Yadi Zhou, Ryan Chornock, and Razvan Bunescu
- Published
- 2019
- Full Text
- View/download PDF
4. A semantic parsing pipeline for context-dependent question answering over temporally structured data
- Author
-
Razvan Bunescu, Charles Chen, and Cindy Marling
- Subjects
Linguistics and Language ,Parsing ,Computer science ,business.industry ,Context (language use) ,computer.software_genre ,Pipeline (software) ,Language and Linguistics ,Artificial Intelligence ,Question answering ,Artificial intelligence ,business ,computer ,Software ,Natural language processing - Abstract
We propose a new setting for question answering (QA) in which users can query the system using both natural language and direct interactions within a graphical user interface that displays multiple time series associated with an entity of interest. The user interacts with the interface in order to understand the entity’s state and behavior, entailing sequences of actions and questions whose answers may depend on previous factual or navigational interactions. We describe a pipeline implementation where spoken questions are first transcribed into text which is then semantically parsed into logical forms that can be used to automatically extract the answer from the underlying database. The speech recognition module is implemented by adapting a pre-trained long short-term memory (LSTM)-based architecture to the user’s speech, whereas for the semantic parsing component we introduce an LSTM-based encoder–decoder architecture that models context dependency through copying mechanisms and multiple levels of attention over inputs and previous outputs. When evaluated separately, with and without data augmentation, both models are shown to substantially outperform several strong baselines. Furthermore, the full pipeline evaluation shows only a small degradation in semantic parsing accuracy, demonstrating that the semantic parser is robust to mistakes in the speech recognition output. The new QA paradigm proposed in this paper has the potential to improve the presentation and navigation of the large amounts of sensor data and life events that are generated in many areas of medicine.
- Published
- 2021
- Full Text
- View/download PDF
5. Hardware-Level Thread Migration to Reduce On-Chip Data Movement Via Reinforcement Learning
- Author
-
Razvan Bunescu, Ahmed Louri, Kyle Shiflett, Quintin Fettes, and Avinash Karanth
- Subjects
Computer science ,Distributed computing ,02 engineering and technology ,Energy consumption ,Thread (computing) ,Computer Graphics and Computer-Aided Design ,Execution time ,020202 computer hardware & architecture ,Instruction set ,Data access ,Scalability ,0202 electrical engineering, electronic engineering, information engineering ,Reinforcement learning ,Electrical and Electronic Engineering ,Latency (engineering) ,Software - Abstract
As the number of processing cores and associated threads in chip multiprocessors (CMPs) continues to scale out, on-chip memory access latency dominates application execution time due to increased data movement. Although tiled CMP architectures with distributed shared caches provide a scalable design, increased physical distance between requesting and responding cores has led to both increased on-chip memory access latency and excess energy consumption. Near data processing is a promising approach that can migrate threads closer to data, however prior hand-engineered rules for fine-grained hardware-level thread migration are either too slow to react to changes in data access patterns, or unable to exploit the large variety of data access patterns. In this article, we propose to use reinforcement learning (RL) to learn relatively complex data access patterns to improve on hardware-level thread migration techniques. By utilizing the recent history of memory access locations as input, each thread learns to recognize the relationship between prior access patterns and future memory access locations. This leads to the unique ability of the proposed technique to make fewer, more effective migrations to intermediate cores that minimize the distance to multiple distinct memory access locations. By allowing a low-overhead RL agent to learn a policy from real interaction with parallel programming benchmarks in a parallel simulator, we show that a migration policy which recognizes more complex data access patterns can be learned. The proposed approach reduces on-chip data movement and energy consumption by an average of 41%, while reducing execution time by 43% when compared to a simple baseline with no thread migration; furthermore, energy consumption and execution time are reduced by an additional 10% when compared to a hand-engineered fine-grained migration policy.
- Published
- 2020
- Full Text
- View/download PDF
6. Dynamic Voltage and Frequency Scaling in NoCs with Supervised and Reinforcement Learning Techniques
- Author
-
Avinash Karanth, Razvan Bunescu, Ahmed Louri, Quintin Fettes, and Mark Clark
- Subjects
Router ,Optimization problem ,General Computer Science ,Computer science ,Reliability (computer networking) ,02 engineering and technology ,Task (project management) ,Theoretical Computer Science ,law.invention ,law ,0202 electrical engineering, electronic engineering, information engineering ,Reinforcement learning ,Frequency scaling ,Throughput (business) ,Leakage (electronics) ,Multi-core processor ,business.industry ,Transistor ,020202 computer hardware & architecture ,Computational Theory and Mathematics ,Computer engineering ,Hardware and Architecture ,Scalability ,Software engineering ,business ,Software ,Voltage - Abstract
Network-on-Chips (NoCs) are the de facto choice for designing the interconnect fabric in multicore chips due to their regularity, efficiency, simplicity, and scalability. However, NoC suffers from excessive static power and dynamic energy due to transistor leakage current and data movement between the cores and caches. Power consumption issues are only exacerbated by ever decreasing technology sizes. Dynamic Voltage and Frequency Scaling (DVFS) is one technique that seeks to reduce dynamic energy; however this often occurs at the expense of performance. In this paper, we propose LEAD Learning-enabled Energy-Aware Dynamic voltage/frequency scaling for multicore architectures using both supervised learning and reinforcement learning approaches. LEAD groups the router and its outgoing links into the same V/F domain and implements proactive DVFS mode management strategies that rely on offline trained machine learning models in order to provide optimal V/F mode selection between different voltage/frequency pairs. We present three supervised learning versions of LEAD that are based on buffer utilization, change in buffer utilization and change in energy/throughput, which allow proactive mode selection based on accurate prediction of future network parameters. We then describe a reinforcement learning approach to LEAD that optimizes the DVFS mode selection directly, obviating the need for label and threshold engineering. Simulation results using PARSEC and Splash-2 benchmarks on a 4 × 4 concentrated mesh architecture show that by using supervised learning LEAD can achieve an average dynamic energy savings of 15.4 percent for a loss in throughput of 0.8 percent with no significant impact on latency. When reinforcement learning is used, LEAD increases average dynamic energy savings to 20.3 percent at the cost of a 1.5 percent decrease in throughput and a 1.7 percent increase in latency. Overall, the more flexible reinforcement learning approach enables learning an optimal behavior for a wider range of load environments under any desired energy versus throughput tradeoff.
- Published
- 2019
- Full Text
- View/download PDF
7. Bitwise Neural Network Acceleration Using Silicon Photonics
- Author
-
Avinash Karanth, Kyle Shiflett, Razvan Bunescu, and Ahmed Louri
- Subjects
Silicon photonics ,Speedup ,Artificial neural network ,Computer science ,Latency (audio) ,02 engineering and technology ,01 natural sciences ,Matrix multiplication ,010309 optics ,Reduction (complexity) ,020210 optoelectronics & photonics ,Computer engineering ,0103 physical sciences ,0202 electrical engineering, electronic engineering, information engineering ,Bitwise operation ,Efficient energy use - Abstract
Hardware accelerators provide significant speedup and improve energy efficiency for several demanding deep neural network (DNN) applications. DNNs have several hidden layers that perform concurrent matrix-vector multiplications (MVMs) between the network weights and input features. As MVMs are critical to the performance of DNNs, previous research has optimized the performance and energy efficiency of MVMs at both the architecture and algorithm levels. In this paper, we propose to use emerging silicon photonics technology to improve parallelism, speed and overall efficiency with the goal of providing real-time inference and fast training of neural nets. We use microring resonators (MRRs) and Mach-Zehnder interferometers (MZIs) to design two versions (all-optical and partial-optical) of hybrid matrix multiplications for DNNs. Our results indicate that our partial optical design gave the best performance in both energy efficiency and latency, with a reduction of 33.1% for energy-delay product (EDP) with conservative estimates and a 76.4% reduction for EDP with aggressive estimates.
- Published
- 2021
- Full Text
- View/download PDF
8. LSTMs and Deep Residual Networks for Carbohydrate and Bolus Recommendations in Type 1 Diabetes Management
- Author
-
Chang Liu, Jeremy Beauchamp, Zhongen Li, Razvan Bunescu, and Cindy Marling
- Subjects
Blood Glucose ,Pancreas, Artificial ,FOS: Computer and information sciences ,Computer Science - Machine Learning ,J.3 ,Computer science ,diabetes management ,medicine.medical_treatment ,030209 endocrinology & metabolism ,TP1-1185 ,Residual ,Machine learning ,computer.software_genre ,Biochemistry ,Artificial pancreas ,Article ,Machine Learning (cs.LG) ,Analytical Chemistry ,03 medical and health sciences ,0302 clinical medicine ,Bolus (medicine) ,Diabetes management ,medicine ,Humans ,Insulin ,Electrical and Electronic Engineering ,I.2.1 ,Instrumentation ,030304 developmental biology ,0303 health sciences ,Type 1 diabetes ,business.industry ,Self-Management ,Deep learning ,Chemical technology ,deep learning ,medicine.disease ,artificial intelligence ,Atomic and Molecular Physics, and Optics ,Diabetes Mellitus, Type 1 ,Chaining ,Artificial intelligence ,business ,computer - Abstract
To avoid serious diabetic complications, people with type 1 diabetes must keep their blood glucose levels (BGLs) as close to normal as possible. Insulin dosages and carbohydrate consumption are important considerations in managing BGLs. Since the 1960s, models have been developed to forecast blood glucose levels based on the history of BGLs, insulin dosages, carbohydrate intake, and other physiological and lifestyle factors. Such predictions can be used to alert people of impending unsafe BGLs or to control insulin flow in an artificial pancreas. In past work, we have introduced an LSTM-based approach to blood glucose level prediction aimed at “what-if” scenarios, in which people could enter foods they might eat or insulin amounts they might take and then see the effect on future BGLs. In this work, we invert the “what-if” scenario and introduce a similar architecture based on chaining two LSTMs that can be trained to make either insulin or carbohydrate recommendations aimed at reaching a desired BG level in the future. Leveraging a recent state-of-the-art model for time series forecasting, we then derive a novel architecture for the same recommendation task, in which the two LSTM chain is used as a repeating block inside a deep residual architecture. Experimental evaluations using real patient data from the OhioT1DM dataset show that the new integrated architecture compares favorably with the previous LSTM-based approach, substantially outperforming the baselines. The promising results suggest that this novel approach could potentially be of practical use to people with type 1 diabetes for self-management of BGLs.
- Published
- 2021
9. CSCNN: Algorithm-hardware Co-design for CNN Accelerators using Centrosymmetric Filters
- Author
-
Jiajun Li, Avinash Karanth, Ahmed Louri, and Razvan Bunescu
- Subjects
010302 applied physics ,Finite impulse response ,business.industry ,Computer science ,Computation ,Deep learning ,02 engineering and technology ,01 natural sciences ,Convolutional neural network ,020202 computer hardware & architecture ,Redundancy (information theory) ,0103 physical sciences ,0202 electrical engineering, electronic engineering, information engineering ,Artificial intelligence ,Pruning (decision trees) ,business ,Algorithm ,Computer hardware ,Energy (signal processing) ,Efficient energy use - Abstract
Convolutional neural networks (CNNs) are at the core of many state-of-the-art deep learning models in computer vision, speech, and text processing. Training and deploying such CNN-based architectures usually require a significant amount of computational resources. Sparsity has emerged as an effective compression approach for reducing the amount of data and computation for CNNs. However, sparsity often results in computational irregularity, which prevents accelerators from fully taking advantage of its benefits for performance and energy improvement. In this paper, we propose CSCNN, an algorithm/hardware co-design framework for CNN compression and acceleration that mitigates the effects of computational irregularity and provides better performance and energy efficiency. On the algorithmic side, CSCNN uses centrosymmetric matrices as convolutional filters. In doing so, it reduces the number of required weights by nearly 50% and enables structured computational reuse without compromising regularity and accuracy. Additionally, complementary pruning techniques are leveraged to further reduce computation by a factor of $2.8-7.2\times $ with a marginal accuracy loss. On the hardware side, we propose a CSCNN accelerator that effectively exploits the structured computational reuse enabled by centrosymmetric filters, and further eliminates zero computations for increased performance and energy efficiency. Compared against a dense accelerator, SCNN and SparTen, the proposed accelerator performs $3.7\times $, $1.6\times $ and $1.3\times $ better, and improves the EDP (Energy Delay Product) by $8.9\times $, $2.8\times $ and $2.0\times $, respectively.
- Published
- 2021
- Full Text
- View/download PDF
10. GCNAX: A Flexible and Energy-efficient Accelerator for Graph Convolutional Neural Networks
- Author
-
Razvan Bunescu, Jiajun Li, Ahmed Louri, and Avinash Karanth
- Subjects
010302 applied physics ,Loop (graph theory) ,Speedup ,business.industry ,Computer science ,Dataflow ,Loop fusion ,Deep learning ,Graph theory ,02 engineering and technology ,Parallel computing ,01 natural sciences ,Convolutional neural network ,020202 computer hardware & architecture ,0103 physical sciences ,0202 electrical engineering, electronic engineering, information engineering ,Artificial intelligence ,business ,Throughput (business) - Abstract
Graph convolutional neural networks (GCNs) have emerged as an effective approach to extend deep learning for graph data analytics. Given that graphs are usually irregular, as nodes in a graph may have a varying number of neighbors, processing GCNs efficiently pose a significant challenge on the underlying hardware. Although specialized GCN accelerators have been proposed to deliver better performance over generic processors, prior accelerators not only under-utilize the compute engine, but also impose redundant data accesses that reduce throughput and energy efficiency. Therefore, optimizing the overall flow of data between compute engines and memory, i.e., the GCN dataflow, which maximizes utilization and minimizes data movement is crucial for achieving efficient GCN processing.In this paper, we propose a flexible and optimized dataflow for GCNs that simultaneously improves resource utilization and reduces data movement. This is realized by fully exploring the design space of GCN dataflows and evaluating the number of execution cycles and DRAM accesses through an analysis framework. Unlike prior GCN dataflows, which employ rigid loop orders and loop fusion strategies, the proposed dataflow can reconFigure the loop order and loop fusion strategy to adapt to different GCN configurations, which results in much improved efficiency. We then introduce a novel accelerator architecture called GCNAX, which tailors the compute engine, buffer structure and size based on the proposed dataflow. Evaluated on five real-world graph datasets, our simulation results show that GCNAX reduces DRAM accesses by a factor of $8.1 \times$ and $2.4 \times$, while achieving $8.9 \times, 1.6 \times$ speedup and $9.5 \times$, $2.3 \times$ energy savings on average over HyGCN and AWB-GCN, respectively.
- Published
- 2021
- Full Text
- View/download PDF
11. An LSTM-based Approach for Insulin and Carbohydrate Recommendations in Type 1 Diabetes Self-Management
- Author
-
Cindy Marling, Razvan Bunescu, and Jeremy Beauchamp
- Subjects
Type 1 diabetes ,Self-management ,business.industry ,Insulin ,medicine.medical_treatment ,Machine learning ,computer.software_genre ,medicine.disease ,Artificial pancreas ,Lifestyle factors ,Medicine ,Artificial intelligence ,business ,computer ,Carbohydrate intake - Abstract
To avoid serious diabetic complications, people with type 1 diabetes must keep their blood glucose levels (BGLs) as close to normal as possible. Insulin dosages and carbohydrate consumption are important considerations in managing BGLs. Since the 1960s, models have been developed to forecast blood glucose levels based on the history of BGLs, insulin dosages, carbohydrate intake, and other physiological and lifestyle factors. Such predictions can be used to alert people to impending unsafe BGLs or to control insulin flow in an artificial pancreas. In past work, we have introduced an LSTM-based approach to blood glucose level prediction aimed at “what if” scenarios, in which people could input the grams of carbohydrates in foods they might eat or units of insulin they might take and then see the effect on future BGLs. Building on these neural models for “what-if” predictions, in this work we derive a novel LSTM-based architecture that can be trained to make either insulin or carbohydrate recommendations to ensure that future BGLs attain the desired level. Experimental evaluations using data from the OhioT1DM dataset show that the neural architecture substantially outperforms the baselines. The promising results suggest that this novel approach could potentially be of practical use to people with type 1 diabetes for self-management of BGLs.
- Published
- 2021
- Full Text
- View/download PDF
12. Primordial non-Gaussianity from the completed SDSS-IV extended Baryon Oscillation Spectroscopic Survey – I: Catalogue preparation and systematic mitigation
- Author
-
Kyle S. Dawson, Jiamin Hou, Ashley J. Ross, Hee-Jong Seo, Mehdi Rezaie, Eva Maria Mueller, Joel R. Brownstein, Axel de la Macorra, Julian E. Bautista, Gong-Bo Zhao, Etienne Burtin, Razvan Bunescu, Grant Merz, Eleanor B. Lyke, Reza Katebi, Will J. Percival, Donald P. Schneider, Graziano Rossi, Pauline Zarrouk, Héctor Gil-Marín, Institut de Recherches sur les lois Fondamentales de l'Univers (IRFU), Commissariat à l'énergie atomique et aux énergies alternatives (CEA)-Université Paris-Saclay, Department of Energy (US), Ohio Supercomputer Center, National Research Foundation of Korea, Ministry of Education, Science and Technology (South Korea), University of Utah, and Alfred P. Sloan Foundation
- Subjects
FOS: Computer and information sciences ,large-scale structure of the Universe ,Computer Science - Machine Learning ,Cosmology and Nongalactic Astrophysics (astro-ph.CO) ,FOS: Physical sciences ,Astrophysics ,Astrophysics::Cosmology and Extragalactic Astrophysics ,01 natural sciences ,Machine Learning (cs.LG) ,Non-Gaussianity ,0103 physical sciences ,Astrophysics::Solar and Stellar Astrophysics ,[INFO]Computer Science [cs] ,Large-scale structure of the Universe ,inflation ,Cluster analysis ,Spurious relationship ,010303 astronomy & astrophysics ,Stellar density ,Astrophysics::Galaxy Astrophysics ,Physics ,010308 nuclear & particles physics ,Astronomy and Astrophysics ,Quasar ,Covariance ,Computational Physics (physics.comp-ph) ,Inflation ,Redshift ,Baryon ,Space and Planetary Science ,Physics - Data Analysis, Statistics and Probability ,[PHYS.ASTR]Physics [physics]/Astrophysics [astro-ph] ,Physics - Computational Physics ,Data Analysis, Statistics and Probability (physics.data-an) ,[PHYS.PHYS.PHYS-DATA-AN]Physics [physics]/Physics [physics]/Data Analysis, Statistics and Probability [physics.data-an] ,Astrophysics - Cosmology and Nongalactic Astrophysics - Abstract
We investigate the large-scale clustering of the final spectroscopic sample of quasars from the recently completed extended Baryon Oscillation Spectroscopic Survey (eBOSS). The sample contains $343708$ objects in the redshift range $0.8, Comment: 17 pages, 13 figures, 2 tables. Accepted for publication in MNRAS. For the associated code and value-added catalogs see https://github.com/mehdirezaie/sysnetdev and https://github.com/mehdirezaie/eBOSSDR16QSOE
- Published
- 2021
- Full Text
- View/download PDF
13. Changing the Narrative Perspective: From Deictic to Anaphoric Point of View
- Author
-
Razvan Bunescu and Mike Chen
- Subjects
FOS: Computer and information sciences ,Computer Science - Artificial Intelligence ,Computer science ,media_common.quotation_subject ,02 engineering and technology ,Library and Information Sciences ,Management Science and Operations Research ,Fiction writing ,Deixis ,J.5 ,computer.software_genre ,050105 experimental psychology ,Reading (process) ,0202 electrical engineering, electronic engineering, information engineering ,Media Technology ,Selection (linguistics) ,Natural (music) ,0501 psychology and cognitive sciences ,Narrative ,media_common ,Computer Science - Computation and Language ,Point (typography) ,business.industry ,I.2.7 ,05 social sciences ,Perspective (graphical) ,Computer Science Applications ,Artificial Intelligence (cs.AI) ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,computer ,Computation and Language (cs.CL) ,Natural language processing ,Information Systems - Abstract
We introduce the task of changing the narrative point of view, where characters are assigned a narrative perspective that is different from the one originally used by the writer. The resulting shift in the narrative point of view alters the reading experience and can be used as a tool in fiction writing or to generate types of text ranging from educational to self-help and self-diagnosis. We introduce a benchmark dataset containing a wide range of types of narratives annotated with changes in point of view from deictic (first or second person) to anaphoric (third person) and describe a pipeline for processing raw text that relies on a neural architecture for mention selection. Evaluations on the new benchmark dataset show that the proposed architecture substantially outperforms the baselines by generating mentions that are less ambiguous and more natural., Comment: To appear in Information Processing & Management, Special Issue on Creative Language Processing
- Published
- 2021
- Full Text
- View/download PDF
14. Energy-Efficient Multiply-and-Accumulate using Silicon Photonics for Deep Neural Networks
- Author
-
Avinash Karanth, Ahmed Louri, Kyle Shiflett, and Razvan Bunescu
- Subjects
Silicon photonics ,020205 medical informatics ,Computer science ,0202 electrical engineering, electronic engineering, information engineering ,Deep neural networks ,02 engineering and technology ,Latency (engineering) ,Efficient energy use ,Computational science - Abstract
We propose two optical hybrid matrix multipliers for deep neural networks. Our results indicate our all-optical design achieved the best performance in energy efficiency and latency, with an energy-delay product reduction of 33.1% and 76.4% for conservative and aggressive estimates, respectively.
- Published
- 2020
- Full Text
- View/download PDF
15. The Promise and Perils of Wearable Physiological Sensors for Diabetes Management
- Author
-
Frank L. Schwartz, Cindy Marling, and Razvan Bunescu
- Subjects
Insulin pump ,Endocrinology, Diabetes and Metabolism ,medicine.medical_treatment ,Biomedical Engineering ,Wearable computer ,Special Section: Combining Diabetes Data from Wearable Devices ,030209 endocrinology & metabolism ,Bioengineering ,Machine Learning ,Smartwatch ,Wearable Electronic Devices ,03 medical and health sciences ,Patient safety ,0302 clinical medicine ,Human–computer interaction ,Diabetes management ,Internal Medicine ,Humans ,Medicine ,030212 general & internal medicine ,Fitness Trackers ,Physiologic monitoring ,business.industry ,Blood Glucose Self-Monitoring ,Insulin ,Diabetes Mellitus, Type 1 ,business - Abstract
Development of truly useful wearable physiologic monitoring devices for use in diabetes management is still in its infancy. From wearable activity monitors such as fitness trackers and smart watches to contact lenses measuring glucose levels in tears, we are just at the threshold of their coming use in medicine. Ultimately, such devices could help to improve the performance of sense-and-respond insulin pumps, illuminate the impact of physical activity on blood glucose levels, and improve patient safety. This is a summary of our experience attempting to use such devices to enhance continuous glucose monitoring–augmented insulin pump therapy. We discuss the current status and present difficulties with available devices, and review the potential for future use.
- Published
- 2018
- Full Text
- View/download PDF
16. Diagnosing Dysarthria with Long Short-Term Memory Networks
- Author
-
Sadegh Mirshekarian, Li Xu, Chang Liu, Zhiwei Mou, Alex Mayle, and Razvan Bunescu
- Subjects
Long short term memory ,Dysarthria ,medicine.medical_specialty ,Physical medicine and rehabilitation ,Computer science ,medicine ,medicine.symptom - Published
- 2019
- Full Text
- View/download PDF
17. Neural caption generation over figures
- Author
-
Razvan Bunescu, Scott Cohen, Charles Chen, Sungchul Kim, Ryan A. Rossi, Ruiyi Zhang, and Tong Yu
- Subjects
Closed captioning ,Parsing ,Artificial neural network ,Computer science ,business.industry ,010401 analytical chemistry ,Process (computing) ,020207 software engineering ,02 engineering and technology ,Attention model ,computer.software_genre ,01 natural sciences ,0104 chemical sciences ,Task (project management) ,0202 electrical engineering, electronic engineering, information engineering ,Artificial intelligence ,business ,computer ,Natural language ,Natural language processing - Abstract
Figures are human-friendly but difficult for computers to process automatically. In this work, we investigate the problem of figure captioning. The goal is to automatically generate a natural language description of a given figure. We create a new dataset for figure captioning, FigCAP. To achieve accurate generation of labels in figures, we propose the Label Maps Attention Model. Extensive experiments show that our method outperforms the baselines. A successful solution to this task allows figure content to be accessible to those with visual impairment by providing input to a text-to-speech system; and enables automatic parsing of vast repositories of documents where figures are pervasive.
- Published
- 2019
- Full Text
- View/download PDF
18. LSTMs and Neural Attention Models for Blood Glucose Prediction: Comparative Experiments on Real and Synthetic Data
- Author
-
Sadegh Mirshekarian, Razvan Bunescu, Cindy Marling, and Hui Shen
- Subjects
Blood Glucose ,Computer science ,02 engineering and technology ,Machine learning ,computer.software_genre ,01 natural sciences ,Article ,Synthetic data ,Set (abstract data type) ,Diabetes mellitus ,0202 electrical engineering, electronic engineering, information engineering ,Feature (machine learning) ,medicine ,Humans ,Attention ,In patient ,Time series ,business.industry ,010401 analytical chemistry ,Skin temperature ,medicine.disease ,0104 chemical sciences ,Diabetes Mellitus, Type 1 ,020201 artificial intelligence & image processing ,Neural Networks, Computer ,Noise (video) ,Artificial intelligence ,business ,computer - Abstract
We have shown in previous work that LSTM networks are effective at predicting blood glucose levels in patients with type I diabetes, outperforming human experts and an SVR model trained with features computed by manually engineered physiological models. In this paper we present the results of a much larger set of experiments on real and synthetic datasets in what-if, agnostic, and inertial scenarios. Experiments on a more recent real-patient dataset, which we are releasing to the research community, demonstrate that LSTMs are robust to noise and can easily incorporate additional features, such as skin temperature, heart rate and skin conductance, without any change in the architecture. A neural attention module that we designed specifically for time series prediction improves prediction performance on synthetic data; however, the improvements do not transfer to real data. Conversely, using time of day as an additional input feature consistently improves the LSTM performance on real data but not on synthetic data. These and other differences show that behavior on synthetic data cannot be assumed to always transfer to real data, highlighting the importance of evaluating physiological models on data from real patients.
- Published
- 2019
- Full Text
- View/download PDF
19. IntelliNoC
- Author
-
Avinash Karanth, Razvan Bunescu, Ahmed Louri, and Ke Wang
- Subjects
010302 applied physics ,Mean time between failures ,Computer science ,Network packet ,business.industry ,02 engineering and technology ,01 natural sciences ,020202 computer hardware & architecture ,Embedded system ,0103 physical sciences ,0202 electrical engineering, electronic engineering, information engineering ,Reinforcement learning ,Holistic design ,Latency (engineering) ,Architecture ,Error detection and correction ,business ,Efficient energy use - Abstract
As technology scales, Network-on-Chips (NoCs), currently being used for on-chip communication in manycore architectures, face several problems including high network latency, excessive power consumption, and low reliability. Simultaneously addressing these problems is proving to be difficult due to the explosion of the design space and the complexity of handling many trade-offs. In this paper, we propose IntelliNoC, an intelligent NoC design framework which introduces architectural innovations and uses reinforcement learning to manage the design complexity and simultaneously optimize performance, energy-efficiency, and reliability in a holistic manner. IntelliNoC integrates three NoC architectural techniques: (1) multifunction adaptive channels (MFACs) to improve energy-efficiency; (2) adaptive error detection/correction and re-transmission control to enhance reliability; and (3) a stress-relaxing bypass feature which dynamically powers off NoC components to prevent overheating and fatigue. To handle the complex dynamic interactions induced by these techniques, we train a dynamic control policy using Q-learning, with the goal of providing improved fault-tolerance and performance while reducing power consumption and area overhead. Simulation using PARSEC benchmarks shows that our proposed IntelliNoC design improves energy-efficiency by 67% and mean-time-to-failure (MTTF) by 77%, and decreases end-to-end packet latency by 32% and area requirements by 25% over baseline NoC architecture.
- Published
- 2019
- Full Text
- View/download PDF
20. High-performance, energy-efficient, fault-tolerant network-on-chip design using reinforcement learning
- Author
-
Ke Wang, Razvan Bunescu, Avinash Karanth, and Ahmed Louri
- Subjects
010302 applied physics ,Computer science ,Network packet ,business.industry ,Retransmission ,Word error rate ,Fault tolerance ,02 engineering and technology ,Chip ,01 natural sciences ,020202 computer hardware & architecture ,Embedded system ,0103 physical sciences ,0202 electrical engineering, electronic engineering, information engineering ,System on a chip ,Error detection and correction ,business ,Decoding methods ,Efficient energy use - Abstract
Network-on-Chips (NoCs) are becoming the standard communication fabric for multi-core and system on a chip (SoC) architectures. As technology continues to scale, transistors and wires on the chip are becoming increasingly vulnerable to various fault mechanisms, especially timing errors, resulting in exacerbation of energy efficiency and performance for NoCs. Typical techniques for handling timing errors are reactive in nature, responding to the faults after their occurrence. They rely on error detection/correction techniques which have resulted in excessive power consumption and degraded performance, since the error detection/correction hardware is constantly enabled. On the other hand, indiscriminately disabling error handling hardware can induce more errors and intrusive retransmission traffic. Therefore, the challenge is to balance the trade-offs among error rate, packet retransmission, performance, and energy. In this paper, we propose a proactive fault-tolerant mechanism to optimize energy efficiency and performance with reinforcement learning (RL). First, we propose a new proactive error handling technique comprised of a dynamic scheme for enabling per-router error detection/correction hardware and an effective retransmission mechanism. Second, we propose the use of RL to train the dynamic control policy with the goals of providing increased fault-tolerance, reduced power consumption and improved performance as compared to conventional techniques. Our evaluation indicates that, on average, end-to-end packet latency is lowered by 55%, energy efficiency is improved by 64%, and retransmission caused by faults is reduced by 48% over the reactive error correction techniques.
- Published
- 2019
- Full Text
- View/download PDF
21. From Physician Queries to Logical Forms for Efficient Exploration of Patient Data
- Author
-
Cindy Marling, Razvan Bunescu, Sadegh Mirshekarian, and Charles Chen
- Subjects
Parsing ,Copying ,Computer science ,business.industry ,computer.software_genre ,Semantics ,Question answering ,Artificial intelligence ,Architecture ,Inference engine ,business ,computer ,Natural language processing ,Natural language ,Graphical user interface - Abstract
We introduce a new question answering paradigm in which users can interact with the system using natural language questions or direct actions within a graphical user interface (GUI). The system displays multiple time series characterizing the behavior of a patient, and a physician interacts with the system through GUI actions and questions, where answers may depend on previous interactions. To find the answers automatically, we propose parsing the questions into logical forms for execution by an inference engine over the underlying database. The semantic parser is implemented as an LSTM-based encoder-decoder that models dependencies between consecutive answers through multiple attention and copying mechanisms. To train and evaluate the model, we created a dataset of semantic parses of real interactions with the system, augmented with a larger dataset of artificial interactions. The proposed architecture obtains promising results, substantially outperforming standard sequence generation baselines.
- Published
- 2019
- Full Text
- View/download PDF
22. Improving Galaxy Clustering Measurements with Deep Learning: analysis of the DECaLS DR7 data
- Author
-
Razvan Bunescu, Hee-Jong Seo, Ashley J. Ross, and Mehdi Rezaie
- Subjects
Physics ,Cosmology and Nongalactic Astrophysics (astro-ph.CO) ,Artificial neural network ,business.industry ,Deep learning ,Dimensionality reduction ,FOS: Physical sciences ,Astronomy and Astrophysics ,Astrophysics::Cosmology and Extragalactic Astrophysics ,Overfitting ,Galaxy ,Space and Planetary Science ,Physics - Data Analysis, Statistics and Probability ,Dark energy ,Artificial intelligence ,Astrophysics - Instrumentation and Methods for Astrophysics ,business ,Cluster analysis ,Algorithm ,Stellar density ,Instrumentation and Methods for Astrophysics (astro-ph.IM) ,Astrophysics::Galaxy Astrophysics ,Data Analysis, Statistics and Probability (physics.data-an) ,Astrophysics - Cosmology and Nongalactic Astrophysics - Abstract
Robust measurements of cosmological parameters from galaxy surveys rely on our understanding of systematic effects that impact the observed galaxy density field. In this paper we present, validate, and implement the idea of adopting the systematics mitigation method of Artificial Neural Networks for modeling the relationship between the target galaxy density field and various observational realities including but not limited to Galactic extinction, seeing, and stellar density. Our method by construction allows a wide class of models and alleviates over-training by performing k-fold cross-validation and dimensionality reduction via backward feature elimination. By permuting the choice of the training, validation, and test sets, we construct a selection mask for the entire footprint. We apply our method on the extended Baryon Oscillation Spectroscopic Survey (eBOSS) Emission Line Galaxies (ELGs) selection from the Dark Energy Camera Legacy Survey (DECaLS) Data Release 7 and show that the spurious large-scale contamination due to imaging systematics can be significantly reduced by up-weighting the observed galaxy density using the selection mask from the neural network and that our method is more effective than the conventional linear and quadratic polynomial functions. We perform extensive analyses on simulated mock datasets with and without systematic effects. Our analyses indicate that our methodology is more robust to overfitting compared to the conventional methods. This method can be utilized in the catalog generation of future spectroscopic galaxy surveys such as eBOSS and Dark Energy Spectroscopic Instrument (DESI) to better mitigate observational systematics., Comment: 31 pages, 25 figures, accepted for publication in MNRAS. Moderate revision throughout the paper. The new version includes a quantitative evaluation of the remaining systematic effects in the DECaLS DR7 data in Summary and Discussion and in Conclusion. Pipeline available at https://github.com/mehdirezaie/SYSNet
- Published
- 2019
- Full Text
- View/download PDF
23. Mapping Bug Reports to Relevant Files: A Ranking Model, a Fine-Grained Benchmark, and Feature Evaluation
- Author
-
Xin Ye, Razvan Bunescu, and Chang Liu
- Subjects
Source code ,Code review ,Java ,business.industry ,Computer science ,media_common.quotation_subject ,020207 software engineering ,02 engineering and technology ,Software maintenance ,computer.software_genre ,Dependency graph ,Software ,Software bug ,Ranking ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Learning to rank ,Data mining ,Software regression ,business ,computer ,computer.programming_language ,media_common - Abstract
When a new bug report is received, developers usually need to reproduce the bug and perform code reviews to find the cause, a process that can be tedious and time consuming. A tool for ranking all the source files with respect to how likely they are to contain the cause of the bug would enable developers to narrow down their search and improve productivity. This paper introduces an adaptive ranking approach that leverages project knowledge through functional decomposition of source code, API descriptions of library components, the bug-fixing history, the code change history, and the file dependency graph. Given a bug report, the ranking score of each source file is computed as a weighted combination of an array of features, where the weights are trained automatically on previously solved bug reports using a learning-to-rank technique. We evaluate the ranking system on six large scale open source Java projects, using the before-fix version of the project for every bug report. The experimental results show that the learning-to-rank approach outperforms three recent state-of-the-art methods. In particular, our method makes correct recommendations within the top 10 ranked source files for over 70 percent of the bug reports in the Eclipse Platform and Tomcat projects.
- Published
- 2016
- Full Text
- View/download PDF
24. Bug Report Classification Using LSTM Architecture for More Accurate Software Defect Locating
- Author
-
Xin Ye, Fan Fang, John Wu, Chang Liu, and Razvan Bunescu
- Subjects
Computer science ,business.industry ,020207 software engineering ,02 engineering and technology ,010501 environmental sciences ,Machine learning ,computer.software_genre ,01 natural sciences ,Ranking (information retrieval) ,Recurrent neural network ,Software ,Software bug ,Multilayer perceptron ,0202 electrical engineering, electronic engineering, information engineering ,False positive paradox ,Code (cryptography) ,Artificial intelligence ,business ,Precision and recall ,computer ,0105 earth and related environmental sciences - Abstract
Recently many information retrieval (IR)-based approaches have been proposed to help locate software defects automatically by using information from bug report contents. However, some bug reports that do not semantically related to the relevant code are not helpful to IR-based systems. Running an IR-based system on these reports may produce false positives. In this paper, we propose a classification model for classifying a bug report as either helpful or unhelpful using a LSTM-network. By filtering our unhelpful reports before running an IR-based bug locating system, our approach helps reduce false positives and improve the ranking performance. We test our model over 9,000 bug reports from three software projects. The evaluation result shows that our model helps improve a state-of-the-art IR-based system's ranking performance under a trade-off between the precision and the recall. Our comparison experiments show that the LSTM-network achieves the best trade-off between precision and recall than other classification models including CNN, multilayer perceptron, and a simple baseline approach that classifies a bug report based its length. In the situation that precision is more important than recall, our classification model helps for bug locating.
- Published
- 2018
- Full Text
- View/download PDF
25. Predictive Analysis by Leveraging Temporal User Behavior and User Embeddings
- Author
-
Branislav Kveton, Razvan Bunescu, Sungchul Kim, Ryan A. Rossi, Charles Chen, Eunyee Koh, and Hung Bui
- Subjects
Computer science ,business.industry ,02 engineering and technology ,Service provider ,Machine learning ,computer.software_genre ,Task (computing) ,Recurrent neural network ,Bag-of-words model ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,tf–idf ,Mobile device ,Feature learning ,computer ,Word (computer architecture) - Abstract
The rapid growth of mobile devices has resulted in the generation of a large number of user behavior logs that contain latent intentions and user interests. However, exploiting such data in real-world applications is still difficult for service providers due to the complexities of user behavior over a sheer number of possible actions that can vary according to time. In this work, a time-aware RNN model, TRNN, is proposed for predictive analysis from user behavior data. First, our approach predicts next user action more accurately than the baselines including the n-gram models as well as two recently introduced time-aware RNN approaches. Second, we use TRNN to learn user embeddings from sequences of user actions and show that overall the TRNN embeddings outperform conventional RNN embeddings. Similar to how word embeddings benefit a wide range of task in natural language processing, the learned user embeddings are general and could be used in a variety of tasks in the digital marketing area. This claim is supported empirically by evaluating their utility in user conversion prediction, and preferred application prediction. According to the evaluation results, TRNN embeddings perform better than the baselines including Bag of Words (BoW), TFIDF and Doc2Vec. We believe that TRNN embeddings provide an effective representation for solving practical tasks such as recommendation, user segmentation and predictive analysis of business metrics.
- Published
- 2018
- Full Text
- View/download PDF
26. Set cover-based methods for motif selection
- Author
-
Lonnie R. Welch, Razvan Bunescu, Frank Drews, David W. Juedes, Yating Liu, and Yichao Li
- Subjects
Statistics and Probability ,Chromatin Immunoprecipitation ,Computer science ,Computational biology ,ENCODE ,Biochemistry ,Cofactor ,03 medical and health sciences ,0302 clinical medicine ,Binding site ,Nucleotide Motifs ,Molecular Biology ,030304 developmental biology ,0303 health sciences ,Binding Sites ,biology ,Set cover problem ,Sequence Analysis, DNA ,Original Papers ,Tabu search ,Computer Science Applications ,DNA binding site ,Computational Mathematics ,Computational Theory and Mathematics ,biology.protein ,Sequence motif ,Sequence Analysis ,030217 neurology & neurosurgery ,Algorithms ,Software ,Transcription Factors - Abstract
Motivation De novo motif discovery algorithms find statistically over-represented sequence motifs that may function as transcription factor binding sites. Current methods often report large numbers of motifs, making it difficult to perform further analyses and experimental validation. The motif selection problem seeks to identify a minimal set of putative regulatory motifs that characterize sequences of interest (e.g. ChIP-Seq binding regions). Results In this study, the motif selection problem is mapped to variants of the set cover problem that are solved via tabu search and by relaxed integer linear programing (RILP). The algorithms are employed to analyze 349 ChIP-Seq experiments from the ENCODE project, yielding a small number of high-quality motifs that represent putative binding sites of primary factors and cofactors. Specifically, when compared with the motifs reported by Kheradpour and Kellis, the set cover-based algorithms produced motif sets covering 35% more peaks for 11 TFs and identified 4 more putative cofactors for 6 TFs. Moreover, a systematic evaluation using nested cross-validation revealed that the RILP algorithm selected fewer motifs and was able to cover 6% more peaks and 3% fewer background regions, which reduced the error rate by 7%. Availability and implementation The source code of the algorithms and all the datasets are available at https://github.com/YichaoOU/Set_cover_tools. Supplementary information Supplementary data are available at Bioinformatics online.
- Published
- 2018
27. LEAD
- Author
-
Ahmed Louri, Razvan Bunescu, Mark Clark, and Avinash Kodi
- Subjects
010302 applied physics ,Router ,Interconnection ,Multi-core processor ,Dynamic voltage frequency scaling ,business.industry ,Energy management ,Computer science ,02 engineering and technology ,01 natural sciences ,020202 computer hardware & architecture ,Embedded system ,0103 physical sciences ,Dynamic demand ,Scalability ,0202 electrical engineering, electronic engineering, information engineering ,business ,Frequency scaling ,Voltage - Abstract
Network on Chips (NoCs) are the interconnect fabric of choice for multicore processors due to their superiority over traditional buses and crossbars in terms of scalability. While NoC’s offer several advantages, they still suffer from high static and dynamic power consumption. Dynamic Voltage and Frequency Scaling (DVFS) is a popular technique that allows dynamic energy to be saved, but it can potentially lead to loss in throughput. In this paper, we propose LEAD - Learning-enabled Energy-Aware Dynamic voltage/frequency scaling for NoC architectures wherein we use machine learning techniques to enable energy-performance trade-offs at reduced overhead cost. LEAD enables a proactive energy management strategy that relies on an offline trained regression model and provides a wide variety of voltage/frequency pairs (modes). LEAD groups each router and the router’s outgoing links locally into the same V/F domain, allowing energy management at a finer granularity without additional timing complications and overhead. Our simulation results using PARSEC and Splash-2 benchmarks on a 4 × 4 concentrated mesh architecture show an average dynamic energy savings of 17% with a minimal loss of 4% in throughput and no latency increase.
- Published
- 2018
- Full Text
- View/download PDF
28. Extending the Power-Efficiency and Performance of Photonic Interconnects for Heterogeneous Multicores with Machine Learning
- Author
-
Ahmed Louri, Razvan Bunescu, Scott Van Winkle, and Avinash Kodi
- Subjects
010302 applied physics ,Router ,Computer science ,business.industry ,Bandwidth (signal processing) ,Control reconfiguration ,02 engineering and technology ,Energy consumption ,Machine learning ,computer.software_genre ,01 natural sciences ,020202 computer hardware & architecture ,0103 physical sciences ,0202 electrical engineering, electronic engineering, information engineering ,Artificial intelligence ,Laser power scaling ,Performance improvement ,business ,Electrical efficiency ,computer ,Performance per watt - Abstract
As communication energy exceeds computation energy in future technologies, traditional on-chip electrical interconnects face fundamental challenges in the many-core era. Photonic interconnects have been proposed as a disruptive technology solution due to superior performance per Watt, distance independent energy consumption and CMOS compatibility for on-chip interconnects. Static power due to the laser being always switched on, varying link utilization due to spatial and temporal traffic fluctuations and thermal sensitivity are some of the critical challenges facing photonics interconnects. In this paper, we propose photonic interconnects for heterogeneous multicores using a checkerboard pattern that clusters CPU-GPU cores together and implements bandwidth reconfiguration using local router information without global coordination. To reduce the static power, we also propose a dynamic laser scaling technique that predicts the power level for the next epoch using the buffer occupancy of previous epoch. To further improve power-performance trade-offs, we also propose a regression-based machine learning technique for scaling the power of the photonic link. Our simulation results demonstrate a 34% performance improvement over a baseline electrical CMESH while consuming 25% less energy per bit when dynamically reallocating bandwidth. When dynamically scaling laser power, our buffer-based reactive and ML-based proactive prediction techniques show 40 - 65% in power savings with 0 - 14% in throughput loss depending on the reservation window size.
- Published
- 2018
- Full Text
- View/download PDF
29. Galaxy morphology prediction using capsule networks
- Author
-
Razvan Bunescu, Ryan Chornock, Yadi Zhou, and Reza Katebi
- Subjects
FOS: Computer and information sciences ,Computer Science - Machine Learning ,media_common.quotation_subject ,Computer Vision and Pattern Recognition (cs.CV) ,Computer Science - Computer Vision and Pattern Recognition ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,FOS: Physical sciences ,02 engineering and technology ,Astrophysics::Cosmology and Extragalactic Astrophysics ,01 natural sciences ,Machine Learning (cs.LG) ,0103 physical sciences ,0202 electrical engineering, electronic engineering, information engineering ,Neural and Evolutionary Computing (cs.NE) ,010303 astronomy & astrophysics ,Instrumentation and Methods for Astrophysics (astro-ph.IM) ,media_common ,Physics ,Artificial neural network ,business.industry ,Computer Science - Neural and Evolutionary Computing ,Astronomy and Astrophysics ,Pattern recognition ,Astrophysics - Astrophysics of Galaxies ,Regression ,Galaxy ,Space and Planetary Science ,Sky ,Astrophysics of Galaxies (astro-ph.GA) ,Neural network architecture ,020201 artificial intelligence & image processing ,Artificial intelligence ,Astrophysics - Instrumentation and Methods for Astrophysics ,business ,Classifier (UML) - Abstract
Understanding morphological types of galaxies is a key parameter for studying their formation and evolution. Neural networks that have been used previously for galaxy morphology classification have some disadvantages, such as not being invariant under rotation. In this work, we studied the performance of Capsule Network, a recently introduced neural network architecture that is rotationally invariant and spatially aware, on the task of galaxy morphology classification. We designed two evaluation scenarios based on the answers from the question tree in the Galaxy Zoo project. In the first scenario, we used Capsule Network for regression and predicted probabilities for all of the questions. In the second scenario, we chose the answer to the first morphology question that had the highest user agreement as the class of the object and trained a Capsule Network classifier, where we also reconstructed galaxy images. We achieved promising results in both of these scenarios. Automated approaches such as the one introduced here will greatly decrease the workload of astronomers and will play a critical role in the upcoming large sky surveys., Comment: 9 pages, 6 figures
- Published
- 2018
- Full Text
- View/download PDF
30. Resilient and Power-Efficient Multi-Function Channel Buffers in Network-on-Chip Architectures
- Author
-
Razvan Bunescu, Dominic DiTomaso, Avinash Kodi, and Ahmed Louri
- Subjects
Router ,business.industry ,Computer science ,Fault tolerance ,Throughput ,Hardware_PERFORMANCEANDRELIABILITY ,Fault (power engineering) ,Theoretical Computer Science ,Traffic flow (computer networking) ,Network on a chip ,Computational Theory and Mathematics ,Hardware and Architecture ,Embedded system ,Scalability ,Link level ,business ,Software - Abstract
Network-on-Chips (NoCs) are quickly becoming the standard communication paradigm for the growing number of cores on the chip. While NoCs can deliver sufficient bandwidth and enhance scalability, NoCs suffer from high power consumption due to the router microarchitecture and communication channels that facilitate inter-core communication. As technology keeps scaling down in the nanometer regime, unpredictable device behavior due to aging, infant mortality, design defects, soft errors, aggressive design, and process-voltage-temperature variations, will increase and will result in a significant increase in faults (both permanent and transient) and hardware failures. In this paper, we propose QORE —a fault tolerant NoC architecture with Multi-Function Channel (MFC) buffers. The use of MFC buffers and their associated control (link and fault controllers) enhance fault-tolerance by allowing the NoC to dynamically adapt to faults at the link level and reverse propagation direction to avoid faulty links. Additionally, MFC buffers reduce router power and improve performance by eliminating in-router buffering. We utilize a machine learning technique in our link controllers to predict the direction of traffic flow in order to more efficiently reverse links. Our simulation results using real benchmarks and synthetic traffic mixes show that QORE improves speedup by 1.3 $\times$ and throughput by 2.3 $\times$ when compared to state-of-the art fault tolerant NoCs designs such as Ariadne and Vicis. Moreover, using Synopsys Design Compiler, we also show that network power in QORE is reduced by 21 percent with minimal control overhead.
- Published
- 2015
- Full Text
- View/download PDF
31. Recognition of Dynamic Hand Gestures from 3D Motion Data Using LSTM and CNN Architectures
- Author
-
Chinmaya R. Naguri and Razvan Bunescu
- Subjects
Flexibility (engineering) ,Computer science ,Speech recognition ,Process (computing) ,020207 software engineering ,02 engineering and technology ,Voice command device ,Virtual reality ,Convolutional neural network ,Gesture recognition ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Augmented reality ,Gesture - Abstract
Hand gestures provide a natural, non-verbal form of communication that can augment or replace other communication modalities such as speech or writing. Along with voice commands, hand gestures are becoming the primary means of interaction in games, augmented reality, and virtual reality platforms. Recognition accuracy, flexibility, and computational cost are some of the primary factors that can impact the incorporation of hand gestures in these new technologies, as well as their subsequent retrieval from multimodal corpora. In this paper, we present fast and highly accurate gesture recognition systems based on long short-term memory (LSTM) and convolutional neural networks (CNN) that are trained to process input sequences of 3D hand positions and velocities acquired from infrared sensors. When evaluated on real time recognition of six types of hand gestures, the proposed architectures obtain 97% F-measure, demonstrating a significant potential for practical applications in novel human-computer interfaces.
- Published
- 2017
- Full Text
- View/download PDF
32. Using LSTMs to learn physiological models of blood glucose behavior
- Author
-
Frank L. Schwartz, Cindy Marling, Sadegh Mirshekarian, and Razvan Bunescu
- Subjects
Blood Glucose ,medicine.medical_specialty ,Memory, Long-Term ,Glucose control ,medicine.medical_treatment ,030209 endocrinology & metabolism ,02 engineering and technology ,Machine learning ,computer.software_genre ,Article ,03 medical and health sciences ,0302 clinical medicine ,Internal medicine ,Diabetes mellitus ,0202 electrical engineering, electronic engineering, information engineering ,medicine ,Humans ,Insulin ,Carbohydrate intake ,Type 1 diabetes ,Artificial neural network ,business.industry ,medicine.disease ,Physiological model ,Endocrinology ,Recurrent neural network ,Diabetes Mellitus, Type 1 ,020201 artificial intelligence & image processing ,Artificial intelligence ,Neural Networks, Computer ,business ,computer - Abstract
For people with type 1 diabetes, good blood glucose control is essential to keeping serious disease complications at bay. This entails carefully monitoring blood glucose levels and taking corrective steps whenever they are too high or too low. If blood glucose levels could be accurately predicted, patients could take proactive steps to prevent blood glucose excursions from occurring. However, accurate predictions require complex physiological models of blood glucose behavior. Factors such as insulin boluses, carbohydrate intake, and exercise influence blood glucose in ways that are difficult to capture through manually engineered equations. In this paper, we describe a recursive neural network (RNN) approach that uses long short-term memory (LSTM) units to learn a physiological model of blood glucose. When trained on raw data from real patients, the LSTM networks (LSTMs) obtain results that are competitive with a previous state-of-the-art model based on manually engineered physiological equations. The RNN approach can incorporate arbitrary physiological parameters without the need for sophisticated manual engineering, thus holding the promise of further improvements in prediction accuracy.
- Published
- 2017
33. Chord Recognition in Symbolic Music: A Segmental CRF Model, Segment-Level Features, and Comparative Evaluations on Classical and Popular Music
- Author
-
Razvan Bunescu and Kristen Masada
- Subjects
lcsh:M1-5000 ,FOS: Computer and information sciences ,Conditional random field ,Sound (cs.SD) ,Computer Science - Machine Learning ,Computer science ,Machine Learning (stat.ML) ,Computer Science - Sound ,Machine Learning (cs.LG) ,Popular music ,Audio and Speech Processing (eess.AS) ,Statistics - Machine Learning ,FOS: Electrical engineering, electronic engineering, information engineering ,segmental CRF ,Segmentation ,chord recognition ,lcsh:Music ,lcsh:T58.5-58.64 ,lcsh:Information technology ,business.industry ,Symbolic music ,Chord recognition ,Pattern recognition ,semi-CRF ,Classical music ,ComputingMethodologies_PATTERNRECOGNITION ,harmonic analysis ,Rock music ,Chord (music) ,Artificial intelligence ,business ,symbolic music ,Electrical Engineering and Systems Science - Audio and Speech Processing - Abstract
We present a new approach to harmonic analysis that is trained to segment music into a sequence of chord spans tagged with chord labels. Formulated as a semi-Markov Conditional Random Field (semi-CRF), this joint segmentation and labeling approach enables the use of a rich set of segment-level features, such as segment purity and chord coverage, that capture the extent to which the events in an entire segment of music are compatible with a candidate chord label. The new chord recognition model is evaluated extensively on three corpora of Western classical music and a newly created corpus of rock music. Experimental results show that the semi-CRF model performs substantially better than previous approaches when trained on a sufficient number of labeled examples and remains competitive when the amount of training data is limited.
- Published
- 2019
- Full Text
- View/download PDF
34. Tone Classification in Mandarin Chinese Using Convolutional Neural Networks
- Author
-
Chang Liu, Li Xu, Razvan Bunescu, and Charles Chen
- Subjects
Tone (musical instrument) ,Computer science ,Speech recognition ,language ,Convolutional neural network ,Mandarin Chinese ,language.human_language - Published
- 2016
- Full Text
- View/download PDF
35. From word embeddings to document similarities for improved information retrieval in software engineering
- Author
-
Chang Liu, Xin Ye, Hui Shen, Xiao Ma, and Razvan Bunescu
- Subjects
Vocabulary ,Information retrieval ,business.industry ,Computer science ,media_common.quotation_subject ,Computer programming ,020207 software engineering ,02 engineering and technology ,Semantics ,computer.software_genre ,Bridging (programming) ,Software ,Software bug ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Question answering ,Artificial intelligence ,business ,Software engineering ,computer ,Natural language ,Natural language processing ,media_common - Abstract
The application of information retrieval techniques to search tasks in software engineering is made difficult by the lexical gap between search queries, usually expressed in natural language (e.g. English), and retrieved documents, usually expressed in code (e.g. programming languages). This is often the case in bug and feature location, community question answering, or more generally the communication between technical personnel and non-technical stake holders in a software project. In this paper, we propose bridging the lexical gap by projecting natural language statements and code snippets as meaning vectors in a shared representation space. In the proposed architecture, word embeddings are first trained on API documents, tutorials, and reference documents, and then aggregated in order to estimate semantic similarities between documents. Empirical evaluations show that the learned vector space embeddings lead to improvements in a previously explored bug localization task and a newly defined task of linking API documents to computer programming questions.
- Published
- 2016
- Full Text
- View/download PDF
36. Mining knowledge from text using information extraction
- Author
-
Razvan Bunescu and Raymond J. Mooney
- Subjects
Text corpus ,Information retrieval ,Noisy text analytics ,Computer science ,business.industry ,Geography, Planning and Development ,Text graph ,Concept mining ,computer.software_genre ,Biomedical text mining ,Relationship extraction ,Information extraction ,Knowledge extraction ,General Earth and Planetary Sciences ,Artificial intelligence ,business ,computer ,Natural language processing ,Water Science and Technology - Abstract
An important approach to text mining involves the use of natural-language information extraction. Information extraction (IE) distills structured data or knowledge from unstructured text by identifying references to named entities as well as stated relationships between such entities. IE systems can be used to directly extricate abstract knowledge from a text corpus, or to extract concrete data from a set of documents which can then be further analyzed with traditional data-mining techniques to discover more general patterns. We discuss methods and implemented systems for both of these approaches and summarize results on mining real text corpora of biomedical abstracts, job announcements, and product descriptions. We also discuss challenges that arise when employing current information extraction technology to discover knowledge in text.
- Published
- 2005
- Full Text
- View/download PDF
37. Comparative experiments on learning information extractors for proteins and their interactions
- Author
-
Raymond J. Mooney, Arun K. Ramani, Edward M. Marcotte, Razvan Bunescu, Yuk Wah Wong, Ruifang Ge, and Rohit J. Kate
- Subjects
Information retrieval ,Computer science ,Rule induction ,MEDLINE ,Principle of maximum entropy ,Information Storage and Retrieval ,Proteins ,Medicine (miscellaneous) ,Expert Systems ,computer.software_genre ,Expert system ,Variety (cybernetics) ,Support vector machine ,Set (abstract data type) ,Information extraction ,ComputingMethodologies_PATTERNRECOGNITION ,Genes ,Artificial Intelligence ,Terminology as Topic ,Humans ,computer ,Algorithms - Abstract
Objective:: Automatically extracting information from biomedical text holds the promise of easily consolidating large amounts of biological knowledge in computer-accessible form. This strategy is particularly attractive for extracting data relevant to genes of the human genome from the 11 million abstracts in Medline. However, extraction efforts have been frustrated by the lack of conventions for describing human genes and proteins. We have developed and evaluated a variety of learned information extraction systems for identifying human protein names in Medline abstracts and subsequently extracting information on interactions between the proteins. Methods and Material:: We used a variety of machine learning methods to automatically develop information extraction systems for extracting information on gene/protein name, function and interactions from Medline abstracts. We present cross-validated results on identifying human proteins and their interactions by training and testing on a set of approximately 1000 manually-annotated Medline abstracts that discuss human genes/proteins. Results:: We demonstrate that machine learning approaches using support vector machines and maximum entropy are able to identify human proteins with higher accuracy than several previous approaches. We also demonstrate that various rule induction methods are able to identify protein interactions with higher precision than manually-developed rules. Conclusion:: Our results show that it is promising to use machine learning to automatically build systems for extracting information from biomedical text. The results also give a broad picture of the relative strengths of a wide variety of methods when tested on a reasonably large human-annotated corpus.
- Published
- 2005
- Full Text
- View/download PDF
38. Learning to rank relevant files for bug reports using domain knowledge
- Author
-
Chang Liu, Xin Ye, and Razvan Bunescu
- Subjects
Source code ,Code review ,Database ,Java ,Computer science ,media_common.quotation_subject ,computer.software_genre ,Ranking ,Software bug ,Domain knowledge ,Learning to rank ,Software regression ,computer ,media_common ,computer.programming_language - Abstract
When a new bug report is received, developers usually need to reproduce the bug and perform code reviews to find the cause, a process that can be tedious and time consuming. A tool for ranking all the source files of a project with respect to how likely they are to contain the cause of the bug would enable developers to narrow down their search and potentially could lead to a substantial increase in productivity. This paper introduces an adaptive ranking approach that leverages domain knowledge through functional decompositions of source code files into methods, API descriptions of library components used in the code, the bug-fixing history, and the code change history. Given a bug report, the ranking score of each source file is computed as a weighted combination of an array of features encoding domain knowledge, where the weights are trained automatically on previously solved bug reports using a learning-to-rank technique. We evaluated our system on six large scale open source Java projects, using the before-fix version of the project for every bug report. The experimental results show that the newly introduced learning-to-rank approach significantly outperforms two recent state-of-the-art methods in recommending relevant files for bug reports. In particular, our method makes correct recommendations within the top 10 ranked source files for over 70% of the bug reports in the Eclipse Platform and Tomcat projects.
- Published
- 2014
- Full Text
- View/download PDF
39. Blood Glucose Level Prediction Using Physiological Models and Support Vector Regression
- Author
-
Jay Shubrook, Razvan Bunescu, Nigel Struble, Frank L. Schwartz, and Cindy Marling
- Subjects
Computer science ,business.industry ,Insulin ,medicine.medical_treatment ,Life events ,medicine.disease ,Machine learning ,computer.software_genre ,Support vector machine ,Diabetes mellitus ,medicine ,Artificial intelligence ,Sugar ,business ,computer ,Normal range - Abstract
Patients with diabetes must continually monitor their blood glucose levels and adjust insulin doses, striving to keep blood glucose levels as close to normal as possible. Blood glucose levels that deviate from the normal range can lead to serious short-term and long-term complications. An automatic prediction model that warned people of imminent changes in their blood glucose levels would enable them to take preventive action. Modeling inter-patient differences and the combined effects of insulin and life events on blood glucose have been particularly challenging in the design of accurate blood glucose forecasting systems. In this paper, we describe a solution that uses a generic physiological model of blood glucose dynamics to generate informative features for a Support Vector Regression model that is trained on patient specific data. Experimental results show that the new prediction model outperforms all three diabetes experts involved in the study, thus demonstrating the utility of using the generic physiological features in machine learning models that are individually trained for every patient.
- Published
- 2013
- Full Text
- View/download PDF
40. A Consensus Perceived Glycemic Variability Metric
- Author
-
Jay H. Shubrook, Frank L. Schwartz, Cindy Marling, Razvan Bunescu, and Nigel Struble
- Subjects
Blood Glucose ,Consensus ,Mean squared error ,Computer science ,Endocrinology, Diabetes and Metabolism ,Biomedical Engineering ,Bioengineering ,Validation Studies as Topic ,Sensitivity and Specificity ,Standard deviation ,Cross-validation ,Control theory ,Statistics ,Internal Medicine ,medicine ,Humans ,Glycemic ,Observer Variation ,Type 1 diabetes ,Blood Glucose Self-Monitoring ,Reproducibility of Results ,Weights and Measures ,medicine.disease ,Regression ,Support vector machine ,Diabetes Mellitus, Type 1 ,Data Interpretation, Statistical ,Original Article ,Perception ,Metric (unit) ,Algorithms - Abstract
Objective: Glycemic variability (GV) is an important component of overall glycemic control for patients with diabetes mellitus. Physicians are able to recognize excessive GV from continuous glucose monitoring (CGM) plots; however, there is currently no universally agreed upon GV metric. The objective of this study was to develop a consensus perceived glycemic variability (CPGV) metric that could be routinely applied to CGM data to assess diabetes mellitus control. Methods: Twelve physicians actively managing patients with type 1 diabetes mellitus rated a total of 250 24 h CGM plots as exhibiting low, borderline, high, or extremely high GV. Ratings were averaged to obtain a consensus and then input into two machine learning algorithms: multilayer perceptrons (MPs) and support vector machines for regression (SVR). In silica experiments were run using each algorithm with different combinations of 12 descriptive input features. Ten-fold cross validation was used to evaluate the performance of each model. Results: The SVR models approximated the physician consensus ratings of unseen CGM plots better than the MP models. When judged by the root mean square error, the best SVR model performed comparably to individual physicians at matching consensus ratings. When applied to 262 different CGM plots as a screen for excessive GV, this model had accuracy, sensitivity, and specificity of 90.1%, 97.0%, and 74.1%, respectively. It significantly outperformed mean amplitude of glycemic excursion, standard deviation, distance traveled, and excursion frequency. Conclusions: This new CPGV metric could be used as a routine measure of overall glucose control to supplement glycosylated hemoglobin in clinical practice.
- Published
- 2013
41. Word Sense Disambiguation Using Wikipedia
- Author
-
Razvan Bunescu, Rada Mihalcea, and Bharath Dandala
- Subjects
Word-sense disambiguation ,Computer science ,business.industry ,Artificial intelligence ,Construct (philosophy) ,computer.software_genre ,business ,Word sense ,GeneralLiterature_REFERENCE(e.g.,dictionaries,encyclopedias,glossaries) ,computer ,Natural language processing ,SemEval - Abstract
This paper describes explorations in word sense disambiguation using Wikipedia as a source of sense annotations. Through experiments on four different languages, we show that the Wikipedia-based sense annotations are reliable and can be used to construct accurate sense classifiers.
- Published
- 2013
- Full Text
- View/download PDF
42. Mandarin tone recognition based on unsupervised feature learning from spectrograms
- Author
-
Li Xu, Charles Chen, Razvan Bunescu, and Chang Liu
- Subjects
Acoustics and Ultrasonics ,Artificial neural network ,Computer science ,Speech recognition ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Mandarin Chinese ,Convolutional neural network ,language.human_language ,Tone (musical instrument) ,ComputingMethodologies_PATTERNRECOGNITION ,Arts and Humanities (miscellaneous) ,language ,Feature (machine learning) ,Spectrogram ,Mel-frequency cepstrum ,Feature learning - Abstract
In tone languages, such as Mandarin Chinese, a syllable with different tones conveys different meanings. Previously, Mandarin tone recognition based on Mel-frequency cepstral coefficients (MFCCs) and Convolutional Neural Networks (CNN) was examined and the results outperformed the model of conventional neural network using manually edited F0 data. In the present study, Mandarin tone recognition based on spectrograms, instead of MFCCs, was explored. Unsupervised feature learning was applied to the unlabeled spectrograms directly with a denoising autoencoder (dAE). Then, the model convolved the labeled spectrograms with the learnt “sound features” and produced a set of feature maps. A dataset that consisted of 4500 monosyllabic words collected from 125 children was used to evaluate the recognition performance. Compared with methods based on MFCCs, there are more parameters to train in the new approach based on spectrograms. As a result, the new model might better capture the statistical distribution in the ...
- Published
- 2016
- Full Text
- View/download PDF
43. Automatic Detection of Excessive Glycemic Variability for Diabetes Management
- Author
-
Jay H. Shubrook, Frank L. Schwartz, Razvan Bunescu, Matthew T. Wiley, and Cindy Marling
- Subjects
Computer science ,business.industry ,Feature extraction ,Feature selection ,Machine learning ,computer.software_genre ,medicine.disease ,Support vector machine ,Naive Bayes classifier ,Diabetes management ,Diabetes mellitus ,Multilayer perceptron ,Pattern recognition (psychology) ,medicine ,Artificial intelligence ,Sugar ,business ,computer ,Glycemic - Abstract
Glycemic variability, or fluctuation in blood glucose levels, is a significant factor in diabetes management. Excessive glycemic variability contributes to oxidative stress, which has been linked to the development of long-term diabetic complications. An automated screen for excessive glycemic variability, based on the readings from continuous glucose monitoring (CGM) systems, would enable early identification of at risk patients. In this paper, we present an automatic approach for learning variability models that can routinely detect excessive glycemic variability when applied to CGM data. Naive Bayes (NB), Multilayer Perceptron (MP), and Support Vector Machine (SVM) models are trained and evaluated on a dataset of CGM plots that have been manually annotated with respect to glycemic variability by two diabetes experts. In order to alleviate the impact of noise, the CGM plots are smoothed using cubic splines. Automatic feature selection is then performed on a rich set of pattern recognition features. Empirical evaluation shows that the top performing model obtains a state of the art accuracy of 93.8%, substantially outperforming a previous NB model.
- Published
- 2011
- Full Text
- View/download PDF
44. The 4 Diabetes Support System: A Case Study in CBR Research and Development
- Author
-
Matthew T. Wiley, Tessa Cooper, Jay H. Shubrook, Cindy Marling, Frank L. Schwartz, and Razvan Bunescu
- Subjects
Insulin pump ,Type 1 diabetes ,Knowledge management ,Computer science ,business.industry ,medicine.disease ,Medical research ,Naive Bayes classifier ,Clinical research ,Diabetes management ,Diabetes mellitus ,medicine ,business ,Glycemic - Abstract
This paper presents the 4 Diabetes Support SystemTM (4DSS) project as a case study in case-based reasoning (CBR) research and development. This project aims to help patients with type 1 diabetes on insulin pump therapy achieve and maintain good blood glucose control. Over the course of seven years and three clinical research studies, a series of defining cases altered the research and development path. Each of these cases suggested a new, unanticipated research direction or clinical application. New AI technologies, including naive Bayes classification and support vector regression, were incorporated. New medical research into glycemic variability and blood glucose prediction was undertaken. The CBR research paradigm has provided a strong framework for medical research as well as for artificial intelligence (AI) research. This new work has the potential to positively impact the health and wellbeing of patients with diabetes. This paper shares the 4DSS project experience.
- Published
- 2011
- Full Text
- View/download PDF
45. Learning with probabilistic features for improved pipeline models
- Author
-
Razvan Bunescu
- Subjects
business.industry ,Computer science ,Pipeline (computing) ,Probabilistic logic ,computer.software_genre ,Machine learning ,Named-entity recognition ,Dependency grammar ,Upstream (networking) ,Artificial intelligence ,Data mining ,State (computer science) ,business ,Baseline (configuration management) ,computer - Abstract
We present a novel learning framework for pipeline models aimed at improving the communication between consecutive stages in a pipeline. Our method exploits the confidence scores associated with outputs at any given stage in a pipeline in order to compute probabilistic features used at other stages downstream. We describe a simple method of integrating probabilistic features into the linear scoring functions used by state of the art machine learning algorithms. Experimental evaluation on dependency parsing and named entity recognition demonstrate the superiority of our approach over the baseline pipeline models, especially when upstream stages in the pipeline exhibit low accuracy.
- Published
- 2008
- Full Text
- View/download PDF
46. From Word Embeddings To Document Similarities for Improved Information Retrieval in Software Engineering.
- Author
-
Xin Ye, Hui Shen, Xiao Ma, Razvan Bunescu, and Chang Liu
- Published
- 2016
- Full Text
- View/download PDF
47. Multiple instance learning for sparse positive bags
- Author
-
Razvan Bunescu and Raymond J. Mooney
- Subjects
Constraint (information theory) ,Support vector machine ,business.industry ,Pattern recognition ,Artificial intelligence ,Machine learning ,computer.software_genre ,business ,computer ,Mathematics ,Image (mathematics) - Abstract
We present a new approach to multiple instance learning (MIL) that is particularly effective when the positive bags are sparse (i.e. contain few positive instances). Unlike other SVM-based MIL methods, our approach more directly enforces the desired constraint that at least one of the instances in a positive bag is positive. Using both artificial and real-world data, we experimentally demonstrate that our approach achieves greater accuracy than state-of-the-art MIL methods when positive bags are sparse, and performs competitively when they are not. In particular, our approach is the best performing method for image region classification.
- Published
- 2007
- Full Text
- View/download PDF
48. Extracting Relations from Text: From Word Sequences to Dependency Paths
- Author
-
Raymond J. Mooney and Razvan Bunescu
- Subjects
Parsing ,Relation (database) ,Computer science ,business.industry ,computer.software_genre ,Relationship extraction ,Task (project management) ,Information extraction ,Chunking (psychology) ,Artificial intelligence ,Set (psychology) ,business ,computer ,Sentence ,Natural language processing - Abstract
Extracting semantic relationships between entities mentioned in text documents is an important task in natural language processing. The various types of relationships that are discovered between mentions of entities can provide useful structured information to a text mining system [1]. Traditionally, the task specifies a predefined set of entity types and relation types that are deemed to be relevant to a potential user and that are likely to occur in a particular text collection. For example, information extraction from newspaper articles is usually concerned with identifying mentions of people, organizations, locations, and extracting useful relations between them. Relevant relation types range from social relationships, to roles that people hold inside an organization, to relations between organizations, to physical locations of people and organizations. Scientific publications in the biomedical domain offer a type of narrative that is very different from the newspaper discourse. A significant effort is currently spent on automatically extracting relevant pieces of information from Medline, an online collection of biomedical abstracts. Proteins, genes, and cells are examples of relevant entities in this task, whereas subcellular localizations and protein-protein interactions are two of the relation types that have received significant attention recently. The inherent difficulty of the relation extraction task is further compounded in the biomedical domain by the relative scarcity of tools able to analyze the corresponding type of narrative. Most existing natural language processing tools, such as tokenizers, sentence segmenters, part-of-speech (POS) taggers, shallow or full parsers are trained on newspaper corpora, and consequently they inccur a loss in accuracy when applied to biomedical literature. Therefore, information extraction systems developed for biological corpora need to be robust to POS or parsing errors, or to give reasonable performance using shallower but more reliable information, such as chunking instead of full parsing.
- Published
- 2007
- Full Text
- View/download PDF
49. Integrating co-occurrence statistics with information extraction for robust retrieval of protein interactions from Medline
- Author
-
Razvan Bunescu, Raymond Mooney, Arun Ramani, and Edward Marcotte
- Published
- 2006
- Full Text
- View/download PDF
50. Using biomedical literature mining to consolidate the set of known human protein-protein interactions
- Author
-
Edward M. Marcotte, Razvan Bunescu, Arun K. Ramani, and Raymond J. Mooney
- Subjects
Conditional random field ,Set (abstract data type) ,Text mining ,Information retrieval ,String kernel ,business.industry ,Computer science ,Construct (python library) ,business ,Protein–protein interaction - Abstract
This paper presents the results of a large-scale effort to construct a comprehensive database of known human protein interactions by combining and linking known interactions from existing databases and then adding to them by automatically mining additional interactions from 750,000 Medline abstracts. The end result is a network of 31,609 interactions amongst 7,748 proteins. The text mining system first identifies protein names in the text using a trained Conditional Random Field (CRF) and then identifies interactions through a filtered co-citation analysis. We also report two new strategies for mining interactions, either by finding explicit statements of interactions in the text using learned pattern-based rules or a Support-Vector Machine using a string kernel. Using information in existing ontologies, the automatically extracted data is shown to be of equivalent accuracy to manually curated data sets.
- Published
- 2005
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.