75 results on '"Sen, Koushik"'
Search Results
2. Analytic approximations for massive close post-mass transfer binary systems
- Author
-
Schürmann, Christoph, Langer, Norbert, Kramer, Joana A., Marchant, Pablo, Wang, Chen, Sen, Koushik, Schürmann, Christoph, Langer, Norbert, Kramer, Joana A., Marchant, Pablo, Wang, Chen, and Sen, Koushik
- Abstract
Massive binary evolution models are needed to predict massive star populations in star forming galaxies, the supernova diversity, and the number and properties of gravitational wave sources. Such models are often computed using so called rapid binary evolution codes, which approximate the evolution of the binary components based on detailed single star models. However, about one third of the interacting massive binary stars undergo mass transfer during core hydrogen burning (Case A mass transfer), whose outcome is difficult to derive from single star models. Here, we use a large grid of detailed binary evolution models for primaries in the initial mass range 10 to 40 Solar masses of LMC and SMC composition, to derive analytic fits for the key quantities needed in rapid binary evolution codes, i.e., the duration of core hydrogen burning, and the resulting donor star mass. Systems with shorter orbital periods produce up to 50% lighter stripped donors and have a up to 30% larger lifetime than wider systems. We find that both quantities depend strongly on the initial binary orbital period, but that the initial mass ratio and the mass transfer efficiency of the binary have little impact on the outcome. Our results are easily parameterisable and can be used to capture the effects of Case A mass transfer more accurately in rapid binary evolution codes., Comment: 10 pages
- Published
- 2024
3. LiveCodeBench: Holistic and Contamination Free Evaluation of Large Language Models for Code
- Author
-
Jain, Naman, Han, King, Gu, Alex, Li, Wen-Ding, Yan, Fanjia, Zhang, Tianjun, Wang, Sida, Solar-Lezama, Armando, Sen, Koushik, Stoica, Ion, Jain, Naman, Han, King, Gu, Alex, Li, Wen-Ding, Yan, Fanjia, Zhang, Tianjun, Wang, Sida, Solar-Lezama, Armando, Sen, Koushik, and Stoica, Ion
- Abstract
Large Language Models (LLMs) applied to code-related applications have emerged as a prominent field, attracting significant interest from both academia and industry. However, as new and improved LLMs are developed, existing evaluation benchmarks (e.g., HumanEval, MBPP) are no longer sufficient for assessing their capabilities. In this work, we propose LiveCodeBench, a comprehensive and contamination-free evaluation of LLMs for code, which continuously collects new problems over time from contests across three competition platforms, namely LeetCode, AtCoder, and CodeForces. Notably, our benchmark also focuses on a broader range of code related capabilities, such as self-repair, code execution, and test output prediction, beyond just code generation. Currently, LiveCodeBench hosts four hundred high-quality coding problems that were published between May 2023 and May 2024. We have evaluated 18 base LLMs and 34 instruction-tuned LLMs on LiveCodeBench. We present empirical findings on contamination, holistic performance comparisons, potential overfitting in existing benchmarks as well as individual model comparisons. We will release all prompts and model completions for further community analysis, along with a general toolkit for adding new scenarios and model, Comment: Website - https://livecodebench.github.io
- Published
- 2024
4. The Counterfeit Conundrum: Can Code Language Models Grasp the Nuances of Their Incorrect Generations?
- Author
-
Gu, Alex, Li, Wen-Ding, Jain, Naman, Olausson, Theo X., Lee, Celine, Sen, Koushik, Solar-Lezama, Armando, Gu, Alex, Li, Wen-Ding, Jain, Naman, Olausson, Theo X., Lee, Celine, Sen, Koushik, and Solar-Lezama, Armando
- Abstract
While language models are increasingly more proficient at code generation, they still frequently generate incorrect programs. Many of these programs are obviously wrong, but others are more subtle and pass weaker correctness checks such as being able to compile. In this work, we focus on these counterfeit samples: programs sampled from a language model that 1) have a high enough log-probability to be generated at a moderate temperature and 2) pass weak correctness checks. Overall, we discover that most models have a very shallow understanding of counterfeits through three clear failure modes. First, models mistakenly classify them as correct. Second, models are worse at reasoning about the execution behaviour of counterfeits and often predict their execution results as if they were correct. Third, when asking models to fix counterfeits, the likelihood of a model successfully repairing a counterfeit is often even lower than that of sampling a correct program from scratch. Counterfeits also have very unexpected properties: first, counterfeit programs for problems that are easier for a model to solve are not necessarily easier to detect and only slightly easier to execute and repair. Second, counterfeits from a given model are just as confusing to the model itself as they are to other models. Finally, both strong and weak models are able to generate counterfeit samples that equally challenge all models. In light of our findings, we recommend that care and caution be taken when relying on models to understand their own samples, especially when no external feedback is incorporated., Comment: 54 pages, 25 figures
- Published
- 2024
5. LLM4Fuzz: Guided Fuzzing of Smart Contracts with Large Language Models
- Author
-
Shou, Chaofan, Liu, Jing, Lu, Doudou, Sen, Koushik, Shou, Chaofan, Liu, Jing, Lu, Doudou, and Sen, Koushik
- Abstract
As blockchain platforms grow exponentially, millions of lines of smart contract code are being deployed to manage extensive digital assets. However, vulnerabilities in this mission-critical code have led to significant exploitations and asset losses. Thorough automated security analysis of smart contracts is thus imperative. This paper introduces LLM4Fuzz to optimize automated smart contract security analysis by leveraging large language models (LLMs) to intelligently guide and prioritize fuzzing campaigns. While traditional fuzzing suffers from low efficiency in exploring the vast state space, LLM4Fuzz employs LLMs to direct fuzzers towards high-value code regions and input sequences more likely to trigger vulnerabilities. Additionally, LLM4Fuzz can leverage LLMs to guide fuzzers based on user-defined invariants, reducing blind exploration overhead. Evaluations of LLM4Fuzz on real-world DeFi projects show substantial gains in efficiency, coverage, and vulnerability detection compared to baseline fuzzing. LLM4Fuzz also uncovered five critical vulnerabilities that can lead to a loss of more than $247k.
- Published
- 2024
6. SlimFit: Memory-Efficient Fine-Tuning of Transformer-based Models Using Training Dynamics
- Author
-
Ardakani, Arash, Haan, Altan, Tan, Shangyin, Popovici, Doru Thom, Cheung, Alvin, Iancu, Costin, Sen, Koushik, Ardakani, Arash, Haan, Altan, Tan, Shangyin, Popovici, Doru Thom, Cheung, Alvin, Iancu, Costin, and Sen, Koushik
- Abstract
Transformer-based models, such as BERT and ViT, have achieved state-of-the-art results across different natural language processing (NLP) and computer vision (CV) tasks. However, these models are extremely memory intensive during their fine-tuning process, making them difficult to deploy on GPUs with limited memory resources. To address this issue, we introduce a new tool called SlimFit that reduces the memory requirements of these models by dynamically analyzing their training dynamics and freezing less-contributory layers during fine-tuning. The layers to freeze are chosen using a runtime inter-layer scheduling algorithm. SlimFit adopts quantization and pruning for particular layers to balance the load of dynamic activations and to minimize the memory footprint of static activations, where static activations refer to those that cannot be discarded regardless of freezing. This allows SlimFit to freeze up to 95% of layers and reduce the overall on-device GPU memory usage of transformer-based models such as ViT and BERT by an average of 2.2x, across different NLP and CV benchmarks/datasets such as GLUE, SQuAD 2.0, CIFAR-10, CIFAR-100 and ImageNet with an average degradation of 0.2% in accuracy. For such NLP and CV tasks, SlimFit can reduce up to 3.1x the total on-device memory usage with an accuracy degradation of only up to 0.4%. As a result, while fine-tuning of ViT on ImageNet and BERT on SQuAD 2.0 with a batch size of 128 requires 3 and 2 32GB GPUs respectively, SlimFit enables their fine-tuning on a single 32GB GPU without any significant accuracy degradation.
- Published
- 2023
7. CodeScholar: Growing Idiomatic Code Examples
- Author
-
Shetty, Manish, Sen, Koushik, Stoica, Ion, Shetty, Manish, Sen, Koushik, and Stoica, Ion
- Abstract
Programmers often search for usage examples for API methods. A tool that could generate realistic, idiomatic, and contextual usage examples for one or more APIs would be immensely beneficial to developers. Such a tool would relieve the need for a deep understanding of the API landscape, augment existing documentation, and help discover interactions among APIs. We present CodeScholar, a tool that generates idiomatic code examples demonstrating the common usage of API methods. It includes a novel neural-guided search technique over graphs that grows the query APIs into idiomatic code examples. Our user study demonstrates that in 70% of cases, developers prefer CodeScholar generated examples over state-of-the-art large language models (LLM) like GPT3.5. We quantitatively evaluate 60 single and 25 multi-API queries from 6 popular Python libraries and show that across-the-board CodeScholar generates more realistic, diverse, and concise examples. In addition, we show that CodeScholar not only helps developers but also LLM-powered programming assistants generate correct code in a program synthesis setting.
- Published
- 2023
8. DSPy Assertions: Computational Constraints for Self-Refining Language Model Pipelines
- Author
-
Singhvi, Arnav, Shetty, Manish, Tan, Shangyin, Potts, Christopher, Sen, Koushik, Zaharia, Matei, Khattab, Omar, Singhvi, Arnav, Shetty, Manish, Tan, Shangyin, Potts, Christopher, Sen, Koushik, Zaharia, Matei, and Khattab, Omar
- Abstract
Chaining language model (LM) calls as composable modules is fueling a new way of programming, but ensuring LMs adhere to important constraints requires heuristic "prompt engineering". We introduce LM Assertions, a programming construct for expressing computational constraints that LMs should satisfy. We integrate our constructs into the recent DSPy programming model for LMs, and present new strategies that allow DSPy to compile programs with LM Assertions into more reliable and accurate systems. We also propose strategies to use assertions at inference time for automatic self-refinement with LMs. We report on four diverse case studies for text generation and find that LM Assertions improve not only compliance with imposed rules but also downstream task performance, passing constraints up to 164% more often and generating up to 37% more higher-quality responses. Our reference implementation of LM Assertions is integrated into DSPy at https://github.com/stanfordnlp/dspy, Comment: Arnav*, Manish*, Shangyin* contributed equally to this work
- Published
- 2023
9. LLM-Assisted Code Cleaning For Training Accurate Code Generators
- Author
-
Jain, Naman, Zhang, Tianjun, Chiang, Wei-Lin, Gonzalez, Joseph E., Sen, Koushik, Stoica, Ion, Jain, Naman, Zhang, Tianjun, Chiang, Wei-Lin, Gonzalez, Joseph E., Sen, Koushik, and Stoica, Ion
- Abstract
Natural language to code generation is an important application area of LLMs and has received wide attention from the community. The majority of relevant studies have exclusively concentrated on increasing the quantity and functional correctness of training sets while disregarding other stylistic elements of programs. More recently, data quality has garnered a lot of interest and multiple works have showcased its importance for improving performance. In this work, we investigate data quality for code and find that making the code more structured and readable leads to improved code generation performance of the system. We build a novel data-cleaning pipeline that uses these principles to transform existing programs by 1.) renaming variables, 2.) modularizing and decomposing complex code into smaller helper sub-functions, and 3.) inserting natural-language based plans via LLM based transformations. We evaluate our approach on two challenging algorithmic code generation benchmarks and find that fine-tuning CodeLLaMa-7B on our transformed modularized programs improves the performance by up to 30% compared to fine-tuning on the original dataset. Additionally, we demonstrate improved performance from using a smaller amount of higher-quality data, finding that a model fine-tuned on the entire original dataset is outperformed by a model trained on 15% of our cleaned dataset. Even in comparison to closed-source models, our models outperform the much larger AlphaCoder models.
- Published
- 2023
10. FlorDB: Multiversion Hindsight Logging for Continuous Training
- Author
-
Garcia, Rolando, Dandamudi, Anusha, Matute, Gabriel, Wan, Lehan, Gonzalez, Joseph, Hellerstein, Joseph M., Sen, Koushik, Garcia, Rolando, Dandamudi, Anusha, Matute, Gabriel, Wan, Lehan, Gonzalez, Joseph, Hellerstein, Joseph M., and Sen, Koushik
- Abstract
Production Machine Learning involves continuous training: hosting multiple versions of models over time, often with many model versions running at once. When model performance does not meet expectations, Machine Learning Engineers (MLEs) debug issues by exploring and analyzing numerous prior versions of code and training data to identify root causes and mitigate problems. Traditional debugging and logging tools often fall short in managing this experimental, multi-version context. FlorDB introduces Multiversion Hindsight Logging, which allows engineers to use the most recent version's logging statements to query past versions, even when older versions logged different data. Log statement propagation enables consistent injection of logging statements into past code versions, regardless of changes to the codebase. Once log statements are propagated across code versions, the remaining challenge in Multiversion Hindsight Logging is to efficiently replay the new log statements based on checkpoints from previous runs. Finally, a coherent user experience is required to help MLEs debug across all versions of code and data. To this end, FlorDB presents a unified relational model for efficient handling of historical queries, offering a comprehensive view of the log history to simplify the exploration of past code iterations. We present a performance evaluation on diverse benchmarks confirming its scalability and the ability to deliver real-time query responses, leveraging query-based filtering and checkpoint-based parallelism for efficient replay.
- Published
- 2023
11. ItyFuzz: Snapshot-Based Fuzzer for Smart Contract
- Author
-
Shou, Chaofan, Tan, Shangyin, Sen, Koushik, Shou, Chaofan, Tan, Shangyin, and Sen, Koushik
- Abstract
Smart contracts are critical financial instruments, and their security is of utmost importance. However, smart contract programs are difficult to fuzz due to the persistent blockchain state behind all transactions. Mutating sequences of transactions are complex and often lead to a suboptimal exploration for both input and program spaces. In this paper, we introduce a novel snapshot-based fuzzer ItyFuzz for testing smart contracts. In ItyFuzz, instead of storing sequences of transactions and mutating from them, we snapshot states and singleton transactions. To explore interesting states, ItyFuzz introduces a dataflow waypoint mechanism to identify states with more potential momentum. ItyFuzz also incorporates comparison waypoints to prune the space of states. By maintaining snapshots of the states, ItyFuzz can synthesize concrete exploits like reentrancy attacks quickly. Because ItyFuzz has second-level response time to test a smart contract, it can be used for on-chain testing, which has many benefits compared to local development testing. Finally, we evaluate ItyFuzz on real-world smart contracts and some hacked on-chain DeFi projects. ItyFuzz outperforms existing fuzzers in terms of instructional coverage and can find and generate realistic exploits for on-chain projects quickly., Comment: ISSTA 2023
- Published
- 2023
12. Efficient and Transferable Adversarial Examples from Bayesian Neural Networks
- Author
-
Gubri, Martin, Cordy, Maxime, Papadakis, Mike, Le Traon, Yves, Sen, Koushik, Gubri, Martin, Cordy, Maxime, Papadakis, Mike, Le Traon, Yves, and Sen, Koushik
- Abstract
An established way to improve the transferability of black-box evasion attacks is to craft the adversarial examples on an ensemble-based surrogate to increase diversity. We argue that transferability is fundamentally related to uncertainty. Based on a state-of-the-art Bayesian Deep Learning technique, we propose a new method to efficiently build a surrogate by sampling approximately from the posterior distribution of neural network weights, which represents the belief about the value of each parameter. Our extensive experiments on ImageNet, CIFAR-10 and MNIST show that our approach improves the success rates of four state-of-the-art attacks significantly (up to 83.2 percentage points), in both intra-architecture and inter-architecture transferability. On ImageNet, our approach can reach 94% of success rate while reducing training computations from 11.6 to 2.4 exaflops, compared to an ensemble of independently trained DNNs. Our vanilla surrogate achieves 87.5% of the time higher transferability than three test-time techniques designed for this purpose. Our work demonstrates that the way to train a surrogate has been overlooked, although it is an important element of transfer-based attacks. We are, therefore, the first to review the effectiveness of several training methods in increasing transferability. We provide new directions to better understand the transferability phenomenon and offer a simple but strong baseline for future work.
- Published
- 2022
13. LGV: Boosting Adversarial Example Transferability from Large Geometric Vicinity
- Author
-
Gubri, Martin, Cordy, Maxime, Papadakis, Mike, Traon, Yves Le, Sen, Koushik, Gubri, Martin, Cordy, Maxime, Papadakis, Mike, Traon, Yves Le, and Sen, Koushik
- Abstract
We propose transferability from Large Geometric Vicinity (LGV), a new technique to increase the transferability of black-box adversarial attacks. LGV starts from a pretrained surrogate model and collects multiple weight sets from a few additional training epochs with a constant and high learning rate. LGV exploits two geometric properties that we relate to transferability. First, models that belong to a wider weight optimum are better surrogates. Second, we identify a subspace able to generate an effective surrogate ensemble among this wider optimum. Through extensive experiments, we show that LGV alone outperforms all (combinations of) four established test-time transformations by 1.8 to 59.9\% points. Our findings shed new light on the importance of the geometry of the weight space to explain the transferability of adversarial examples.
- Published
- 2022
14. Efficient and Transferable Adversarial Examples from Bayesian Neural Networks
- Author
-
Gubri, Martin, Cordy, Maxime, Papadakis, Mike, Le Traon, Yves, Sen, Koushik, Gubri, Martin, Cordy, Maxime, Papadakis, Mike, Le Traon, Yves, and Sen, Koushik
- Abstract
An established way to improve the transferability of black-box evasion attacks is to craft the adversarial examples on an ensemble-based surrogate to increase diversity. We argue that transferability is fundamentally related to uncertainty. Based on a state-of-the-art Bayesian Deep Learning technique, we propose a new method to efficiently build a surrogate by sampling approximately from the posterior distribution of neural network weights, which represents the belief about the value of each parameter. Our extensive experiments on ImageNet, CIFAR-10 and MNIST show that our approach improves the success rates of four state-of-the-art attacks significantly (up to 83.2 percentage points), in both intra-architecture and inter-architecture transferability. On ImageNet, our approach can reach 94% of success rate while reducing training computations from 11.6 to 2.4 exaflops, compared to an ensemble of independently trained DNNs. Our vanilla surrogate achieves 87.5% of the time higher transferability than three test-time techniques designed for this purpose. Our work demonstrates that the way to train a surrogate has been overlooked, although it is an important element of transfer-based attacks. We are, therefore, the first to review the effectiveness of several training methods in increasing transferability. We provide new directions to better understand the transferability phenomenon and offer a simple but strong baseline for future work.
- Published
- 2022
15. PALT: Parameter-Lite Transfer of Language Models for Knowledge Graph Completion
- Author
-
Shen, Jianhao, Wang, Chenguang, Yuan, Ye, Han, Jiawei, Ji, Heng, Sen, Koushik, Zhang, Ming, Song, Dawn, Shen, Jianhao, Wang, Chenguang, Yuan, Ye, Han, Jiawei, Ji, Heng, Sen, Koushik, Zhang, Ming, and Song, Dawn
- Abstract
This paper presents a parameter-lite transfer learning approach of pretrained language models (LM) for knowledge graph (KG) completion. Instead of finetuning, which modifies all LM parameters, we only tune a few new parameters while keeping the original LM parameters fixed. We establish this via reformulating KG completion as a "fill-in-the-blank" task, and introducing a parameter-lite encoder on top of the original LMs. We show that, by tuning far fewer parameters than finetuning, LMs transfer non-trivially to most tasks and reach competitiveness with prior state-of-the-art approaches. For instance, we outperform the fully finetuning approaches on a KG completion benchmark by tuning only 1% of the parameters. The code and datasets are available at \url{https://github.com/yuanyehome/PALT}., Comment: Findings of EMNLP 2022
- Published
- 2022
16. LGV: Boosting Adversarial Example Transferability from Large Geometric Vicinity
- Author
-
Gubri, Martin, Cordy, Maxime, Papadakis, Mike, Traon, Yves Le, Sen, Koushik, Gubri, Martin, Cordy, Maxime, Papadakis, Mike, Traon, Yves Le, and Sen, Koushik
- Abstract
We propose transferability from Large Geometric Vicinity (LGV), a new technique to increase the transferability of black-box adversarial attacks. LGV starts from a pretrained surrogate model and collects multiple weight sets from a few additional training epochs with a constant and high learning rate. LGV exploits two geometric properties that we relate to transferability. First, models that belong to a wider weight optimum are better surrogates. Second, we identify a subspace able to generate an effective surrogate ensemble among this wider optimum. Through extensive experiments, we show that LGV alone outperforms all (combinations of) four established test-time transformations by 1.8 to 59.9\% points. Our findings shed new light on the importance of the geometry of the weight space to explain the transferability of adversarial examples.
- Published
- 2022
17. Benchmarking Language Models for Code Syntax Understanding
- Author
-
Shen, Da, Chen, Xinyun, Wang, Chenguang, Sen, Koushik, Song, Dawn, Shen, Da, Chen, Xinyun, Wang, Chenguang, Sen, Koushik, and Song, Dawn
- Abstract
Pre-trained language models have demonstrated impressive performance in both natural language processing and program understanding, which represent the input as a token sequence without explicitly modeling its structure. Some prior works show that pre-trained language models can capture the syntactic rules of natural languages without finetuning on syntax understanding tasks. However, there is limited understanding of how well pre-trained models understand the code structure so far. In this work, we perform the first thorough benchmarking of the state-of-the-art pre-trained models for identifying the syntactic structures of programs. Specifically, we introduce CodeSyntax, a large-scale dataset of programs annotated with the syntactic relationships in their corresponding abstract syntax trees. Our key observation is that existing language models pretrained on code still lack the understanding of code syntax. In fact, these pre-trained programming language models fail to match the performance of simple baselines based on positional offsets and keywords. We also present a natural language benchmark to highlight the differences between natural languages and programming languages in terms of syntactic structure understanding. Our findings point out key limitations of existing pre-training methods for programming languages, and suggest the importance of modeling code syntactic structures., Comment: Findings of EMNLP 2022
- Published
- 2022
18. LGV: Boosting Adversarial Example Transferability from Large Geometric Vicinity
- Author
-
Gubri, Martin, Cordy, Maxime, Papadakis, Mike, Traon, Yves Le, Sen, Koushik, Gubri, Martin, Cordy, Maxime, Papadakis, Mike, Traon, Yves Le, and Sen, Koushik
- Abstract
We propose transferability from Large Geometric Vicinity (LGV), a new technique to increase the transferability of black-box adversarial attacks. LGV starts from a pretrained surrogate model and collects multiple weight sets from a few additional training epochs with a constant and high learning rate. LGV exploits two geometric properties that we relate to transferability. First, models that belong to a wider weight optimum are better surrogates. Second, we identify a subspace able to generate an effective surrogate ensemble among this wider optimum. Through extensive experiments, we show that LGV alone outperforms all (combinations of) four established test-time transformations by 1.8 to 59.9 percentage points. Our findings shed new light on the importance of the geometry of the weight space to explain the transferability of adversarial examples., Comment: Accepted at ECCV 2022
- Published
- 2022
19. The Sky Above The Clouds
- Author
-
Chasins, Sarah, Cheung, Alvin, Crooks, Natacha, Ghodsi, Ali, Goldberg, Ken, Gonzalez, Joseph E., Hellerstein, Joseph M., Jordan, Michael I., Joseph, Anthony D., Mahoney, Michael W., Parameswaran, Aditya, Patterson, David, Popa, Raluca Ada, Sen, Koushik, Shenker, Scott, Song, Dawn, Stoica, Ion, Chasins, Sarah, Cheung, Alvin, Crooks, Natacha, Ghodsi, Ali, Goldberg, Ken, Gonzalez, Joseph E., Hellerstein, Joseph M., Jordan, Michael I., Joseph, Anthony D., Mahoney, Michael W., Parameswaran, Aditya, Patterson, David, Popa, Raluca Ada, Sen, Koushik, Shenker, Scott, Song, Dawn, and Stoica, Ion
- Abstract
Technology ecosystems often undergo significant transformations as they mature. For example, telephony, the Internet, and PCs all started with a single provider, but in the United States each is now served by a competitive market that uses comprehensive and universal technology standards to provide compatibility. This white paper presents our view on how the cloud ecosystem, barely over fifteen years old, could evolve as it matures., Comment: 35 pages
- Published
- 2022
20. The Nature of Unseen Companions in Massive Single-Line Spectroscopic Binaries
- Author
-
Sana, Hugues, Abdul-Masih, Michael, Banyard, Gareth, Bodensteiner, Julia, Bowman, Dominic M., Dsilva, Karan, Eldridge, C., Fabry, Matthias, Frost, Abigail J., Hawcroft, Calum, Janssens, Soetkin, Mahy, Laurent, Marchant, Pablo, Langer, Norbert, Van Reeth, Timothy, Sen, Koushik, Shenar, Tomer, Sana, Hugues, Abdul-Masih, Michael, Banyard, Gareth, Bodensteiner, Julia, Bowman, Dominic M., Dsilva, Karan, Eldridge, C., Fabry, Matthias, Frost, Abigail J., Hawcroft, Calum, Janssens, Soetkin, Mahy, Laurent, Marchant, Pablo, Langer, Norbert, Van Reeth, Timothy, Sen, Koushik, and Shenar, Tomer
- Abstract
Massive stars are predominantly found in binaries and higher order multiples. While the period and eccentricity distributions of OB stars are now well established across different metallicity regimes, the determination of mass-ratios has been mostly limited to double-lined spectroscopic binaries. As a consequence, the mass-ratio distribution remains subject to significant uncertainties. Open questions include the shape and extent of the companion mass-function towards its low-mass end and the nature of undetected companions in single-lined spectroscopic binaries. In this contribution, we present the results of a large and systematic analysis of a sample of over 80 single-lined O-type spectroscopic binaries (SB1s) in the Milky Way and in the Large Magellanic Cloud (LMC). We report on the developed methodology, the constraints obtained on the nature of SB1 companions, the distribution of O star mass-ratios at LMC metallicity and the occurrence of quiescent OB+black hole binaries., Comment: To appear in the proceedings of IAUS361: Massive stars near and far; 6 pages, 1 figure
- Published
- 2022
21. The Tarantula Massive Binary Monitoring VI: Characterisation of hidden companions in 51 single-lined O-type binaries, a flat mass-ratio distribution, and black-hole binary candidates
- Author
-
Shenar, Tomer, Sana, Hugues, Mahy, Laurent, Apellaniz, Jesus Maiz, Crowther, Paul A., Gromadzki, Mariusz, Herrero, Artemio, Langer, Norbert, Marchant, Pablo, Schneider, Fabian R. N., Sen, Koushik, Soszynski, Igor, Toonen, S., Shenar, Tomer, Sana, Hugues, Mahy, Laurent, Apellaniz, Jesus Maiz, Crowther, Paul A., Gromadzki, Mariusz, Herrero, Artemio, Langer, Norbert, Marchant, Pablo, Schneider, Fabian R. N., Sen, Koushik, Soszynski, Igor, and Toonen, S.
- Abstract
We aim to hunt for massive binaries hosting a black hole companion (OB+BH) and establish the natal mass-ratio distribution of massive stars at the subsolar metallicity environment of the Large Magellanic Cloud (LMC). We use the shift-and-add grid disentangling technique to characterize the hidden companions in 51 SB1 O-type and evolved B-type binaries in the LMC monitored in the framework of the Tarantula Massive Binary Monitoring (TMBM). Out of the 51 SB1 systems, 43 (84%) are found to have non-degenerate stellar companions, of which 28 are confident detections, and 15 are less certain (SB1: or SB2:). Of these 43 targets, one is found to be a triple (VFTS 64), and two are found to be quadruples (VFTS 120, 702). The remaining eight targets (16%) retain an SB1 classification. Aside from the unambiguous case of VFTS 243, analysed in detailed in a separate paper, we identify two additional OB+BH candidates: VFTS 514 and VFTS 779. Additional black holes may be present in the sample but at a lower probability. Our study firmly establishes a virtually flat natal mass-ratio distribution for O-type stars at LMC metallicity, covering the entire mass-ratio range (0.05 < q < 1) and periods in the range 0 < log P < 3 [d]. The nature of the OB+BH candidates should be verified through future monitoring, but the frequency of OB+BH candidates is generally in line with recent predictions at LMC metallicity., Comment: 41 pages (14 main article + 27 appendix), accepted to A&A
- Published
- 2022
- Full Text
- View/download PDF
22. An X-ray quiet black hole born with a negligible kick in a massive binary within the Large Magellanic Cloud
- Author
-
Shenar, Tomer, Sana, Hugues, Mahy, Laurent, El-Badry, Kareem, Marchant, Pablo, Langer, Norbert, Hawcroft, Calum, Fabry, Matthias, Sen, Koushik, Almeida, Leonardo A., Abdul-Masih, Michael, Bodensteiner, Julia, Crowther, Paul A., Gieles, Mark, Gromadzki, Mariusz, Henault-Brunet, Vincent, Herrero, Artemio, de Koter, Alex, Iwanek, Patryk, Kozłowski, Szymon, Lennon, Daniel J., Apellaniz, Jesus Maız, Mroz, Przemysław, Moffat, Anthony F. J., Picco, Annachiara, Pietrukowicz, Paweł, Poleski, Radosław, Rybicki, Krzysztof, Schneider, Fabian R. N., Skowron, Dorota M., Skowron, Jan, Soszynski, Igor, Szymanski, Michał K., Toonen, Silvia, Udalski, Andrzej, Ulaczyk, Krzysztof, Vink, Jorick S., Wrona, Marcin, Shenar, Tomer, Sana, Hugues, Mahy, Laurent, El-Badry, Kareem, Marchant, Pablo, Langer, Norbert, Hawcroft, Calum, Fabry, Matthias, Sen, Koushik, Almeida, Leonardo A., Abdul-Masih, Michael, Bodensteiner, Julia, Crowther, Paul A., Gieles, Mark, Gromadzki, Mariusz, Henault-Brunet, Vincent, Herrero, Artemio, de Koter, Alex, Iwanek, Patryk, Kozłowski, Szymon, Lennon, Daniel J., Apellaniz, Jesus Maız, Mroz, Przemysław, Moffat, Anthony F. J., Picco, Annachiara, Pietrukowicz, Paweł, Poleski, Radosław, Rybicki, Krzysztof, Schneider, Fabian R. N., Skowron, Dorota M., Skowron, Jan, Soszynski, Igor, Szymanski, Michał K., Toonen, Silvia, Udalski, Andrzej, Ulaczyk, Krzysztof, Vink, Jorick S., and Wrona, Marcin
- Abstract
Stellar-mass black holes are the final remnants of stars born with more than 15 solar masses. Billions are expected to reside in the Local Group, yet only few are known, mostly detected through X-rays emitted as they accrete material from a companion star. Here, we report on VFTS 243: a massive X-ray faint binary in the Large Magellanic Cloud. With an orbital period of 10.4-d, it comprises an O-type star of 25 solar masses and an unseen companion of at least nine solar masses. Our spectral analysis excludes a non-degenerate companion at a 5-sigma confidence level. The minimum companion mass implies that it is a black hole. No other X-ray quiet black hole is unambiguously known outside our Galaxy. The (near-)circular orbit and kinematics of VFTS 243 imply that the collapse of the progenitor into a black hole was associated with little or no ejected material or black-hole kick. Identifying such unique binaries substantially impacts the predicted rates of gravitational-wave detections and properties of core-collapse supernovae across the Cosmos., Comment: Accepted to Nature Astronomy, 64 pages, 15 figures, 2 tables; ESO press release: https://www.eso.org/public/news/eso2210/; Nat Asr paper URL: https://www.nature.com/articles/s41550-022-01730-y
- Published
- 2022
- Full Text
- View/download PDF
23. Detailed evolutionary models of massive contact binaries - I. Model grids and synthetic populations for the Magellanic Clouds
- Author
-
Menon, Athira, Langer, Norbert, de Mink, Selma E., Justham, Stephen, Sen, Koushik, Szecsi, Dorottya, de Koter, Alex, Abdul-Masih, Michael, Sana, Hugues, Mahy, Laurent, Marchant, Pablo, Menon, Athira, Langer, Norbert, de Mink, Selma E., Justham, Stephen, Sen, Koushik, Szecsi, Dorottya, de Koter, Alex, Abdul-Masih, Michael, Sana, Hugues, Mahy, Laurent, and Marchant, Pablo
- Abstract
The majority of close massive binary stars with initial periods of a few days experience a contact phase, in which both stars overflow their Roche lobes simultaneously. We perform the first dedicated study of the evolution of massive contact binaries and provide a comprehensive prediction of their observed properties. We compute 2790 detailed binary models for the Large and Small Magellanic Clouds each, assuming mass transfer to be conservative. The initial parameter space for both grids span total masses from 20 to 80 M-circle dot, orbital periods of 0.6-2 d and mass ratios of 0.6-1.0. We find that models that remain in contact over nuclear time-scales evolve towards equal masses, echoing the mass ratios of their observed counterparts. Ultimately, the fate of our nuclear-time-scale models is to merge on the main sequence. Our predicted period-mass ratio distributions of O-type contact binaries are similar for both galaxies, and we expect 10 such systems together in both Magellanic Clouds. While we can largely reproduce the observed distribution, we overestimate the population of equal-mass contact binaries. This situation is somewhat remedied if we also account for binaries that are nearly in contact. Our theoretical distributions work particularly well for contact binaries with periods <2 d and total masses less than or similar to 45 M-circle dot. We expect stellar winds, non-conservative mass transfer, and envelope inflation to have played a role in the formation of the more massive and longer-period contact binaries.
- Published
- 2021
24. Detailed evolutionary models of massive contact binaries - I. Model grids and synthetic populations for the Magellanic Clouds
- Author
-
Menon, Athira, Langer, Norbert, de Mink, Selma E., Justham, Stephen, Sen, Koushik, Szecsi, Dorottya, de Koter, Alex, Abdul-Masih, Michael, Sana, Hugues, Mahy, Laurent, Marchant, Pablo, Menon, Athira, Langer, Norbert, de Mink, Selma E., Justham, Stephen, Sen, Koushik, Szecsi, Dorottya, de Koter, Alex, Abdul-Masih, Michael, Sana, Hugues, Mahy, Laurent, and Marchant, Pablo
- Abstract
The majority of close massive binary stars with initial periods of a few days experience a contact phase, in which both stars overflow their Roche lobes simultaneously. We perform the first dedicated study of the evolution of massive contact binaries and provide a comprehensive prediction of their observed properties. We compute 2790 detailed binary models for the Large and Small Magellanic Clouds each, assuming mass transfer to be conservative. The initial parameter space for both grids span total masses from 20 to 80 M-circle dot, orbital periods of 0.6-2 d and mass ratios of 0.6-1.0. We find that models that remain in contact over nuclear time-scales evolve towards equal masses, echoing the mass ratios of their observed counterparts. Ultimately, the fate of our nuclear-time-scale models is to merge on the main sequence. Our predicted period-mass ratio distributions of O-type contact binaries are similar for both galaxies, and we expect 10 such systems together in both Magellanic Clouds. While we can largely reproduce the observed distribution, we overestimate the population of equal-mass contact binaries. This situation is somewhat remedied if we also account for binaries that are nearly in contact. Our theoretical distributions work particularly well for contact binaries with periods <2 d and total masses less than or similar to 45 M-circle dot. We expect stellar winds, non-conservative mass transfer, and envelope inflation to have played a role in the formation of the more massive and longer-period contact binaries.
- Published
- 2021
25. Learning Highly Recursive Input Grammars
- Author
-
Kulkarni, Neil, Lemieux, Caroline, Sen, Koushik, Kulkarni, Neil, Lemieux, Caroline, and Sen, Koushik
- Abstract
This paper presents Arvada, an algorithm for learning context-free grammars from a set of positive examples and a Boolean-valued oracle. Arvada learns a context-free grammar by building parse trees from the positive examples. Starting from initially flat trees, Arvada builds structure to these trees with a key operation: it bubbles sequences of sibling nodes in the trees into a new node, adding a layer of indirection to the tree. Bubbling operations enable recursive generalization in the learned grammar. We evaluate Arvada against GLADE and find it achieves on average increases of 4.98x in recall and 3.13x in F1 score, while incurring only a 1.27x slowdown and requiring only 0.87x as many calls to the oracle. Arvada has a particularly marked improvement over GLADE on grammars with highly recursive structure, like those of programming languages.
- Published
- 2021
26. Growing a Test Corpus with Bonsai Fuzzing
- Author
-
Vikram, Vasudev, Padhye, Rohan, Sen, Koushik, Vikram, Vasudev, Padhye, Rohan, and Sen, Koushik
- Abstract
This paper presents a coverage-guided grammar-based fuzzing technique for automatically generating a corpus of concise test inputs for programs such as compilers. We walk-through a case study of a compiler designed for education and the corresponding problem of generating meaningful test cases to provide to students. The prior state-of-the-art solution is a combination of fuzzing and test-case reduction techniques such as variants of delta-debugging. Our key insight is that instead of attempting to minimize convoluted fuzzer-generated test inputs, we can instead grow concise test inputs by construction using a form of iterative deepening. We call this approach Bonsai Fuzzing. Experimental results show that Bonsai Fuzzing can generate test corpora having inputs that are 16--45% smaller in size on average as compared to a fuzz-then-reduce approach, while achieving approximately the same code coverage and fault-detection capability., Comment: Accepted at the 43rd International Conference on Software Engineering (ICSE 2021)
- Published
- 2021
27. Selecting fault revealing mutants
- Author
-
Titcheu Chekam, Thierry, Papadakis, Mike, Bissyande, Tegawendé François D Assise, Le Traon, Yves, Sen, Koushik, Titcheu Chekam, Thierry, Papadakis, Mike, Bissyande, Tegawendé François D Assise, Le Traon, Yves, and Sen, Koushik
- Published
- 2020
- Full Text
- View/download PDF
28. Detailed evolutionary models of massive contact binaries: I. Model grids and synthetic populations for the Magellanic Clouds
- Author
-
Menon, Athira, Langer, Norbert, de Mink, Selma E., Justham, Stephen, Sen, Koushik, Szécsi, Dorottya, de Koter, Alex, Abdul-Masih, Michael, Sana, Hugues, Mahy, Laurent, Marchant, Pablo, Menon, Athira, Langer, Norbert, de Mink, Selma E., Justham, Stephen, Sen, Koushik, Szécsi, Dorottya, de Koter, Alex, Abdul-Masih, Michael, Sana, Hugues, Mahy, Laurent, and Marchant, Pablo
- Abstract
The majority of close massive binary stars with initial periods of a few days experience a contact phase, in which both stars overflow their Roche lobes simultaneously. We perform the first dedicated study of the evolution of massive contact binaries and provide a comprehensive prediction of their observed properties. We compute 2790 detailed binary models for the Large and Small Magellanic Clouds each, assuming a conservative mass transfer. The initial parameter space for both grids span total masses from 20 to 80$\,\textrm{M}_{\odot}$, orbital periods of 0.6 to 2 days and mass ratios of 0.6 to 1.0. We find that models that remain in contact over nuclear timescales evolve towards equal masses, echoing the mass ratios of their observed counterparts. Ultimately, the fate of our nuclear-timescale models is to merge on the main sequence. Our predicted period-mass ratio distributions of O-type contact binaries are similar for both galaxies, and we expect 10 such systems together in both Magellanic Clouds. While we can largely reproduce the observed distribution, we over-estimate the population of equal-mass contact binaries. This situation is somewhat remedied if we also account for binaries that are approaching contact. Our theoretical distributions work particularly well for contact binaries with periods $<$2 days and total masses $\lessapprox45\,\textrm{M}_{\odot}$. We expect stellar winds, non-conservative mass transfer and envelope inflation to have played a role in the formation of the more massive and longer-period contact binaries., Comment: 23 pages, 11 figures, accepted for publication in MNRAS
- Published
- 2020
- Full Text
- View/download PDF
29. Hindsight Logging for Model Training
- Author
-
Garcia, Rolando, Liu, Eric, Sreekanti, Vikram, Yan, Bobby, Dandamudi, Anusha, Gonzalez, Joseph E., Hellerstein, Joseph M., Sen, Koushik, Garcia, Rolando, Liu, Eric, Sreekanti, Vikram, Yan, Bobby, Dandamudi, Anusha, Gonzalez, Joseph E., Hellerstein, Joseph M., and Sen, Koushik
- Abstract
In modern Machine Learning, model training is an iterative, experimental process that can consume enormous computation resources and developer time. To aid in that process, experienced model developers log and visualize program variables during training runs. Exhaustive logging of all variables is infeasible. Optimistic logging can be accompanied by program checkpoints; this allows developers to add log statements post-hoc, and "replay" desired log statements from checkpoint -- a process we refer to as hindsight logging. Unfortunately, hindsight logging raises tricky problems in data management and software engineering. Done poorly, hindsight logging can waste resources and generate technical debt embodied in multiple variants of training code. In this paper, we present methodologies for efficient and effective logging practices for model training, with a focus on techniques for hindsight logging. Our goal is for experienced model developers to learn and adopt these practices. To make this easier, we provide an open-source suite of tools for Fast Low-Overhead Recovery (flor) that embodies our design across three tasks: (i) efficient background logging in Python, (ii) adaptable periodic checkpointing, and (iii) an instrumentation library that codifies hindsight logging for efficient and automatic record-replay of model-training. Model developers can use each flor tool separately as they see fit, or they can use flor in hands-free mode, entrusting it to instrument their code end-to-end for efficient record-replay. Our solutions leverage techniques from physiological transaction logs and recovery in database systems. Evaluations on modern ML benchmarks demonstrate that flor can produce fast checkpointing with small user-specifiable overheads (e.g. 7%), and still provide hindsight log replay times orders of magnitude faster than restarting training from scratch.
- Published
- 2020
- Full Text
- View/download PDF
30. Ansor: Generating High-Performance Tensor Programs for Deep Learning
- Author
-
Zheng, Lianmin, Jia, Chengfan, Sun, Minmin, Wu, Zhao, Yu, Cody Hao, Haj-Ali, Ameer, Wang, Yida, Yang, Jun, Zhuo, Danyang, Sen, Koushik, Gonzalez, Joseph E., Stoica, Ion, Zheng, Lianmin, Jia, Chengfan, Sun, Minmin, Wu, Zhao, Yu, Cody Hao, Haj-Ali, Ameer, Wang, Yida, Yang, Jun, Zhuo, Danyang, Sen, Koushik, Gonzalez, Joseph E., and Stoica, Ion
- Abstract
High-performance tensor programs are crucial to guarantee efficient execution of deep neural networks. However, obtaining performant tensor programs for different operators on various hardware platforms is notoriously challenging. Currently, deep learning systems rely on vendor-provided kernel libraries or various search strategies to get performant tensor programs. These approaches either require significant engineering effort to develop platform-specific optimization code or fall short of finding high-performance programs due to restricted search space and ineffective exploration strategy. We present Ansor, a tensor program generation framework for deep learning applications. Compared with existing search strategies, Ansor explores many more optimization combinations by sampling programs from a hierarchical representation of the search space. Ansor then fine-tunes the sampled programs with evolutionary search and a learned cost model to identify the best programs. Ansor can find high-performance programs that are outside the search space of existing state-of-the-art approaches. In addition, Ansor utilizes a task scheduler to simultaneously optimize multiple subgraphs in deep neural networks. We show that Ansor improves the execution performance of deep neural networks relative to the state-of-the-art on the Intel CPU, ARM CPU, and NVIDIA GPU by up to $3.8\times$, $2.6\times$, and $1.7\times$, respectively., Comment: OSDI 2020
- Published
- 2020
31. Efficient and Transferable Adversarial Examples from Bayesian Neural Networks
- Author
-
Gubri, Martin, Cordy, Maxime, Papadakis, Mike, Traon, Yves Le, Sen, Koushik, Gubri, Martin, Cordy, Maxime, Papadakis, Mike, Traon, Yves Le, and Sen, Koushik
- Abstract
An established way to improve the transferability of black-box evasion attacks is to craft the adversarial examples on an ensemble-based surrogate to increase diversity. We argue that transferability is fundamentally related to uncertainty. Based on a state-of-the-art Bayesian Deep Learning technique, we propose a new method to efficiently build a surrogate by sampling approximately from the posterior distribution of neural network weights, which represents the belief about the value of each parameter. Our extensive experiments on ImageNet, CIFAR-10 and MNIST show that our approach improves the success rates of four state-of-the-art attacks significantly (up to 83.2 percentage points), in both intra-architecture and inter-architecture transferability. On ImageNet, our approach can reach 94% of success rate while reducing training computations from 11.6 to 2.4 exaflops, compared to an ensemble of independently trained DNNs. Our vanilla surrogate achieves 87.5% of the time higher transferability than three test-time techniques designed for this purpose. Our work demonstrates that the way to train a surrogate has been overlooked, although it is an important element of transfer-based attacks. We are, therefore, the first to review the effectiveness of several training methods in increasing transferability. We provide new directions to better understand the transferability phenomenon and offer a simple but strong baseline for future work., Comment: Accepted at UAI 2022
- Published
- 2020
32. Exempla Gratis (E.G.): Code Examples for Free
- Author
-
Barnaby, Celeste, Sen, Koushik, Zhang, Tianyi, Glassman, Elena, Chandra, Satish, Barnaby, Celeste, Sen, Koushik, Zhang, Tianyi, Glassman, Elena, and Chandra, Satish
- Abstract
Modern software engineering often involves using many existing APIs, both open source and, in industrial coding environments, proprietary. Programmers reference documentation and code search tools to remind themselves of proper common usage patterns of APIs. However, high-quality API usage examples are computationally expensive to curate and maintain, and API usage examples retrieved from company-wide code search can be tedious to review. We present a tool, EG, that mines codebases and shows the common, idiomatic usage examples for API methods. EG was integrated into Facebook's internal code search tool for the Hack language and evaluated on open-source GitHub projects written in Python. EG was also compared against code search results and hand-written examples from a popular programming website called ProgramCreek. Compared with these two baselines, examples generated by EG are more succinct and representative with less extraneous statements. In addition, a survey with Facebook developers shows that EG examples are preferred in 97 percent of cases.
- Published
- 2020
33. Selecting fault revealing mutants
- Author
-
Interdisciplinary Centre for Security, Reliability and Trust (SnT) > Security Design and Validation Research Group (SerVal) [research center], Fonds National de la Recherche - FnR [sponsor], Titcheu Chekam, Thierry, Papadakis, Mike, Bissyande, Tegawendé François D Assise, Le Traon, Yves, Sen, Koushik, Interdisciplinary Centre for Security, Reliability and Trust (SnT) > Security Design and Validation Research Group (SerVal) [research center], Fonds National de la Recherche - FnR [sponsor], Titcheu Chekam, Thierry, Papadakis, Mike, Bissyande, Tegawendé François D Assise, Le Traon, Yves, and Sen, Koushik
- Abstract
Mutant selection refers to the problem of choosing, among a large number of mutants, the (few) ones that should be used by the testers. In view of this, we investigate the problem of selecting the fault revealing mutants, i.e., the mutants that are killable and lead to test cases that uncover unknown program faults. We formulate two variants of this problem: the fault revealing mutant selection and the fault revealing mutant prioritization. We argue and show that these problems can be tackled through a set of ‘static’ program features and propose a machine learning approach, named FaRM, that learns to select and rank killable and fault revealing mutants. Experimental results involving 1,692 real faults show the practical benefits of our approach in both examined problems. Our results show that FaRM achieves a good trade-off between application cost and effectiveness (measured in terms of faults revealed). We also show that FaRM outperforms all the existing mutant selection methods, i.e., the random mutant sampling, the selective mutation and defect prediction (mutating the code areas pointed by defect prediction). In particular, our results show that with respect to mutant selection, our approach reveals 23% to 34% more faults than any of the baseline methods, while, with respect to mutant prioritization, it achieves higher average percentage of revealed faults with a median difference between 4% and 9% (from the random mutant orderings).
- Published
- 2019
34. Semantic Fuzzing with Zest
- Author
-
Padhye, Rohan, Lemieux, Caroline, Sen, Koushik, Papadakis, Mike, Le Traon, Yves, Padhye, Rohan, Lemieux, Caroline, Sen, Koushik, Papadakis, Mike, and Le Traon, Yves
- Published
- 2019
- Full Text
- View/download PDF
35. Selecting fault revealing mutants
- Author
-
Interdisciplinary Centre for Security, Reliability and Trust (SnT) > Security Design and Validation Research Group (SerVal) [research center], Fonds National de la Recherche - FnR [sponsor], Titcheu Chekam, Thierry, Papadakis, Mike, Bissyande, Tegawendé François D Assise, Le Traon, Yves, Sen, Koushik, Interdisciplinary Centre for Security, Reliability and Trust (SnT) > Security Design and Validation Research Group (SerVal) [research center], Fonds National de la Recherche - FnR [sponsor], Titcheu Chekam, Thierry, Papadakis, Mike, Bissyande, Tegawendé François D Assise, Le Traon, Yves, and Sen, Koushik
- Abstract
Mutant selection refers to the problem of choosing, among a large number of mutants, the (few) ones that should be used by the testers. In view of this, we investigate the problem of selecting the fault revealing mutants, i.e., the mutants that are killable and lead to test cases that uncover unknown program faults. We formulate two variants of this problem: the fault revealing mutant selection and the fault revealing mutant prioritization. We argue and show that these problems can be tackled through a set of ‘static’ program features and propose a machine learning approach, named FaRM, that learns to select and rank killable and fault revealing mutants. Experimental results involving 1,692 real faults show the practical benefits of our approach in both examined problems. Our results show that FaRM achieves a good trade-off between application cost and effectiveness (measured in terms of faults revealed). We also show that FaRM outperforms all the existing mutant selection methods, i.e., the random mutant sampling, the selective mutation and defect prediction (mutating the code areas pointed by defect prediction). In particular, our results show that with respect to mutant selection, our approach reveals 23% to 34% more faults than any of the baseline methods, while, with respect to mutant prioritization, it achieves higher average percentage of revealed faults with a median difference between 4% and 9% (from the random mutant orderings).
- Published
- 2019
36. Semantic Fuzzing with Zest
- Author
-
Padhye, Rohan, Lemieux, Caroline, Sen, Koushik, Papadakis, Mike, Le Traon, Yves, Padhye, Rohan, Lemieux, Caroline, Sen, Koushik, Papadakis, Mike, and Le Traon, Yves
- Published
- 2019
- Full Text
- View/download PDF
37. Heuristics for Quantum Compiling with a Continuous Gate Set
- Author
-
Davis, Marc Grau, Smith, Ethan, Tudor, Ana, Sen, Koushik, Siddiqi, Irfan, Iancu, Costin, Davis, Marc Grau, Smith, Ethan, Tudor, Ana, Sen, Koushik, Siddiqi, Irfan, and Iancu, Costin
- Abstract
We present an algorithm for compiling arbitrary unitaries into a sequence of gates native to a quantum processor. As accurate CNOT gates are hard for the foreseeable Noisy- Intermediate-Scale Quantum devices era, our A* inspired algorithm attempts to minimize their count, while accounting for connectivity. We discuss the search strategy together with metrics to expand the solution frontier. For a workload of circuits with complexity appropriate for the NISQ era, we produce solutions well within the best upper bounds published in literature and match or exceed hand tuned implementations, as well as other existing synthesis alternatives. In particular, when comparing against state-of-the-art available synthesis packages we show 2.4x average (up to 5.3x) reduction in CNOT count. We also show how to re-target the algorithm for a different chip topology and native gate set, while obtaining similar quality results. We believe that empirical tools like ours can facilitate algorithmic exploration, gate set discovery for quantum processor designers, as well as providing useful optimization blocks within the quantum compilation tool-chain., Comment: Presented at the 3rd International Workshop on Quantum Compilation as part of the International Conference On Computer Aided Design 2019
- Published
- 2019
38. When Deep Learning Met Code Search
- Author
-
Cambronero, Jose, Li, Hongyu, Kim, Seohyun, Sen, Koushik, Chandra, Satish, Cambronero, Jose, Li, Hongyu, Kim, Seohyun, Sen, Koushik, and Chandra, Satish
- Abstract
There have been multiple recent proposals on using deep neural networks for code search using natural language. Common across these proposals is the idea of $\mathit{embedding}$ code and natural language queries, into real vectors and then using vector distance to approximate semantic correlation between code and the query. Multiple approaches exist for learning these embeddings, including $\mathit{unsupervised}$ techniques, which rely only on a corpus of code examples, and $\mathit{supervised}$ techniques, which use an $\mathit{aligned}$ corpus of paired code and natural language descriptions. The goal of this supervision is to produce embeddings that are more similar for a query and the corresponding desired code snippet. Clearly, there are choices in whether to use supervised techniques at all, and if one does, what sort of network and training to use for supervision. This paper is the first to evaluate these choices systematically. To this end, we assembled implementations of state-of-the-art techniques to run on a common platform, training and evaluation corpora. To explore the design space in network complexity, we also introduced a new design point that is a $\mathit{minimal}$ supervision extension to an existing unsupervised technique. Our evaluation shows that: 1. adding supervision to an existing unsupervised technique can improve performance, though not necessarily by much; 2. simple networks for supervision can be more effective that more sophisticated sequence-based networks for code search; 3. while it is common to use docstrings to carry out supervision, there is a sizeable gap between the effectiveness of docstrings and a more query-appropriate supervision corpus. The evaluation dataset is now available at arXiv:1908.09804
- Published
- 2019
39. Sub-photospheric fluctuations in magnetized radiative envelopes: contribution from unstable magnetosonic waves
- Author
-
Sen, Koushik, Fernández, Rodrigo, Socrates, Aristotle, Sen, Koushik, Fernández, Rodrigo, and Socrates, Aristotle
- Abstract
We examine the excitation of unstable magnetosonic waves in the radiative envelopes of intermediate- and high-mass stars with a magnetic field of ~kG strength. Wind clumping close to the star and microturbulence can often be accounted for when including small-scale, sub-photospheric density or velocity perturbations. Compressional waves - with wavelengths comparable to or shorter than the gas pressure scale height - can be destabilized by the radiative flux in optically-thick media when a magnetic field is present, in a process called the Radiation-Driven Magneto-Acoustic Instability (RMI). The instability does not require radiation or magnetic pressure to dominate over gas pressure, and acts independently of sub-surface convection zones. Here we evaluate the conditions for the RMI to operate on a grid of stellar models covering a mass range $3-40M_\odot$ at solar metallicity. For a uniform 1kG magnetic field, fast magnetosonic modes are unstable down to an optical depth of a few tens, while unstable slow modes extend beyond the depth of the iron convection zone. The qualitative behavior is robust to magnetic field strength variations by a factor of a few. When combining our findings with previous results for the saturation amplitude of the RMI, we predict velocity fluctuations in the range ~0.1-10 km/s. These amplitudes are a monotonically increasing function of the ratio of radiation to gas pressure, or alternatively, of the zero-age main sequence mass., Comment: Accepted by MNRAS
- Published
- 2018
- Full Text
- View/download PDF
40. Sub-photospheric fluctuations in magnetized radiative envelopes: contribution from unstable magnetosonic waves
- Author
-
Sen, Koushik, Fernández, Rodrigo, Socrates, Aristotle, Sen, Koushik, Fernández, Rodrigo, and Socrates, Aristotle
- Abstract
We examine the excitation of unstable magnetosonic waves in the radiative envelopes of intermediate- and high-mass stars with a magnetic field of ~kG strength. Wind clumping close to the star and microturbulence can often be accounted for when including small-scale, sub-photospheric density or velocity perturbations. Compressional waves - with wavelengths comparable to or shorter than the gas pressure scale height - can be destabilized by the radiative flux in optically-thick media when a magnetic field is present, in a process called the Radiation-Driven Magneto-Acoustic Instability (RMI). The instability does not require radiation or magnetic pressure to dominate over gas pressure, and acts independently of sub-surface convection zones. Here we evaluate the conditions for the RMI to operate on a grid of stellar models covering a mass range $3-40M_\odot$ at solar metallicity. For a uniform 1kG magnetic field, fast magnetosonic modes are unstable down to an optical depth of a few tens, while unstable slow modes extend beyond the depth of the iron convection zone. The qualitative behavior is robust to magnetic field strength variations by a factor of a few. When combining our findings with previous results for the saturation amplitude of the RMI, we predict velocity fluctuations in the range ~0.1-10 km/s. These amplitudes are a monotonically increasing function of the ratio of radiation to gas pressure, or alternatively, of the zero-age main sequence mass., Comment: Accepted by MNRAS
- Published
- 2018
- Full Text
- View/download PDF
41. Semantic Fuzzing with Zest
- Author
-
Padhye, Rohan, Lemieux, Caroline, Sen, Koushik, Papadakis, Mike, Traon, Yves Le, Padhye, Rohan, Lemieux, Caroline, Sen, Koushik, Papadakis, Mike, and Traon, Yves Le
- Abstract
Programs expecting structured inputs often consist of both a syntactic analysis stage, which parses raw input, and a semantic analysis stage, which conducts checks on the parsed input and executes the core logic of the program. Generator-based testing tools in the lineage of QuickCheck are a promising way to generate random syntactically valid test inputs for these programs. We present Zest, a technique which automatically guides QuickCheck-like randominput generators to better explore the semantic analysis stage of test programs. Zest converts random-input generators into deterministic parametric generators. We present the key insight that mutations in the untyped parameter domain map to structural mutations in the input domain. Zest leverages program feedback in the form of code coverage and input validity to perform feedback-directed parameter search. We evaluate Zest against AFL and QuickCheck on five Java programs: Maven, Ant, BCEL, Closure, and Rhino. Zest covers 1.03x-2.81x as many branches within the benchmarks semantic analysis stages as baseline techniques. Further, we find 10 new bugs in the semantic analysis stages of these benchmarks. Zest is the most effective technique in finding these bugs reliably and quickly, requiring at most 10 minutes on average to find each bug., Comment: To appear in Proceedings of 28th ACM SIGSOFT International Symposium on Software Testing and Analysis (ISSTA'19)
- Published
- 2018
- Full Text
- View/download PDF
42. Aroma: Code Recommendation via Structural Code Search
- Author
-
Luan, Sifei, Yang, Di, Barnaby, Celeste, Sen, Koushik, Chandra, Satish, Luan, Sifei, Yang, Di, Barnaby, Celeste, Sen, Koushik, and Chandra, Satish
- Abstract
Programmers often write code that has similarity to existing code written somewhere. A tool that could help programmers to search such similar code would be immensely useful. Such a tool could help programmers to extend partially written code snippets to completely implement necessary functionality, help to discover extensions to the partial code which are commonly included by other programmers, help to cross-check against similar code written by other programmers, or help to add extra code which would fix common mistakes and errors. We propose Aroma, a tool and technique for code recommendation via structural code search. Aroma indexes a huge code corpus including thousands of open-source projects, takes a partial code snippet as input, searches the corpus for method bodies containing the partial code snippet, and clusters and intersects the results of the search to recommend a small set of succinct code snippets which both contain the query snippet and appear as part of several methods in the corpus. We evaluated Aroma on 2000 randomly selected queries created from the corpus, as well as 64 queries derived from code snippets obtained from Stack Overflow, a popular website for discussing code. We implemented Aroma for 4 different languages, and developed an IDE plugin for Aroma. Furthermore, we conducted a study where we asked 12 programmers to complete programming tasks using Aroma, and collected their feedback. Our results indicate that Aroma is capable of retrieving and recommending relevant code snippets efficiently.
- Published
- 2018
- Full Text
- View/download PDF
43. Context2Name: A Deep Learning-Based Approach to Infer Natural Variable Names from Usage Contexts
- Author
-
Bavishi, Rohan, Pradel, Michael, Sen, Koushik, Bavishi, Rohan, Pradel, Michael, and Sen, Koushik
- Abstract
Most of the JavaScript code deployed in the wild has been minified, a process in which identifier names are replaced with short, arbitrary and meaningless names. Minified code occupies less space, but also makes the code extremely difficult to manually inspect and understand. This paper presents Context2Name, a deep learningbased technique that partially reverses the effect of minification by predicting natural identifier names for minified names. The core idea is to predict from the usage context of a variable a name that captures the meaning of the variable. The approach combines a lightweight, token-based static analysis with an auto-encoder neural network that summarizes usage contexts and a recurrent neural network that predict natural names for a given usage context. We evaluate Context2Name with a large corpus of real-world JavaScript code and show that it successfully predicts 47.5% of all minified identifiers while taking only 2.9 milliseconds on average to predict a name. A comparison with the state-of-the-art tools JSNice and JSNaughty shows that our approach performs comparably in terms of accuracy while improving in terms of efficiency. Moreover, Context2Name complements the state-of-the-art by predicting 5.3% additional identifiers that are missed by both existing tools.
- Published
- 2018
44. DeepBugs: A Learning Approach to Name-based Bug Detection
- Author
-
Pradel, Michael, Sen, Koushik, Pradel, Michael, and Sen, Koushik
- Abstract
Natural language elements in source code, e.g., the names of variables and functions, convey useful information. However, most existing bug detection tools ignore this information and therefore miss some classes of bugs. The few existing name-based bug detection approaches reason about names on a syntactic level and rely on manually designed and tuned algorithms to detect bugs. This paper presents DeepBugs, a learning approach to name-based bug detection, which reasons about names based on a semantic representation and which automatically learns bug detectors instead of manually writing them. We formulate bug detection as a binary classification problem and train a classifier that distinguishes correct from incorrect code. To address the challenge that effectively learning a bug detector requires examples of both correct and incorrect code, we create likely incorrect code examples from an existing corpus of code through simple code transformations. A novel insight learned from our work is that learning from artificially seeded bugs yields bug detectors that are effective at finding bugs in real-world code. We implement our idea into a framework for learning-based and name-based bug detection. Three bug detectors built on top of the framework detect accidentally swapped function arguments, incorrect binary operators, and incorrect operands in binary operations. Applying the approach to a corpus of 150,000 JavaScript files yields bug detectors that have a high accuracy (between 89% and 95%), are very efficient (less than 20 milliseconds per analyzed file), and reveal 102 programming mistakes (with 68% true positive rate) in real-world code.
- Published
- 2018
45. Selecting Fault Revealing Mutants
- Author
-
Chekam, Thierry Titcheu, Papadakis, Mike, Bissyandé, Tegawendé, Traon, Yves Le, Sen, Koushik, Chekam, Thierry Titcheu, Papadakis, Mike, Bissyandé, Tegawendé, Traon, Yves Le, and Sen, Koushik
- Abstract
Mutant selection refers to the problem of choosing, among a large number of mutants, the (few) ones that should be used by the testers. In view of this, we investigate the problem of selecting the fault revealing mutants, i.e., the mutants that are most likely to be killable and lead to test cases that uncover unknown program faults. We formulate two variants of this problem: the fault revealing mutant selection and the fault revealing mutant prioritization. We argue and show that these problems can be tackled through a set of 'static' program features and propose a machine learning approach, named FaRM, that learns to select and rank killable and fault revealing mutants. Experimental results involving 1,692 real faults show the practical benefits of our approach in both examined problems. Our results show that FaRM achieves a good trade-off between application cost and effectiveness (measured in terms of faults revealed). We also show that FaRM outperforms all the existing mutant selection methods, i.e., the random mutant sampling, the selective mutation and defect prediction (mutating the code areas pointed by defect prediction). In particular, our results show that with respect to mutant selection, our approach reveals 23% to 34% more faults than any of the baseline methods, while, with respect to mutant prioritization, it achieves higher average percentage of revealed faults with a median difference between 4% and 9% (from the random mutant orderings).
- Published
- 2018
46. Report of the HPC Correctness Summit, Jan 25--26, 2017, Washington, DC
- Author
-
Gopalakrishnan, Ganesh, Hovland, Paul D., Iancu, Costin, Krishnamoorthy, Sriram, Laguna, Ignacio, Lethin, Richard A., Sen, Koushik, Siegel, Stephen F., Solar-Lezama, Armando, Gopalakrishnan, Ganesh, Hovland, Paul D., Iancu, Costin, Krishnamoorthy, Sriram, Laguna, Ignacio, Lethin, Richard A., Sen, Koushik, Siegel, Stephen F., and Solar-Lezama, Armando
- Abstract
Maintaining leadership in HPC requires the ability to support simulations at large scales and fidelity. In this study, we detail one of the most significant productivity challenges in achieving this goal, namely the increasing proclivity to bugs, especially in the face of growing hardware and software heterogeneity and sheer system scale. We identify key areas where timely new research must be proactively begun to address these challenges, and create new correctness tools that must ideally play a significant role even while ramping up toward exacale. We close with the proposal for a two-day workshop in which the problems identified in this report can be more broadly discussed, and specific plans to launch these new research thrusts identified., Comment: 57 pages
- Published
- 2017
47. FairFuzz: Targeting Rare Branches to Rapidly Increase Greybox Fuzz Testing Coverage
- Author
-
Lemieux, Caroline, Sen, Koushik, Lemieux, Caroline, and Sen, Koushik
- Abstract
In recent years, fuzz testing has proven itself to be one of the most effective techniques for finding correctness bugs and security vulnerabilities in practice. One particular fuzz testing tool, American Fuzzy Lop or AFL, has become popular thanks to its ease-of-use and bug-finding power. However, AFL remains limited in the depth of program coverage it achieves, in particular because it does not consider which parts of program inputs should not be mutated in order to maintain deep program coverage. We propose an approach, FairFuzz, that helps alleviate this limitation in two key steps. First, FairFuzz automatically prioritizes inputs exercising rare parts of the program under test. Second, it automatically adjusts the mutation of inputs so that the mutated inputs are more likely to exercise these same rare parts of the program. We conduct evaluation on real-world programs against state-of-the-art versions of AFL, thoroughly repeating experiments to get good measures of variability. We find that on certain benchmarks FairFuzz shows significant coverage increases after 24 hours compared to state-of-the-art versions of AFL, while on others it achieves high program coverage at a significantly faster rate.
- Published
- 2017
- Full Text
- View/download PDF
48. OPR
- Author
-
Qian, Xuehai, Qian, Xuehai, Sen, Koushik, Hargrove, Paul, Iancu, Costin, Qian, Xuehai, Qian, Xuehai, Sen, Koushik, Hargrove, Paul, and Iancu, Costin
- Abstract
The ability to reproduce a parallel execution is desirable for debugging and program reliability purposes. In debugging (13), the programmer needs to manually step back in time, while for resilience (6) this is automatically performed by the the application upon failure. To be useful, replay has to faithfully reproduce the original execution. For parallel programs the main challenge is inferring and maintaining the order of conflicting operations (data races). Deterministic record and replay (R&R) techniques have been developed for multithreaded shared memory programs (5), as well as distributed memory programs (14). Our main interest is techniques for large scale scientific (3; 4) programming models.
- Published
- 2016
49. Trace Typing: An Approach for Evaluating Retrofitted Type Systems
- Author
-
Andreasen, Esben, Gordon, Colin S., Chandra, Satish, Sridharan, Manu, Tip, Frank, Sen, Koushik, Andreasen, Esben, Gordon, Colin S., Chandra, Satish, Sridharan, Manu, Tip, Frank, and Sen, Koushik
- Abstract
Recent years have seen growing interest in the retrofitting of type systems onto dynamically-typed programming languages, in order to improve type safety, programmer productivity, or performance. In such cases, type system developers must strike a delicate balance between disallowing certain coding patterns to keep the type system simple, or including them at the expense of additional complexity and effort. Thus far, the process for designing retrofitted type systems has been largely ad hoc, because evaluating multiple variations of a type system on large bodies of existing code is a significant undertaking. We present trace typing: a framework for automatically and quantitatively evaluating variations of a retrofitted type system on large code bases. The trace typing approach involves gathering traces of program executions, inferring types for instances of variables and expressions occurring in a trace, and merging types according to merge strategies that reflect specific (combinations of) choices in the source-level type system design space. We evaluated trace typing through several experiments. We compared several variations of type systems retrofitted onto JavaScript, measuring the number of program locations with type errors in each case on a suite of over fifty thousand lines of JavaScript code. We also used trace typing to validate and guide the design of a new retrofitted type system that enforces fixed object layout for JavaScript objects. Finally, we leveraged the types computed by trace typing to automatically identify tag tests --- dynamic checks that refine a type --- and examined the variety of tests identified.
- Published
- 2016
- Full Text
- View/download PDF
50. OPR
- Author
-
Qian, Xuehai, Qian, Xuehai, Sen, Koushik, Hargrove, Paul, Iancu, Costin, Qian, Xuehai, Qian, Xuehai, Sen, Koushik, Hargrove, Paul, and Iancu, Costin
- Abstract
The ability to reproduce a parallel execution is desirable for debugging and program reliability purposes. In debugging (13), the programmer needs to manually step back in time, while for resilience (6) this is automatically performed by the the application upon failure. To be useful, replay has to faithfully reproduce the original execution. For parallel programs the main challenge is inferring and maintaining the order of conflicting operations (data races). Deterministic record and replay (R&R) techniques have been developed for multithreaded shared memory programs (5), as well as distributed memory programs (14). Our main interest is techniques for large scale scientific (3; 4) programming models.
- Published
- 2016
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.