33 results on '"Alvarsson J"'
Search Results
2. Pain rather than induced emotions and ICU sound increases skin conductance variability in healthy volunteers
- Author
-
Gunther, A. C., Schandl, A. R., Bernhardsson, Jens, Bjärtå, Anna, Wållgren, M., Sundin, Örjan, Alvarsson, J., Bottai, M., Martling, C. -R, Sackey, P. V., Gunther, A. C., Schandl, A. R., Bernhardsson, Jens, Bjärtå, Anna, Wållgren, M., Sundin, Örjan, Alvarsson, J., Bottai, M., Martling, C. -R, and Sackey, P. V.
- Abstract
Background: Assessing pain in critically ill patients is difficult. Skin conductance variability (SCV), induced by the sympathetic response to pain, has been suggested as a method to identify pain in poorly communicating patients. However, SCV, a derivate of conventional skin conductance, could potentially also be sensitive to emotional stress. The purpose of the study was to investigate if pain and emotional stress can be distinguished with SCV. Methods: In a series of twelve 1-min sessions with SCV recording, 18 healthy volunteers were exposed to standardized electric pain stimulation during blocks of positive, negative, or neutral emotion, induced with pictures from the International Affective PictureSystem (IAPS). Additionally, authentic intensive care unit (ICU) sound was included in half of the sessions. All possible combinations of pain and sound occurred in each block of emotion, and blocks were presented in randomized order. Results: Pain stimulation resulted in increases in the number of skin conductance fluctuations (NSCF) in all but one participant. During pain-free baseline sessions, the median NSCF was 0.068 (interquartile range 0.013-0.089) and during pain stimulation median NSCF increased to 0.225 (interquartile range 0.146-0.3175). Only small increases in NSCF were found during negative emotions. Pain, assessed with the numeric rating scale, during the sessions with pain stimulation was not altered significantly by other ongoing sensory input. Conclusion: In healthy volunteers, NSCF appears to reflect ongoing autonomous reactions mainly to pain and to a lesser extent, reactions to emotion induced with IAPS pictures or ICU sound.
- Published
- 2016
- Full Text
- View/download PDF
3. Pain rather than induced emotions and ICU sound increases skin conductance variability in healthy volunteers
- Author
-
Günther, A. C., primary, Schandl, A. R., additional, Berhardsson, J., additional, Bjärtå, A., additional, Wållgren, M., additional, Sundin, Ö., additional, Alvarsson, J., additional, Bottai, M., additional, Martling, C.-R., additional, and Sackey, P. V., additional
- Published
- 2016
- Full Text
- View/download PDF
4. Bioclipse-R: integrating management and visualization of life science data with statistical analysis
- Author
-
Spjuth, O., Spjuth, O., Georgiev, V., Carlsson, L., Alvarsson, J., Berg, A., Willighagen, E., Wikberg, J. E., Eklund, M., Spjuth, O., Spjuth, O., Georgiev, V., Carlsson, L., Alvarsson, J., Berg, A., Willighagen, E., Wikberg, J. E., and Eklund, M.
- Abstract
SUMMARY: Bioclipse, a graphical workbench for the life sciences, provides functionality for managing and visualizing life science data. We introduce Bioclipse-R, which integrates Bioclipse and the statistical programming language R. The synergy between Bioclipse and R is demonstrated by the construction of a decision support system for anticancer drug screening and mutagenicity prediction, which shows how Bioclipse-R can be used to perform complex tasks from within a single software system. Availability and implementation: Bioclipse-R is implemented as a set of Java plug-ins for Bioclipse based on the R-package rj. Source code and binary packages are available from https://github.com/bioclipse and http://www.bioclipse.net/bioclipse-r, respectively. CONTACT: martin.eklund@farmbio.uu.se SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online.
- Published
- 2013
5. Loudness of fountain and road traffic sounds in a city park
- Author
-
Nilsson, M. E., Alvarsson, J., Rådsten-Ekman, M., Bolin, Karl, Nilsson, M. E., Alvarsson, J., Rådsten-Ekman, M., and Bolin, Karl
- Abstract
Auditory masking of unwanted sounds by wanted sounds has been suggested as an approach to soundscape improvement. Anecdotal evidence exists on successful applications, for instance use of fountain sounds for masking road traffic noise in urban parks. However, basic research on auditory masking of environmental sounds is lacking. Therefore, we conducted two listening experiments on auditory masking, using binaural recordings from a city park in Stockholm exposed to traffic noise from a main road and sound from a large fountain located in the centre of the park. In Experiment 1, 12 listeners assessed the loudness of road traffic noise and fountain sounds from recordings at various distances from road, with or without the fountain turned on. In Experiment 2, the same listeners assessed loudness of manipulated sound levels of singular or combined road traffic or fountain sounds. The results of Experiment 1 showed that the fountain sound reduced the loudness of road traffic noise close to the fountain, and that the fountain sound was equally loud or louder than the road traffic noise in a region 20-30 m around the fountain. This suggests that fountain sounds may add to the quality of city park soundscape by reducing the loudness of the (presumably unwanted) traffic noise. On the other hand, results from both experiments showed that road traffic noise was harder to mask than fountain sound. Furthermore, Experiment 2 showed that partial loudness of both sources was considerably less than expected from a model of energetic masking. This suggests that informational masking due to target-masker similarity may reduce the overall masking effect of environmental sounds., QC 20140923
- Published
- 2009
6. The Chemistry Development Kit (CDK) v2.0: atom typing, depiction, molecular formulas, and substructure searching
- Author
-
Egon Willighagen, Jw, Mayfield, Alvarsson J, Berg A, Carlsson L, Jeliazkova N, Kuhn S, Pluskal T, Rojas-Chertó M, Spjuth O, Torrance G, Ct, Evelo, Guha R, and Steinbeck C
7. Linking the Resource Description Framework to cheminformatics and proteochemometrics
- Author
-
Willighagen Egon L, Alvarsson Jonathan, Andersson Annsofie, Eklund Martin, Lampa Samuel, Lapins Maris, Spjuth Ola, and Wikberg Jarl ES
- Subjects
Computer applications to medicine. Medical informatics ,R858-859.7 - Abstract
Abstract Background Semantic web technologies are finding their way into the life sciences. Ontologies and semantic markup have already been used for more than a decade in molecular sciences, but have not found widespread use yet. The semantic web technology Resource Description Framework (RDF) and related methods show to be sufficiently versatile to change that situation. Results The work presented here focuses on linking RDF approaches to existing molecular chemometrics fields, including cheminformatics, QSAR modeling and proteochemometrics. Applications are presented that link RDF technologies to methods from statistics and cheminformatics, including data aggregation, visualization, chemical identification, and property prediction. They demonstrate how this can be done using various existing RDF standards and cheminformatics libraries. For example, we show how IC50 and Ki values are modeled for a number of biological targets using data from the ChEMBL database. Conclusions We have shown that existing RDF standards can suitably be integrated into existing molecular chemometrics methods. Platforms that unite these technologies, like Bioclipse, makes this even simpler and more transparent. Being able to create and share workflows that integrate data aggregation and analysis (visual and statistical) is beneficial to interoperability and reproducibility. The current work shows that RDF approaches are sufficiently powerful to support molecular chemometrics workflows.
- Published
- 2011
- Full Text
- View/download PDF
8. Open Data, Open Source and Open Standards in chemistry: The Blue Obelisk five years on
- Author
-
O'Boyle Noel M, Guha Rajarshi, Willighagen Egon L, Adams Samuel E, Alvarsson Jonathan, Bradley Jean-Claude, Filippov Igor V, Hanson Robert M, Hanwell Marcus D, Hutchison Geoffrey R, James Craig A, Jeliazkova Nina, Lang Andrew SID, Langner Karol M, Lonie David C, Lowe Daniel M, Pansanel Jérôme, Pavlov Dmitry, Spjuth Ola, Steinbeck Christoph, Tenderholt Adam L, Theisen Kevin J, and Murray-Rust Peter
- Subjects
Information technology ,T58.5-58.64 ,Chemistry ,QD1-999 - Abstract
Abstract Background The Blue Obelisk movement was established in 2005 as a response to the lack of Open Data, Open Standards and Open Source (ODOSOS) in chemistry. It aims to make it easier to carry out chemistry research by promoting interoperability between chemistry software, encouraging cooperation between Open Source developers, and developing community resources and Open Standards. Results This contribution looks back on the work carried out by the Blue Obelisk in the past 5 years and surveys progress and remaining challenges in the areas of Open Data, Open Standards, and Open Source in chemistry. Conclusions We show that the Blue Obelisk has been very successful in bringing together researchers and developers with common interests in ODOSOS, leading to development of many useful resources freely available to the chemistry community.
- Published
- 2011
- Full Text
- View/download PDF
9. Brunn: An open source laboratory information system for microplates with a graphical plate layout design process
- Author
-
Larsson Rolf, Spjuth Ola, Andersson Claes, Alvarsson Jonathan, and Wikberg Jarl ES
- Subjects
Computer applications to medicine. Medical informatics ,R858-859.7 ,Biology (General) ,QH301-705.5 - Abstract
Abstract Background Compound profiling and drug screening generates large amounts of data and is generally based on microplate assays. Current information systems used for handling this are mainly commercial, closed source, expensive, and heavyweight and there is a need for a flexible lightweight open system for handling plate design, and validation and preparation of data. Results A Bioclipse plugin consisting of a client part and a relational database was constructed. A multiple-step plate layout point-and-click interface was implemented inside Bioclipse. The system contains a data validation step, where outliers can be removed, and finally a plate report with all relevant calculated data, including dose-response curves. Conclusions Brunn is capable of handling the data from microplate assays. It can create dose-response curves and calculate IC50 values. Using a system of this sort facilitates work in the laboratory. Being able to reuse already constructed plates and plate layouts by starting out from an earlier step in the plate layout design process saves time and cuts down on error sources.
- Published
- 2011
- Full Text
- View/download PDF
10. Bioclipse 2: A scriptable integration platform for the life sciences
- Author
-
Wagener Johannes, Torrance Gilleain, Mäsak Carl, Kuhn Stefan, Eklund Martin, Berg Arvid, Alvarsson Jonathan, Spjuth Ola, Willighagen Egon L, Steinbeck Christoph, and Wikberg Jarl ES
- Subjects
Computer applications to medicine. Medical informatics ,R858-859.7 ,Biology (General) ,QH301-705.5 - Abstract
Abstract Background Contemporary biological research integrates neighboring scientific domains to answer complex questions in fields such as systems biology and drug discovery. This calls for tools that are intuitive to use, yet flexible to adapt to new tasks. Results Bioclipse is a free, open source workbench with advanced features for the life sciences. Version 2.0 constitutes a complete rewrite of Bioclipse, and delivers a stable, scalable integration platform for developers and an intuitive workbench for end users. All functionality is available both from the graphical user interface and from a built-in novel domain-specific language, supporting the scientist in interdisciplinary research and reproducible analyses through advanced visualization of the inputs and the results. New components for Bioclipse 2 include a rewritten editor for chemical structures, a table for multiple molecules that supports gigabyte-sized files, as well as a graphical editor for sequences and alignments. Conclusion Bioclipse 2 is equipped with advanced tools required to carry out complex analysis in the fields of bio- and cheminformatics. Developed as a Rich Client based on Eclipse, Bioclipse 2 leverages on today's powerful desktop computers for providing a responsive user interface, but also takes full advantage of the Web and networked (Web/Cloud) services for more demanding calculations or retrieval of data. The fact that Bioclipse 2 is based on an advanced and widely used service platform ensures wide extensibility, making it easy to add new algorithms, visualizations, as well as scripting commands. The intuitive tools for end users and the extensible architecture make Bioclipse 2 ideal for interdisciplinary and integrative research. Bioclipse 2 is released under the Eclipse Public License (EPL), a flexible open source license that allows additional plugins to be of any license. Bioclipse 2 is implemented in Java and supported on all major platforms; Source code and binaries are freely available at http://www.bioclipse.net.
- Published
- 2009
- Full Text
- View/download PDF
11. CPSign: conformal prediction for cheminformatics modeling.
- Author
-
Arvidsson McShane S, Norinder U, Alvarsson J, Ahlberg E, Carlsson L, and Spjuth O
- Abstract
Conformal prediction has seen many applications in pharmaceutical science, being able to calibrate outputs of machine learning models and producing valid prediction intervals. We here present the open source software CPSign that is a complete implementation of conformal prediction for cheminformatics modeling. CPSign implements inductive and transductive conformal prediction for classification and regression, and probabilistic prediction with the Venn-ABERS methodology. The main chemical representation is signatures but other types of descriptors are also supported. The main modeling methodology is support vector machines (SVMs), but additional modeling methods are supported via an extension mechanism, e.g. DeepLearning4J models. We also describe features for visualizing results from conformal models including calibration and efficiency plots, as well as features to publish predictive models as REST services. We compare CPSign against other common cheminformatics modeling approaches including random forest, and a directed message-passing neural network. The results show that CPSign produces robust predictive performance with comparative predictive efficiency, with superior runtime and lower hardware requirements compared to neural network based models. CPSign has been used in several studies and is in production-use in multiple organizations. The ability to work directly with chemical input files, perform descriptor calculation and modeling with SVM in the conformal prediction framework, with a single software package having a low footprint and fast execution time makes CPSign a convenient and yet flexible package for training, deploying, and predicting on chemical data. CPSign can be downloaded from GitHub at https://github.com/arosbio/cpsign .Scientific contribution CPSign provides a single software that allows users to perform data preprocessing, modeling and make predictions directly on chemical structures, using conformal and probabilistic prediction. Building and evaluating new models can be achieved at a high abstraction level, without sacrificing flexibility and predictive performance-showcased with a method evaluation against contemporary modeling approaches, where CPSign performs on par with a state-of-the-art deep learning based model., (© 2024. The Author(s).)
- Published
- 2024
- Full Text
- View/download PDF
12. Longitudinal Effects of Screen Time on Depressive Symptoms among Swedish Adolescents: The Moderating and Mediating Role of Coping Engagement Behavior.
- Author
-
Hökby S, Westerlund J, Alvarsson J, Carli V, and Hadlaczky G
- Subjects
- Male, Humans, Adolescent, Female, Sweden, Adaptation, Psychological, Emotions, Depression psychology, Screen Time
- Abstract
Studies suggest that hourly digital screen time increases adolescents' depressive symptoms and emotional regulation difficulties. However, causal mechanisms behind such associations remain unclear. We hypothesized that problem-focused and/or emotion-focused engagement coping moderates and possibly mediates this association over time. Questionnaire data were collected in three waves from a representative sample of Swedish adolescents (0, 3 and 12 months; n = 4793; 51% boys; 99% aged 13-15). Generalized Estimating Equations estimated the main effects and moderation effects, and structural regression estimated the mediation pathways. The results showed that problem-focused coping had a main effect on future depression ( b = 0.030; p < 0.001) and moderated the effect of screen time ( b = 0.009; p < 0.01). The effect size of this moderation was maximum 3.4 BDI-II scores. The mediation results corroborated the finding that future depression was only indirectly correlated with baseline screen time, conditional upon intermittent problem-coping interference (C'-path: Std. beta = 0.001; p = 0.018). The data did not support direct effects, emotion-focused coping effects, or reversed causality. We conclude that hourly screen time can increase depressive symptoms in adolescent populations through interferences with problem-focused coping and other emotional regulation behaviors. Preventive programs could target coping interferences to improve public health. We discuss psychological models of why screen time may interfere with coping, including displacement effects and echo chamber phenomena.
- Published
- 2023
- Full Text
- View/download PDF
13. In Silico Prediction of Human Clinical Pharmacokinetics with ANDROMEDA by Prosilico: Predictions for an Established Benchmarking Data Set, a Modern Small Drug Data Set, and a Comparison with Laboratory Methods.
- Author
-
Fagerholm U, Hellberg S, Alvarsson J, and Spjuth O
- Subjects
- Animals, Humans, Permeability, Pharmacokinetics, Pharmaceutical Preparations, Computer Simulation, Benchmarking, Models, Biological
- Abstract
There is an ongoing aim to replace animal and in vitro laboratory models with in silico methods. Such replacement requires the successful validation and comparably good performance of the alternative methods. We have developed an in silico prediction system for human clinical pharmacokinetics, based on machine learning, conformal prediction and a new physiologically-based pharmacokinetic model, i.e. ANDROMEDA. The objectives of this study were: a) to evaluate how well ANDROMEDA predicts the human clinical pharmacokinetics of a previously proposed benchmarking data set comprising 24 physicochemically diverse drugs and 28 small drug molecules new to the market in 2021; b) to compare its predictive performance with that of laboratory methods; and c) to investigate and describe the pharmacokinetic characteristics of the modern drugs. Median and maximum prediction errors for the selected major parameters were ca 1.2 to 2.5-fold and 16-fold for both data sets, respectively. Prediction accuracy was on par with, or better than, the best laboratory-based prediction methods (superior performance for a vast majority of the comparisons), and the prediction range was considerably broader. The modern drugs have higher average molecular weight than those in the benchmarking set from 15 years earlier ( ca 200 g/mol higher), and were predicted to (generally) have relatively complex pharmacokinetics, including permeability and dissolution limitations and significant renal, biliary and/or gut-wall elimination. In conclusion, the results were overall better than those obtained with laboratory methods, and thus serve to further validate the ANDROMEDA in silico system for the prediction of human clinical pharmacokinetics of modern and physicochemically diverse drugs.
- Published
- 2023
- Full Text
- View/download PDF
14. In Silico Predictions of the Gastrointestinal Uptake of Macrocycles in Man Using Conformal Prediction Methodology.
- Author
-
Fagerholm U, Hellberg S, Alvarsson J, and Spjuth O
- Subjects
- Administration, Oral, Caco-2 Cells, Computer Simulation, Humans, Permeability, Pharmaceutical Preparations, Solubility, Intestinal Absorption, Models, Biological
- Abstract
The gastrointestinal uptake of macrocyclic compounds is not fully understood. Here we applied our previously validated integrated system based on machine learning and conformal prediction to predict the passive fraction absorbed (f
a ), maximum fraction dissolved (fdiss ), substrate specificities for major efflux transporters and total fraction absorbed (fa,tot ) for a selected set of designed macrocyclic compounds (n = 37; MW 407-889 g/mol) and macrocyclic drugs (n = 16; MW 734-1203 g/mole) in vivo in man. Major aims were to increase the understanding of oral absorption of macrocycles and further validate our methodology. We predicted designed macrocycles to have high fa and low to high fdiss and fa,tot , and average estimates were higher than for the larger macrocyclic drugs. With few exceptions, compounds were predicted to be effluxed and well absorbed. A 2-fold median prediction error for fa,tot was achieved for macrocycles (validation set). Advantages with our methodology include that it enables predictions for macrocycles with low permeability, Caco-2 recovery and solubility (BCS IV), and provides prediction intervals and guides optimization of absorption. The understanding of oral absorption of macrocycles was increased and the methodology was validated for prediction of the uptake of macrocycles in man., Competing Interests: Declaration of Competing Interest The authors declare the following financial interests/personal relationships which may be considered as potential competing interests: Urban Fagerholm, Sven Hellberg and Ola Spjuth declare shares in Prosilico AB, a Swedish company that develops solutions for human clinical ADME/PK predictions. Ola Spjuth declares shares in Aros Bio AB, a company developing the CPSign software., (Copyright © 2022 American Pharmacists Association. Published by Elsevier Inc. All rights reserved.)- Published
- 2022
- Full Text
- View/download PDF
15. In silico predictions of the human pharmacokinetics/toxicokinetics of 65 chemicals from various classes using conformal prediction methodology.
- Author
-
Fagerholm U, Hellberg S, Alvarsson J, and Spjuth O
- Subjects
- Biological Availability, Computer Simulation, Humans, Kinetics, Pharmaceutical Preparations, Toxicokinetics, Models, Biological, Pharmacokinetics
- Abstract
Pharmacokinetic/toxicokinetic (PK/TK) information for chemicals in humans is generally lacking. Here we applied machine learning, conformal prediction and a new physiologically-based PK/TK model for prediction of the human PK/TK of 65 chemicals from different classes, including carcinogens, food constituents and preservatives, vitamins, sweeteners, dyes and colours, pesticides, alternative medicines, flame retardants, psychoactive drugs, dioxins, poisons, UV-absorbents, surfactants, solvents and cosmetics.About 80% of the main human PK/TK (fraction absorbed, oral bioavailability, half-life, unbound fraction in plasma, clearance, volume of distribution, fraction excreted) for the selected chemicals was missing in the literature. This information was now added (from in silico predictions). Median and mean prediction errors for these parameters were 1.3- to 2.7-fold and 1.4- to 4.8-fold, respectively. In total, 59 and 86% of predictions had errors <2- and <5-fold, respectively. Predicted and observed PK/TK for the chemicals was generally within the range for pharmaceutical drugs.The results validated the new integrated system for prediction of the human PK/TK for different chemicals and added important missing information. No general difference in PK/TK-characteristics was found between the selected chemicals and pharmaceutical drugs.
- Published
- 2022
- Full Text
- View/download PDF
16. In silico prediction of volume of distribution of drugs in man using conformal prediction performs on par with animal data-based models.
- Author
-
Fagerholm U, Hellberg S, Alvarsson J, Arvidsson McShane S, and Spjuth O
- Subjects
- Animals, Drug Discovery, Models, Animal, Pharmacokinetics, Rats, Models, Biological, Pharmaceutical Preparations
- Abstract
Volume of distribution at steady state (V
ss ) is an important pharmacokinetic endpoint. In this study we apply machine learning and conformal prediction for human Vss prediction, and make a head-to-head comparison with rat-to-man scaling, allometric scaling and the Rodgers-Lukova method on combined in silico and in vitro data, using a test set of 105 compounds with experimentally observed Vss .The mean prediction error and % with <2-fold prediction error for our method were 2.4-fold and 64%, respectively. 69% of test compounds had an observed Vss within the prediction interval at a 70% confidence level. In comparison, 2.2-, 2.9- and 3.1-fold mean errors and 69, 64 and 61% of predictions with <2-fold error was reached with rat-to-man and allometric scaling and Rodgers-Lukova method, respectively.We conclude that our method has theoretically proven validity that was empirically confirmed, and showing predictive accuracy on par with animal models and superior to an alternative widely used in silico -based method. The option for the user to select the level of confidence in predictions offers better guidance on how to optimise Vss in drug discovery applications.- Published
- 2021
- Full Text
- View/download PDF
17. Predicting With Confidence: Using Conformal Prediction in Drug Discovery.
- Author
-
Alvarsson J, Arvidsson McShane S, Norinder U, and Spjuth O
- Subjects
- Algorithms, Molecular Conformation, Quantitative Structure-Activity Relationship, Reproducibility of Results, Drug Discovery, Machine Learning
- Abstract
One of the challenges with predictive modeling is how to quantify the reliability of the models' predictions on new objects. In this work we give an introduction to conformal prediction, a framework that sits on top of traditional machine learning algorithms and which outputs valid confidence estimates to predictions from QSAR models in the form of prediction intervals that are specific to each predicted object. For regression, a prediction interval consists of an upper and a lower bound. For classification, a prediction interval is a set that contains none, one, or many of the potential classes. The size of the prediction interval is affected by a user-specified confidence/significance level, and by the nonconformity of the predicted object; i.e., the strangeness as defined by a nonconformity function. Conformal prediction provides a rigorous and mathematically proven framework for in silico modeling with guarantees on error rates as well as a consistent handling of the models' applicability domain intrinsically linked to the underlying machine learning model. Apart from introducing the concepts and types of conformal prediction, we also provide an example application for modeling ABC transporters using conformal prediction, as well as a discussion on general implications for drug discovery., (Copyright © 2020 The Authors. Published by Elsevier Inc. All rights reserved.)
- Published
- 2021
- Full Text
- View/download PDF
18. SciPipe: A workflow library for agile development of complex and dynamic bioinformatics pipelines.
- Author
-
Lampa S, Dahlö M, Alvarsson J, and Spjuth O
- Subjects
- Gene Library, Machine Learning, Programming Languages, Workflow, Computational Biology, Genomics, Software
- Abstract
Background: The complex nature of biological data has driven the development of specialized software tools. Scientific workflow management systems simplify the assembly of such tools into pipelines, assist with job automation, and aid reproducibility of analyses. Many contemporary workflow tools are specialized or not designed for highly complex workflows, such as with nested loops, dynamic scheduling, and parametrization, which is common in, e.g., machine learning., Findings: SciPipe is a workflow programming library implemented in the programming language Go, for managing complex and dynamic pipelines in bioinformatics, cheminformatics, and other fields. SciPipe helps in particular with workflow constructs common in machine learning, such as extensive branching, parameter sweeps, and dynamic scheduling and parametrization of downstream tasks. SciPipe builds on flow-based programming principles to support agile development of workflows based on a library of self-contained, reusable components. It supports running subsets of workflows for improved iterative development and provides a data-centric audit logging feature that saves a full audit trace for every output file of a workflow, which can be converted to other formats such as HTML, TeX, and PDF on demand. The utility of SciPipe is demonstrated with a machine learning pipeline, a genomics, and a transcriptomics pipeline., Conclusions: SciPipe provides a solution for agile development of complex and dynamic pipelines, especially in machine learning, through a flexible application programming interface suitable for scientists used to programming or scripting., (© The Author(s) 2019. Published by Oxford University Press.)
- Published
- 2019
- Full Text
- View/download PDF
19. Predicting Off-Target Binding Profiles With Confidence Using Conformal Prediction.
- Author
-
Lampa S, Alvarsson J, Arvidsson Mc Shane S, Berg A, Ahlberg E, and Spjuth O
- Abstract
Ligand-based models can be used in drug discovery to obtain an early indication of potential off-target interactions that could be linked to adverse effects. Another application is to combine such models into a panel, allowing to compare and search for compounds with similar profiles. Most contemporary methods and implementations however lack valid measures of confidence in their predictions, and only provide point predictions. We here describe a methodology that uses Conformal Prediction for predicting off-target interactions, with models trained on data from 31 targets in the ExCAPE-DB dataset selected for their utility in broad early hazard assessment. Chemicals were represented by the signature molecular descriptor and support vector machines were used as the underlying machine learning method. By using conformal prediction, the results from predictions come in the form of confidence p -values for each class. The full pre-processing and model training process is openly available as scientific workflows on GitHub, rendering it fully reproducible. We illustrate the usefulness of the developed methodology on a set of compounds extracted from DrugBank. The resulting models are published online and are available via a graphical web interface and an OpenAPI interface for programmatic access.
- Published
- 2018
- Full Text
- View/download PDF
20. Evaluating parameters for ligand-based modeling with random forest on sparse data sets.
- Author
-
Kensert A, Alvarsson J, Norinder U, and Spjuth O
- Abstract
Ligand-based predictive modeling is widely used to generate predictive models aiding decision making in e.g. drug discovery projects. With growing data sets and requirements on low modeling time comes the necessity to analyze data sets efficiently to support rapid and robust modeling. In this study we analyzed four data sets and studied the efficiency of machine learning methods on sparse data structures, utilizing Morgan fingerprints of different radii and hash sizes, and compared with molecular signatures descriptor of different height. We specifically evaluated the effect these parameters had on modeling time, predictive performance, and memory requirements using two implementations of random forest; Scikit-learn as well as FEST. We also compared with a support vector machine implementation. Our results showed that unhashed fingerprints yield significantly better accuracy than hashed fingerprints ([Formula: see text]), with no pronounced deterioration in modeling time and memory usage. Furthermore, the fast execution and low memory usage of the FEST algorithm suggest that it is a good alternative for large, high dimensional sparse data. Both support vector machines and random forest performed equally well but results indicate that the support vector machine was better at using the extra information from larger values of the Morgan fingerprint's radius.
- Published
- 2018
- Full Text
- View/download PDF
21. A confidence predictor for logD using conformal regression and a support-vector machine.
- Author
-
Lapins M, Arvidsson S, Lampa S, Berg A, Schaal W, Alvarsson J, and Spjuth O
- Abstract
Lipophilicity is a major determinant of ADMET properties and overall suitability of drug candidates. We have developed large-scale models to predict water-octanol distribution coefficient (logD) for chemical compounds, aiding drug discovery projects. Using ACD/logD data for 1.6 million compounds from the ChEMBL database, models are created and evaluated by a support-vector machine with a linear kernel using conformal prediction methodology, outputting prediction intervals at a specified confidence level. The resulting model shows a predictive ability of [Formula: see text] and with the best performing nonconformity measure having median prediction interval of [Formula: see text] log units at 80% confidence and [Formula: see text] log units at 90% confidence. The model is available as an online service via an OpenAPI interface, a web page with a molecular editor, and we also publish predictive values at 90% confidence level for 91 M PubChem structures in RDF format for download and as an URI resolver service.
- Published
- 2018
- Full Text
- View/download PDF
22. Erratum to: The Chemistry Development Kit (CDK) v2.0: atom typing, depiction, molecular formulas, and substructure searching.
- Author
-
Willighagen EL, Mayfield JW, Alvarsson J, Berg A, Carlsson L, Jeliazkova N, Kuhn S, Pluskal T, Rojas-Chertó M, Spjuth O, Torrance G, Evelo CT, Guha R, and Steinbeck C
- Published
- 2017
- Full Text
- View/download PDF
23. The Chemistry Development Kit (CDK) v2.0: atom typing, depiction, molecular formulas, and substructure searching.
- Author
-
Willighagen EL, Mayfield JW, Alvarsson J, Berg A, Carlsson L, Jeliazkova N, Kuhn S, Pluskal T, Rojas-Chertó M, Spjuth O, Torrance G, Evelo CT, Guha R, and Steinbeck C
- Abstract
Background: The Chemistry Development Kit (CDK) is a widely used open source cheminformatics toolkit, providing data structures to represent chemical concepts along with methods to manipulate such structures and perform computations on them. The library implements a wide variety of cheminformatics algorithms ranging from chemical structure canonicalization to molecular descriptor calculations and pharmacophore perception. It is used in drug discovery, metabolomics, and toxicology. Over the last 10 years, the code base has grown significantly, however, resulting in many complex interdependencies among components and poor performance of many algorithms., Results: We report improvements to the CDK v2.0 since the v1.2 release series, specifically addressing the increased functional complexity and poor performance. We first summarize the addition of new functionality, such atom typing and molecular formula handling, and improvement to existing functionality that has led to significantly better performance for substructure searching, molecular fingerprints, and rendering of molecules. Second, we outline how the CDK has evolved with respect to quality control and the approaches we have adopted to ensure stability, including a code review mechanism., Conclusions: This paper highlights our continued efforts to provide a community driven, open source cheminformatics library, and shows that such collaborative projects can thrive over extended periods of time, resulting in a high-quality and performant library. By taking advantage of community support and contributions, we show that an open source cheminformatics project can act as a peer reviewed publishing platform for scientific computing software. Graphical abstract CDK 2.0 provides new features and improved performance.
- Published
- 2017
- Full Text
- View/download PDF
24. Towards agile large-scale predictive modelling in drug discovery with flow-based programming design principles.
- Author
-
Lampa S, Alvarsson J, and Spjuth O
- Abstract
Predictive modelling in drug discovery is challenging to automate as it often contains multiple analysis steps and might involve cross-validation and parameter tuning that create complex dependencies between tasks. With large-scale data or when using computationally demanding modelling methods, e-infrastructures such as high-performance or cloud computing are required, adding to the existing challenges of fault-tolerant automation. Workflow management systems can aid in many of these challenges, but the currently available systems are lacking in the functionality needed to enable agile and flexible predictive modelling. We here present an approach inspired by elements of the flow-based programming paradigm, implemented as an extension of the Luigi system which we name SciLuigi. We also discuss the experiences from using the approach when modelling a large set of biochemical interactions using a shared computer cluster.Graphical abstract.
- Published
- 2016
- Full Text
- View/download PDF
25. Large-scale ligand-based predictive modelling using support vector machines.
- Author
-
Alvarsson J, Lampa S, Schaal W, Andersson C, Wikberg JE, and Spjuth O
- Abstract
The increasing size of datasets in drug discovery makes it challenging to build robust and accurate predictive models within a reasonable amount of time. In order to investigate the effect of dataset sizes on predictive performance and modelling time, ligand-based regression models were trained on open datasets of varying sizes of up to 1.2 million chemical structures. For modelling, two implementations of support vector machines (SVM) were used. Chemical structures were described by the signatures molecular descriptor. Results showed that for the larger datasets, the LIBLINEAR SVM implementation performed on par with the well-established libsvm with a radial basis function kernel, but with dramatically less time for model building even on modest computer resources. Using a non-linear kernel proved to be infeasible for large data sizes, even with substantial computational resources on a computer cluster. To deploy the resulting models, we extended the Bioclipse decision support framework to support models from LIBLINEAR and made our models of logD and solubility available from within Bioclipse.
- Published
- 2016
- Full Text
- View/download PDF
26. Ex Vivo Assessment of Drug Activity in Patient Tumor Cells as a Basis for Tailored Cancer Therapy.
- Author
-
Blom K, Nygren P, Alvarsson J, Larsson R, and Andersson CR
- Subjects
- Acoustics, Cell Survival drug effects, Cells, Cultured, Humans, Neoplasms, Antineoplastic Agents pharmacology, Cytological Techniques methods, Drug Screening Assays, Antitumor methods
- Abstract
Although medical cancer treatment has improved during the past decades, it is difficult to choose between several first-line treatments supposed to be equally active in the diagnostic group. It is even more difficult to select a treatment after the standard protocols have failed. Any guidance for selection of the most effective treatment is valuable at these critical stages. We describe the principles and procedures for ex vivo assessment of drug activity in tumor cells from patients as a basis for tailored cancer treatment. Patient tumor cells are assayed for cytotoxicity with a panel of drugs. Acoustic drug dispensing provides great flexibility in the selection of drugs for testing; currently, up to 80 compounds and/or combinations thereof may be tested for each patient. Drug response predictions are obtained by classification using an empirical model based on historical responses for the diagnosis. The laboratory workflow is supported by an integrated system that enables rapid analysis and automatic generation of the clinical referral response., (© 2015 Society for Laboratory Automation and Screening.)
- Published
- 2016
- Full Text
- View/download PDF
27. Scaling predictive modeling in drug development with cloud computing.
- Author
-
Moghadam BT, Alvarsson J, Holm M, Eklund M, Carlsson L, and Spjuth O
- Subjects
- Databases, Factual, Internet, Ligands, Software, Computational Biology methods, Computing Methodologies, Databases, Chemical, Drug Discovery methods, Quantitative Structure-Activity Relationship
- Abstract
Growing data sets with increased time for analysis is hampering predictive modeling in drug discovery. Model building can be carried out on high-performance computer clusters, but these can be expensive to purchase and maintain. We have evaluated ligand-based modeling on cloud computing resources where computations are parallelized and run on the Amazon Elastic Cloud. We trained models on open data sets of varying sizes for the end points logP and Ames mutagenicity and compare with model building parallelized on a traditional high-performance computing cluster. We show that while high-performance computing results in faster model building, the use of cloud computing resources is feasible for large data sets and scales well within cloud instances. An additional advantage of cloud computing is that the costs of predictive models can be easily quantified, and a choice can be made between speed and economy. The easy access to computational resources with no up-front investments makes cloud computing an attractive alternative for scientists, especially for those without access to a supercomputer, and our study shows that it enables cost-efficient modeling of large data sets on demand within reasonable time.
- Published
- 2015
- Full Text
- View/download PDF
28. Benchmarking study of parameter variation when using signature fingerprints together with support vector machines.
- Author
-
Alvarsson J, Eklund M, Andersson C, Carlsson L, Spjuth O, and Wikberg JE
- Subjects
- Benchmarking, Quantitative Structure-Activity Relationship, Drug Evaluation, Preclinical methods, Support Vector Machine
- Abstract
QSAR modeling using molecular signatures and support vector machines with a radial basis function is increasingly used for virtual screening in the drug discovery field. This method has three free parameters: C, γ, and signature height. C is a penalty parameter that limits overfitting, γ controls the width of the radial basis function kernel, and the signature height determines how much of the molecule is described by each atom signature. Determination of optimal values for these parameters is time-consuming. Good default values could therefore save considerable computational cost. The goal of this project was to investigate whether such default values could be found by using seven public QSAR data sets spanning a wide range of end points and using both a bit version and a count version of the molecular signatures. On the basis of the experiments performed, we recommend a parameter set of heights 0 to 2 for the count version of the signature fingerprints and heights 0 to 3 for the bit version. These are in combination with a support vector machine using C in the range of 1 to 100 and γ in the range of 0.001 to 0.1. When data sets are small or longer run times are not a problem, then there is reason to consider the addition of height 3 to the count fingerprint and a wider grid search. However, marked improvements should not be expected.
- Published
- 2014
- Full Text
- View/download PDF
29. Ligand-based target prediction with signature fingerprints.
- Author
-
Alvarsson J, Eklund M, Engkvist O, Spjuth O, Carlsson L, Wikberg JE, and Noeske T
- Subjects
- Area Under Curve, Computer Simulation, Databases, Chemical, Ligands, Molecular Structure, ROC Curve, Drug Design, Models, Chemical, Molecular Imprinting methods, Software
- Abstract
When evaluating a potential drug candidate it is desirable to predict target interactions in silico prior to synthesis in order to assess, e.g., secondary pharmacology. This can be done by looking at known target binding profiles of similar compounds using chemical similarity searching. The purpose of this study was to construct and evaluate the performance of chemical fingerprints based on the molecular signature descriptor for performing target binding predictions. For the comparison we used the area under the receiver operating characteristics curve (AUC) complemented with net reclassification improvement (NRI). We created two open source signature fingerprints, a bit and a count version, and evaluated their performance compared to a set of established fingerprints with regards to predictions of binding targets using Tanimoto-based similarity searching on publicly available data sets extracted from ChEMBL. The results showed that the count version of the signature fingerprint performed on par with well-established fingerprints such as ECFP. The count version outperformed the bit version slightly; however, the count version is more complex and takes more computing time and memory to run so its usage should probably be evaluated on a case-by-case basis. The NRI based tests complemented the AUC based ones and showed signs of higher power.
- Published
- 2014
- Full Text
- View/download PDF
30. Bioclipse-R: integrating management and visualization of life science data with statistical analysis.
- Author
-
Spjuth O, Georgiev V, Carlsson L, Alvarsson J, Berg A, Willighagen E, Wikberg JE, and Eklund M
- Subjects
- Antineoplastic Agents chemistry, Antineoplastic Agents pharmacology, Antineoplastic Agents toxicity, Data Interpretation, Statistical, Mutagenesis, Programming Languages, Quantitative Structure-Activity Relationship, Systems Integration, Biological Science Disciplines, Computer Graphics, Software
- Abstract
Summary: Bioclipse, a graphical workbench for the life sciences, provides functionality for managing and visualizing life science data. We introduce Bioclipse-R, which integrates Bioclipse and the statistical programming language R. The synergy between Bioclipse and R is demonstrated by the construction of a decision support system for anticancer drug screening and mutagenicity prediction, which shows how Bioclipse-R can be used to perform complex tasks from within a single software system., Availability and Implementation: Bioclipse-R is implemented as a set of Java plug-ins for Bioclipse based on the R-package rj. Source code and binary packages are available from https://github.com/bioclipse and http://www.bioclipse.net/bioclipse-r, respectively., Contact: martin.eklund@farmbio.uu.se, Supplementary Information: Supplementary data are available at Bioinformatics online.
- Published
- 2013
- Full Text
- View/download PDF
31. Open source drug discovery with bioclipse.
- Author
-
Spjuth O, Carlsson L, Alvarsson J, Georgiev V, Willighagen E, and Eklund M
- Subjects
- Absorption, Algorithms, Decision Support Techniques, Drug Evaluation, Preclinical, Pharmacokinetics, Toxicology methods, Drug Discovery, Software
- Abstract
We present the open source components for drug discovery that has been developed and integrated into the graphical workbench Bioclipse. Building on a solid open source cheminformatics core, Bioclipse has advanced functionality for managing and visualizing chemical structures and related information. The features presented here include QSAR/QSPR modeling, various predictive solutions such as decision support for chemical liability assessment, site-ofmetabolism prediction, virtual screening, and knowledge discovery and integration. We demonstrate the utility of the described tools with examples from computational pharmacology, toxicology, and ADME. Bioclipse is used in both academia and industry, and is a good example of open source leading to new solutions for drug discovery.
- Published
- 2012
- Full Text
- View/download PDF
32. Brunn: an open source laboratory information system for microplates with a graphical plate layout design process.
- Author
-
Alvarsson J, Andersson C, Spjuth O, Larsson R, and Wikberg JE
- Subjects
- Databases, Factual, Dose-Response Relationship, Drug, Cytological Techniques methods, Drug Evaluation, Preclinical methods, High-Throughput Screening Assays methods, Software
- Abstract
Background: Compound profiling and drug screening generates large amounts of data and is generally based on microplate assays. Current information systems used for handling this are mainly commercial, closed source, expensive, and heavyweight and there is a need for a flexible lightweight open system for handling plate design, and validation and preparation of data., Results: A Bioclipse plugin consisting of a client part and a relational database was constructed. A multiple-step plate layout point-and-click interface was implemented inside Bioclipse. The system contains a data validation step, where outliers can be removed, and finally a plate report with all relevant calculated data, including dose-response curves., Conclusions: Brunn is capable of handling the data from microplate assays. It can create dose-response curves and calculate IC50 values. Using a system of this sort facilitates work in the laboratory. Being able to reuse already constructed plates and plate layouts by starting out from an earlier step in the plate layout design process saves time and cuts down on error sources.
- Published
- 2011
- Full Text
- View/download PDF
33. Bioclipse 2: a scriptable integration platform for the life sciences.
- Author
-
Spjuth O, Alvarsson J, Berg A, Eklund M, Kuhn S, Mäsak C, Torrance G, Wagener J, Willighagen EL, Steinbeck C, and Wikberg JE
- Subjects
- Algorithms, Biological Science Disciplines, Databases, Factual, Computational Biology methods, Software
- Abstract
Background: Contemporary biological research integrates neighboring scientific domains to answer complex questions in fields such as systems biology and drug discovery. This calls for tools that are intuitive to use, yet flexible to adapt to new tasks., Results: Bioclipse is a free, open source workbench with advanced features for the life sciences. Version 2.0 constitutes a complete rewrite of Bioclipse, and delivers a stable, scalable integration platform for developers and an intuitive workbench for end users. All functionality is available both from the graphical user interface and from a built-in novel domain-specific language, supporting the scientist in interdisciplinary research and reproducible analyses through advanced visualization of the inputs and the results. New components for Bioclipse 2 include a rewritten editor for chemical structures, a table for multiple molecules that supports gigabyte-sized files, as well as a graphical editor for sequences and alignments., Conclusion: Bioclipse 2 is equipped with advanced tools required to carry out complex analysis in the fields of bio- and cheminformatics. Developed as a Rich Client based on Eclipse, Bioclipse 2 leverages on today's powerful desktop computers for providing a responsive user interface, but also takes full advantage of the Web and networked (Web/Cloud) services for more demanding calculations or retrieval of data. The fact that Bioclipse 2 is based on an advanced and widely used service platform ensures wide extensibility, making it easy to add new algorithms, visualizations, as well as scripting commands. The intuitive tools for end users and the extensible architecture make Bioclipse 2 ideal for interdisciplinary and integrative research.Bioclipse 2 is released under the Eclipse Public License (EPL), a flexible open source license that allows additional plugins to be of any license. Bioclipse 2 is implemented in Java and supported on all major platforms; Source code and binaries are freely available at http://www.bioclipse.net.
- Published
- 2009
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.