418 results on '"Sequence prediction"'
Search Results
252. Racing tracks improvisation
- Author
-
Jiao Jian Wang and Olana Missura
- Subjects
Improvisation ,Video game development ,business.industry ,Computer science ,ComputingMilieux_PERSONALCOMPUTING ,Markov process ,Context (language use) ,Track (rail transport) ,Prediction algorithms ,symbols.namesake ,Sequence prediction ,Content generation ,symbols ,Artificial intelligence ,business - Abstract
Procedural content generation is a popular technique in the game development. One of its typical applications is generation of game levels. This paper presents a method to generate tracks for racing games, by viewing racing track generation as a discrete sequence prediction problem. To solve it we combine two techniques from music improvisation. We show that this method is capable of generating new racing tracks which appear to be interesting.
- Published
- 2014
- Full Text
- View/download PDF
253. A time series based sequence prediction algorithm to detect activities of daily living in smart home
- Author
-
Mohd. Marufuzzaman, Md. Mamun Ibne Reaz, Labonnah Farzana Rahman, and Masni Mohd Ali
- Subjects
Activities of daily living ,Computer science ,Monitoring, Ambulatory ,Health Informatics ,02 engineering and technology ,computer.software_genre ,Health Information Management ,Home automation ,Sequence prediction ,Software Design ,020204 information systems ,Component (UML) ,Activities of Daily Living ,0202 electrical engineering, electronic engineering, information engineering ,Intelligent environment ,Humans ,Duration (project management) ,Advanced and Specialized Nursing ,Series (stratigraphy) ,business.industry ,Housing ,020201 artificial intelligence & image processing ,Data mining ,State (computer science) ,business ,Algorithm ,computer ,Algorithms ,Forecasting - Abstract
SummaryObjectives: The goal of smart homes is to create an intelligent environment adapting the inhabitants need and assisting the person who needs special care and safety in their daily life. This can be reached by collecting the ADL (activities of daily living) data and further analysis within existing computing elements. In this research, a very recent algorithm named sequence prediction via enhanced episode discovery (SPEED) is modified and in order to improve accuracy time component is included.Methods: The modified SPEED or M-SPEED is a sequence prediction algorithm, which modified the previous SPEED algorithm by using time duration of appliance’s ON-OFF states to decide the next state. M-SPEED discovered periodic episodes of inhabitant behavior, trained it with learned episodes, and made decisions based on the obtained knowledge.Results: The results showed that M-SPEED achieves 96.8% prediction accuracy, which is better than other time prediction algorithms like PUBS, ALZ with temporal rules and the previous SPEED.Conclusions: Since human behavior shows natural temporal patterns, duration times can be used to predict future events more accurately. This inhabitant activity prediction system will certainly improve the smart homes by ensuring safety and better care for elderly and handicapped people.
- Published
- 2014
254. Reply to Liu: Amino acid 104 asparagine/glutamic acid of p53 is an adaptively selected site for extreme environments in mammals of the Tibet plateau
- Author
-
Hu-Yue Zu, Sheng-Ting Zhang, Min Li, Yu Liu, Yi-Bin Cao, Chung-I Wu, Yang Zhao, Ming-Yang Wang, Xue-Qun Chen, Xiao-Cheng Chen, Ji-Long Ren, Ji-Zeng Du, and Eviatar Nevo
- Subjects
chemistry.chemical_classification ,Multidisciplinary ,Ecology ,Arvicolinae ,Apoptosis ,Glutamic acid ,Biology ,Plateau (mathematics) ,Adaptation, Physiological ,Amino acid ,Cold Temperature ,Evolution, Molecular ,chemistry ,Evolutionary biology ,Sequence prediction ,Stress, Physiological ,Extreme environment ,Animals ,Humans ,Asparagine ,Letters ,Tumor Suppressor Protein p53 ,Hypoxia - Abstract
In our paper, we substantiate the fact that codon 104 variation is adaptive in highland Tibet plateau mammals (1). Liu’s letter (2) claims that codon 104 is not adaptive, which is only based on searching the sequence prediction in National Center for Biotechnology Information (NCBI) without experimental evidence.
- Published
- 2014
255. Computational protein design
- Author
-
Christina M. Kraemer-Pecore, Andrew M. Wollacott, and John R. Desjarlais
- Subjects
Models, Molecular ,Fold (higher-order function) ,Protein Conformation ,Surface Properties ,Molecular Sequence Data ,Protein design ,Chemical biology ,Proteins ,Peptidylprolyl Isomerase ,Biology ,computer.software_genre ,Biochemistry ,Combinatorial chemistry ,Analytical Chemistry ,NIMA-Interacting Peptidylprolyl Isomerase ,Protein structure ,Sequence prediction ,Optimization methods ,Combinatorial Chemistry Techniques ,Computer-Aided Design ,Computer Aided Design ,Amino Acid Sequence ,Biochemical engineering ,computer - Abstract
The field of computational protein design is reaching its adolescence. Protein design algorithms have been applied to design or engineer proteins that fold, fold faster, catalyze, catalyze faster, signal, and adopt preferred conformational states. Further developments of scoring functions, sampling strategies, and optimization methods will expand the range of applicability of computational protein design to larger and more varied systems, with greater incidence of success. Developments in this field are beginning to have significant impact on biotechnology and chemical biology.
- Published
- 2001
- Full Text
- View/download PDF
256. Estimating the significance of sequence order in protein secondary structure and prediction
- Author
-
Liisa Holm, Andreas Heger, Sabine Dietmann, and Jong Park
- Subjects
Statistics and Probability ,Databases, Factual ,Globular protein ,Molecular Sequence Data ,Protein Data Bank (RCSB PDB) ,Computational biology ,Biology ,Biochemistry ,Protein Structure, Secondary ,Protein sequencing ,Sequence prediction ,Amino Acid Sequence ,Molecular Biology ,Protein secondary structure ,Sequence (medicine) ,chemistry.chemical_classification ,Computational Biology ,Proteins ,Computer Science Applications ,Computational Mathematics ,Order (biology) ,Computational Theory and Mathematics ,chemistry ,Algorithm ,Algorithms ,Software - Abstract
Motivation: How critical is the sequence order information in predicting protein secondary structure segments? We tried to get a rough insight on it from a theoretical approach using both a prediction algorithm and structural fragments from Protein Databank (PDB). Results: Using reverse protein sequences and PDB structural fragments, we theoretically estimated the significance of the order for protein secondary structure and prediction. On average: (1) 79% of protein sequence segments resulted in the same prediction in both normal and reverse directions, which indicated a relatively high conservation of secondary structure propensity in the reverse direction; (2) the reversed sequence prediction alone performed less accurately than the normal forward sequence prediction, but comparably high (2% difference); (3) the commonly predicted regions showed a slightly higher prediction accuracy (4%) than the normal sequences prediction; and (4) structural fragments which have counterparts in reverse direction in the same protein showed a comparable degree of secondary structure conservation (73% identity with reversed structures on average for pentamers). Contact: jong@biosophy.org; dietmann@ebi.ac.uk; heger@ebi.ac.uk; holm@ebi.ac.uk *** Present address: EBI, Genome Campus, Hinxton, Cambridge, CB10 1SD, UK. To whom correspondence should be addressed.
- Published
- 2000
- Full Text
- View/download PDF
257. Antipredictable Sequences: Harder to Predict Than Random Sequences
- Author
-
Wolfgang Kinzel and Huaiyu Zhu
- Subjects
Sequence ,Computational complexity theory ,Artificial neural network ,Cognitive Neuroscience ,Existential quantification ,Models, Neurological ,Reproducibility of Results ,Models, Psychological ,Random sequence ,Random Allocation ,Arts and Humanities (miscellaneous) ,Margin (machine learning) ,Simple (abstract algebra) ,Sequence prediction ,Learning ,Neural Networks, Computer ,Algorithm ,Algorithms ,Mathematics - Abstract
For any discrete-state sequence prediction algorithm A, it is always possible, using an algorithm B no more complicated than A, to generate a sequence for which A's prediction is always wrong. For any prediction algorithm A and sequence x, there exists a sequence y no more complicated than x, such that if A performs better than random on x, then it will perform worse than random on y by the same margin. An example of a simple neural network predicting a bit sequence is used to illustrate this very general but not widely recognized phenomenon. This implies that any predictor with good performance must rely on some (usually implicitly) assumed prior distributions of the problem.
- Published
- 1998
- Full Text
- View/download PDF
258. Sequential Prediction for Information Fusion and Control
- Author
-
Maxim Raginsky and Rebecca Willett
- Subjects
Stochastic control ,Intersection (set theory) ,business.industry ,Computer science ,Control (management) ,Adversary ,Machine learning ,computer.software_genre ,Information fusion ,Adversarial system ,Control theory ,Sequence prediction ,Artificial intelligence ,business ,computer - Abstract
The goal of this work was to develop sequential prediction methods for online information fusion and control, with methods designed to handle unknown, environmental dynamics, potentially stemming from an adversary who reacts to sensing actions, active sensing paradigms, and external feedback mechanisms. Online prediction and targeted collection of information is an emerging paradigm at the intersection of optimization, machine learning and control theory, which is concerned with real-time sequential planning of actions or decisions in the presence of model uncertainty, nonstationarity, and possibly adversarial disturbances. Several methods and underlying supporting theory which meet these objectives are described in detail in the following final report.
- Published
- 2013
- Full Text
- View/download PDF
259. Context codes and the effect of noisy learning on a simplified hippocampal CA3 model
- Author
-
William B. Levy, Xiangbao Wu, and Robert Baxter
- Subjects
Neurons ,Sequence ,Quantitative Biology::Neurons and Cognition ,General Computer Science ,Computer science ,business.industry ,Models, Neurological ,SIGNAL (programming language) ,Complex system ,Context (language use) ,Hippocampal formation ,Hippocampus ,Noise ,nervous system ,Sequence prediction ,Learning ,Artificial intelligence ,Sequence learning ,Artifacts ,business ,Mathematics ,Biotechnology - Abstract
This paper investigates how noise affects a minimal computational model of the hippocampus and, in particular, region CA3. The architecture and physiology employed are consistent with the known anatomy and physiology of this region. Here, we use computer simulations to demonstrate and quantify the ability of this model to create context codes in sequential learning problems. These context codes are mediated by local context neurons which are analogous to hippocampal place-coding cells. These local context neurons endow the network with many of its problem-solving abilities. Our results show that the network encodes context on its own and then uses context to solve sequence prediction under ambiguous conditions. Noise during learning affects performance, and it also affects the development of context codes. The relationship between noise and performance in a sequence prediction is simple and corresponds to a disruption of local context neuron firing. As noise exceeds the signal, sequence completion and local context neuron firing are both lost. For the parameters investigated, extra learning trials and slower learning rates do not overcome either of the effects of noise. The results are consistent with the important role played, in this hippocampal model, by local context neurons in sequence prediction and for disambiguation across time.
- Published
- 1996
- Full Text
- View/download PDF
260. Modelling splice sites with locality-sensitive sequence features
- Author
-
Yin-Fu Huang and Sing-Wu Liou
- Subjects
Sequence ,Receiver operating characteristic ,business.industry ,Computer science ,RNA Splicing ,Locality ,Pattern recognition ,Library and Information Sciences ,Models, Theoretical ,computer.software_genre ,General Biochemistry, Genetics and Molecular Biology ,ROC Curve ,Sequence prediction ,Prediction methods ,Outlier ,RNA Precursors ,splice ,Artificial intelligence ,Data mining ,RNA Splice Sites ,business ,computer ,Information Systems - Abstract
The splice sites are essential for pre-mRNA maturation and crucial for Splice Site Modelling (SSM); however, there are gaps between the splicing signals and the computationally identified sequence features. In this paper, the Locality Sensitive Features (LSFs) are proposed to reduce the gaps by homogenising their contexts. Under the skewness-kurtosis based statistics and data analysis, SSM attributed with LSFs is fulfilled by double-boundary outlier filters. The LSF-based SSM had been applied to six model organisms of diverse species; by the accuracy and Receiver Operating Characteristic (ROC) analysis, the promising results show the proposed methodology is versatile and robust for the splice-site classification. It is prospective the LSF-based SSM can serve as a new infrastructure for developing effective splice-site prediction methods and have the potential to be applied to other sequence prediction problems.
- Published
- 2013
261. Predicting Mobile Call Behavior via Subspace Methods
- Author
-
Shen-Shyang Ho, Peng Dai, and Wanqing Yang
- Subjects
Computer science ,business.industry ,computer.software_genre ,Machine learning ,Independent component analysis ,Task (project management) ,Behavioral data ,Sequence prediction ,Principal component analysis ,Data mining ,Artificial intelligence ,business ,computer ,Subspace topology - Abstract
We investigate behavioral prediction approaches based on subspace methods such as principal component analysis (PCA) and independent component analysis (ICA). Moreover, we propose a personalized sequential prediction approach to predict next day behavior based on features extracted from past behavioral data using subspace methods. The proposed approach is applied to the individual call (voice calls and short messages) behavior prediction task. Experimental results on the Nokia mobility data challenge (MDC) dataset are used to show the feasibility of our proposed prediction approach. Furthermore, we investigate whether prediction accuracy can be improved (i) when specific call type (voice call or short message), instead of the general call behavior prediction, is considered in the prediction task, and (ii) when workday and weekend scenarios are considered separately.
- Published
- 2013
- Full Text
- View/download PDF
262. Universal codes from switching strategies
- Author
-
Steven de Rooij, Wouter M. Koolen, and Algorithms and Complexity
- Subjects
FOS: Computer and information sciences ,Theoretical computer science ,Computer science ,Information Theory (cs.IT) ,Computer Science - Information Theory ,Parameterized complexity ,Regret ,Library and Information Sciences ,Machine Learning (cs.LG) ,Computer Science Applications ,Computer Science - Learning ,Sequence prediction ,Time series ,Hidden Markov model ,Information Systems - Abstract
We discuss algorithms for combining sequential prediction strategies, a task which can be viewed as a natural generalization of the concept of universal coding. We describe a graphical language based on hidden Markov models for defining prediction strategies, and we provide both existing and new models as examples. The models include efficient, parameterless models for switching between the input strategies over time, including a model for the case where switches tend to occur in clusters, and finally a new model for the scenario where the prediction strategies have a known relationship, and where jumps are typically between strongly related ones. This last model is relevant for coding time series data where parameter drift is expected. As theoretical contributions, we introduce an interpolation construction that is useful in the development and analysis of new algorithms, and we establish a new sophisticated lemma for analyzing the individual sequence regret of parameterized models.
- Published
- 2013
263. Epilogue: Where to go from here
- Author
-
Peter Flach
- Subjects
Preference learning ,Computer science ,business.industry ,Active learning (machine learning) ,Online learning ,Deep learning ,Multi-task learning ,Machine learning ,computer.software_genre ,Sequence prediction ,Reinforcement learning ,Artificial intelligence ,Transfer of learning ,business ,computer - Published
- 2012
- Full Text
- View/download PDF
264. Residue-specific immunochemical sequence prediction
- Author
-
Jens F. Rehfeld and Anders H. Johnsen
- Subjects
Residue (complex analysis) ,Binding Sites ,Molecular Sequence Data ,Snails ,Immunology ,Computational biology ,Biology ,Immunohistochemistry ,Sensitivity and Specificity ,Peptide Fragments ,Ganglia, Invertebrate ,Epitopes ,Sequence prediction ,Gastrins ,Animals ,Immunology and Allergy ,Amino Acid Sequence ,Cholecystokinin ,Sequence Analysis - Published
- 1994
- Full Text
- View/download PDF
265. LEARNABILITY IN PROBLEMS OF SEQUENTIAL INFERENCE
- Author
-
Ryabko, Daniil, Sequential Learning (SEQUEL), Laboratoire d'Informatique Fondamentale de Lille (LIFL), Université de Lille, Sciences et Technologies-Institut National de Recherche en Informatique et en Automatique (Inria)-Université de Lille, Sciences Humaines et Sociales-Centre National de la Recherche Scientifique (CNRS)-Université de Lille, Sciences et Technologies-Institut National de Recherche en Informatique et en Automatique (Inria)-Université de Lille, Sciences Humaines et Sociales-Centre National de la Recherche Scientifique (CNRS)-Laboratoire d'Automatique, Génie Informatique et Signal (LAGIS), Université de Lille, Sciences et Technologies-Centrale Lille-Centre National de la Recherche Scientifique (CNRS)-Centrale Lille-Centre National de la Recherche Scientifique (CNRS)-Inria Lille - Nord Europe, Institut National de Recherche en Informatique et en Automatique (Inria), Université des Sciences et Technologie de Lille - Lille I, Max Dauchet(max.dauchet@inria.fr), Université de Lille, Sciences et Technologies-Institut National de Recherche en Informatique et en Automatique (Inria)-Université de Lille, Sciences Humaines et Sociales-Centre National de la Recherche Scientifique (CNRS)-Université de Lille, Sciences et Technologies-Institut National de Recherche en Informatique et en Automatique (Inria)-Université de Lille, Sciences Humaines et Sociales-Centre National de la Recherche Scientifique (CNRS)-Inria Lille - Nord Europe, Institut National de Recherche en Informatique et en Automatique (Inria)-Laboratoire d'Automatique, Génie Informatique et Signal (LAGIS), and Université de Lille, Sciences et Technologies-Centrale Lille-Centre National de la Recherche Scientifique (CNRS)-Centrale Lille-Centre National de la Recherche Scientifique (CNRS)
- Subjects
learnability ,[INFO.INFO-LG]Computer Science [cs]/Machine Learning [cs.LG] ,[STAT.ML]Statistics [stat]/Machine Learning [stat.ML] ,[INFO.INFO-IT]Computer Science [cs]/Information Theory [cs.IT] ,Machine learning ,hypothesis testing ,[MATH.MATH-IT]Mathematics [math]/Information Theory [math.IT] ,[STAT.TH]Statistics [stat]/Statistics Theory [stat.TH] ,time series ,sequence prediction ,apprentissage artificiel ,[INFO.INFO-AI]Computer Science [cs]/Artificial Intelligence [cs.AI] - Abstract
Given a growing sequence of observations x_1,...,x_n,..., one is required, at each time step n, to make some inference about the stochastic mechanism generating the sequence. Several problems that have numerous applications in different branches of mathematics and computer science can be formulated in this way. For example, one may want to forecast probabilities of the next outcome x_{n+1} (sequence prediction); to make a decision on whether the mechanism generating the sequence belongs to a certain family $H_0$ versus it belongs to a different family $H_1$ (hypothesis testing); to take an action in order to maximize some utility function. In each of these problems, as well as in many others, in order to be able to make inference, one has to make some assumptions on the probabilistic mechanism generating the data. Typical assumptions are that x_i are independent and identically distributed, or that the distribution generating the sequence belongs to a certain parametric family. The central question addressed in this work is: under which assumptions is inference possible? This question is considered for several problems of inference, including sequence prediction, hypothesis testing, classification and reinforcement learning.; Les travaux présentés sont dédiés à la possibilité de faire de l'inférence statistique à partir de données séquentielles. Le problème est le suivant. Étant donnée une suite d'observations x_1,...,x_n,..., on cherche à faire de l'inférence sur le processus aléatoire ayant produit la suite. Plusieurs problèmes, qui d'ailleurs ont des applications multiples dans différents domaines des mathématiques et de l'informatique, peuvent être formulés ainsi. Par exemple, on peut vouloir prédire la probabilité d'apparition de l'observation suivante, x_{n+1} (le problème de prédiction séquentielle); ou répondre à la question de savoir si le processus aléatoire qui produit la suite appartient à un certain ensemble H_0 versus appartient à un ensemble différent H_1 (test d'hypothèse) ; ou encore, effectuer une action avec le but de maximiser une certain fonction d'utilité. Dans chacun de ces problèmes, pour rendre l'inférence possible il faut d'abord faire certaines hypothèses sur le processus aléatoire qui produit les données. La question centrale adressée dans les travaux présentés est la suivante : sous quelles hypothèses l'inférence est-elle possible ? Cette question est posée et analysée pour des problèmes d'inférence différents, parmi lesquels se trouvent la prédiction séquentielle, les tests d'hypothèse, la classification et l'apprentissage par renforcement.
- Published
- 2011
266. APPRENABILITÉ DANS LES PROBLÈMES DE L'INFÉRENCE SÉQUENTIELLE
- Author
-
Ryabko, Daniil, Sequential Learning (SEQUEL), Laboratoire d'Informatique Fondamentale de Lille (LIFL), Université de Lille, Sciences et Technologies-Institut National de Recherche en Informatique et en Automatique (Inria)-Université de Lille, Sciences Humaines et Sociales-Centre National de la Recherche Scientifique (CNRS)-Université de Lille, Sciences et Technologies-Institut National de Recherche en Informatique et en Automatique (Inria)-Université de Lille, Sciences Humaines et Sociales-Centre National de la Recherche Scientifique (CNRS)-Laboratoire d'Automatique, Génie Informatique et Signal (LAGIS), Université de Lille, Sciences et Technologies-Centrale Lille-Centre National de la Recherche Scientifique (CNRS)-Centrale Lille-Centre National de la Recherche Scientifique (CNRS)-Inria Lille - Nord Europe, Institut National de Recherche en Informatique et en Automatique (Inria), Université des Sciences et Technologie de Lille - Lille I, and Max Dauchet(max.dauchet@inria.fr)
- Subjects
learnability ,[INFO.INFO-LG]Computer Science [cs]/Machine Learning [cs.LG] ,[STAT.ML]Statistics [stat]/Machine Learning [stat.ML] ,[INFO.INFO-IT]Computer Science [cs]/Information Theory [cs.IT] ,Machine learning ,hypothesis testing ,[MATH.MATH-IT]Mathematics [math]/Information Theory [math.IT] ,[STAT.TH]Statistics [stat]/Statistics Theory [stat.TH] ,time series ,sequence prediction ,apprentissage artificiel ,[INFO.INFO-AI]Computer Science [cs]/Artificial Intelligence [cs.AI] - Abstract
Given a growing sequence of observations x_1,...,x_n,..., one is required, at each time step n, to make some inference about the stochastic mechanism generating the sequence. Several problems that have numerous applications in different branches of mathematics and computer science can be formulated in this way. For example, one may want to forecast probabilities of the next outcome x_{n+1} (sequence prediction); to make a decision on whether the mechanism generating the sequence belongs to a certain family $H_0$ versus it belongs to a different family $H_1$ (hypothesis testing); to take an action in order to maximize some utility function. In each of these problems, as well as in many others, in order to be able to make inference, one has to make some assumptions on the probabilistic mechanism generating the data. Typical assumptions are that x_i are independent and identically distributed, or that the distribution generating the sequence belongs to a certain parametric family. The central question addressed in this work is: under which assumptions is inference possible? This question is considered for several problems of inference, including sequence prediction, hypothesis testing, classification and reinforcement learning.; Les travaux présentés sont dédiés à la possibilité de faire de l'inférence statistique à partir de données séquentielles. Le problème est le suivant. Étant donnée une suite d'observations x_1,...,x_n,..., on cherche à faire de l'inférence sur le processus aléatoire ayant produit la suite. Plusieurs problèmes, qui d'ailleurs ont des applications multiples dans différents domaines des mathématiques et de l'informatique, peuvent être formulés ainsi. Par exemple, on peut vouloir prédire la probabilité d'apparition de l'observation suivante, x_{n+1} (le problème de prédiction séquentielle); ou répondre à la question de savoir si le processus aléatoire qui produit la suite appartient à un certain ensemble H_0 versus appartient à un ensemble différent H_1 (test d'hypothèse) ; ou encore, effectuer une action avec le but de maximiser une certain fonction d'utilité. Dans chacun de ces problèmes, pour rendre l'inférence possible il faut d'abord faire certaines hypothèses sur le processus aléatoire qui produit les données. La question centrale adressée dans les travaux présentés est la suivante : sous quelles hypothèses l'inférence est-elle possible ? Cette question est posée et analysée pour des problèmes d'inférence différents, parmi lesquels se trouvent la prédiction séquentielle, les tests d'hypothèse, la classification et l'apprentissage par renforcement.
- Published
- 2011
267. Assessing time-varying causality network of ensemble neural spiking activity
- Author
-
Sanggyun Kim, Marcelo Aguilar, and Todd P. Coleman
- Subjects
Computer science ,business.industry ,General Neuroscience ,lcsh:QP351-495 ,Stimulus (physiology) ,Point process ,lcsh:RC321-571 ,Cellular and Molecular Neuroscience ,Exponential family ,Probabilistic method ,lcsh:Neurophysiology and neuropsychology ,Granger causality ,Sequence prediction ,Simulated data ,Poster Presentation ,Neural system ,Artificial intelligence ,business ,lcsh:Neurosciences. Biological psychiatry. Neuropsychiatry - Abstract
Neurons in many brain regions change their spiking responses and interactions among them to relevant stimuli. Tracking the dynamics of neural system is crucial for understanding how neural systems adapt their responses to relevant biological information. Granger causality [1] has been effectively used to assess directional interactions between continuous neural signals, but it cannot be directly applied to neural spike trains viewed as point processes. Recently, methods that extend Granger’s viewpoint to the point process modality have been developed [2], [3] to identify causal interactions between neural spike trains. These methods, however, depend upon stationarity assumptions – which might not be valid when the interactive causal influences themselves are time-varying. We propose a novel probabilistic method for tracking the time-varying causal neural interactions based on sequential prediction of point process models. The time-varying causality from neuron x to y is assessed by the variability of a windowed version of the point process log-likelihood ratio (LLR), where one model incorporates only the past y and the other incorporates the past of both x and y. The proposed method successfully tracks the time-varying causal network in simulated data, and when applied to real neural data recorded in the rat insular cortex, it identifies the change of causal relationships between neurons to a relevant behavioral stimulus (see Figure Figure11). Figure 1 Tracking time-varying causality network. Sij(t) represents time-varying causality effect from neuron i to j. A. Simulation: Proposed time-varying causality measure had larger values when x caused y than when x did not cause y. B. Real data analysis: Time-varying ...
- Published
- 2011
268. A generalized prediction framework for granger causality
- Author
-
Christopher J. Quinn, Todd P. Coleman, and Negar Kiyavash
- Subjects
Granger causality ,Computer science ,Stochastic process ,Sequence prediction ,Statistics ,Econometrics ,Entropy (information theory) ,Regret ,Side information ,Minimax ,Electronic mail - Abstract
In his 1969 paper, Granger proposed a statistical definition of causality between stochastic processes. It is based on whether causal side information helps in a sequential prediction task. However, his formulation was limited to linear predictors. We describe a generalized framework, where predictions are beliefs and compare the best predictor with side information to the best predictor without side information. The difference in the prediction performance, i.e., regret of such predictors, is used as a measure of causal influence of the side information. Specifically when log loss is used to quantify each predictor's loss and an expectation over the outcomes is used to quantify the regret, we show that the directed information, an information theoretic quantity, quantifies Granger causality. We also explore a more pessimistic setup perhaps better suited for adversarial settings where minimax criterion is used to quantify the regret.
- Published
- 2011
- Full Text
- View/download PDF
269. Nonparametric sequential prediction for stationary processes
- Author
-
Benjamin Weiss and Gusztáv Morvai
- Subjects
Statistics and Probability ,Nonparametric predicton ,Probability (math.PR) ,Nonparametric statistics ,Combinatorics ,Sequence prediction ,Scheme (mathematics) ,TheoryofComputation_ANALYSISOFALGORITHMSANDPROBLEMCOMPLEXITY ,stationary processes ,FOS: Mathematics ,Ergodic theory ,60G25 ,62G05 ,Statistics, Probability and Uncertainty ,60G45 ,Mathematics - Probability ,Mathematics - Abstract
We study the problem of finding an universal estimation scheme $h_n:\mathbb{R}^n\to \mathbb{R}$, $n=1,2,...$ which will satisfy \lim_{t\rightarrow\infty}{\frac{1}{t}}\sum_{i=1}^t|h_ i(X_0,X_1,...,X_{i-1})-E(X_i|X_0,X_1,...,X_{i-1})|^p=0 a.s. for all real valued stationary and ergodic processes that are in $L^p$. We will construct a single such scheme for all $1, Comment: Published in at http://dx.doi.org/10.1214/10-AOP576 the Annals of Probability (http://www.imstat.org/aop/) by the Institute of Mathematical Statistics (http://www.imstat.org)
- Published
- 2011
- Full Text
- View/download PDF
270. Learning from Partially Annotated Sequences
- Author
-
Ulf Brefeld and Eraldo R. Fernandes
- Subjects
Computer science ,business.industry ,Hide markov model ,computer.software_genre ,Perceptron ,Machine learning ,Task (project management) ,ComputingMethodologies_PATTERNRECOGNITION ,Named-entity recognition ,Sequence prediction ,Wisdom of the crowd ,Labeled data ,Artificial intelligence ,business ,computer - Abstract
We study sequential prediction models in cases where only fragments of the sequences are annotated with the ground-truth. The task does not match the standard semi-supervised setting and is highly relevant in areas such as natural language processing, where completely labeled instances are expensive and require editorial data. We propose to generalize the semi-supervised setting and devise a simple transductive loss-augmented perceptron to learn from inexpensive partially annotated sequences that could for instance be provided by laymen, the wisdom of the crowd, or even automatically. Experiments on mono- and crosslingual named entity recognition tasks with automatically generated partially annotated sentences from Wikipedia demonstrate the effectiveness of the proposed approach. Our results show that learning from partially labeled data is never worse than standard supervised and semi-supervised approaches trained on data with the same ratio of labeled and unlabeled tokens.
- Published
- 2011
- Full Text
- View/download PDF
271. Characterization and sequence prediction of structural variations in ?-helix
- Author
-
Pramod P. Wangikar and Ashish V. Tendulkar
- Subjects
Context (language use) ,Biology ,Characterization (mathematics) ,Machine learning ,computer.software_genre ,Biochemistry ,Protein Structure, Secondary ,Protein structure ,Structural Biology ,Position (vector) ,Amino Acids ,Molecular Biology ,Sequence (medicine) ,business.industry ,Applied Mathematics ,Research ,Computational Biology ,Proteins ,Dot product ,Computer Science Applications ,Support vector machine ,Helix ,Aromatic amino acid ,Computational design ,Polar amino acids ,Sequence prediction ,Sequence structure ,Structural distortions ,Structural perturbation ,Structural variations ,Forecasting ,Support vector machines ,Bioinformatics ,Amino acids ,amino acid ,protein ,algorithm ,biology ,chemistry ,computer program ,methodology ,protein secondary structure ,Algorithms ,Software ,Artificial intelligence ,Biological system ,business ,computer - Abstract
Background: The structure conservation in various ?-helix subclasses reveals the sequence and context dependent factors causing distortions in the ?-helix. The sequence-structure relationship in these subclasses can be used to predict structural variations in ?-helix purely based on its sequence. We train support vector machine(SVM) with dot product kernel function to discriminate between regular ?-helix and non-regular ?-helices purely based on the sequences, which are represented with various overall and position specific propensities of amino acids.Results: We characterize the structural distortions in five ?-helix subclasses. The sequence structure correlation in the subclasses reveals that the increased propensity of proline, histidine, serine, aspartic acid and aromatic amino acids are responsible for the distortions in regular ?-helix. The N-terminus of regular ?-helix prefers neutral and acidic polar amino acids, while the C-terminus prefers basic polar amino acid. Proline is preferred in the first turn of regular ?-helix , while it is preferred to produce kinked and curved subclasses. The SVM discriminates between regular ?-helix and the rest with precision of 80.97% and recall of 88.05%.Conclusions: The correlation between structural variation in helices and their sequences is manifested by the performance of SVM based on sequence features. The results presented here are useful for computational design of helices. The results are also useful for prediction of structural perturbations in helix sequence purely based on its sequence. � 2011 Tendulkar and Wangikar; licensee BioMed Central Ltd.
- Published
- 2011
272. Brain Journal - Some Consequences Of The Complexity Of Intelligent Prediction
- Author
-
Joel Ratsaby
- Subjects
descriptive complexity ,learning ,sequence prediction - Abstract
What is the relationship between the complexity of a learner and the randomness of his mistakes ? This question was posed in [4] who showed that the more complex the learner the higher the possibility that his mistakes deviate from a true random sequence. In the current paper we report on an empirical investigation of this problem. We investigate two characteristics of randomness, the stochastic and algorithmic complexity of the binary sequence of mistakes. A learner with a Markov model of order k is trained on a finite binary sequence produced by a Markov source of order k ∗ and is tested on a different random sequence. As a measure of learner’s complexity we define a quantity called the sysRatio, denoted by ρ, which is the ratio between the compressed and uncompressed lengths of the binary string whose i th bit represents the maximum a posteriori decision made at state i of the learner’s model. The quantity ρ is a measure of information density. The main result of the paper shows that this ratio is crucial in answering the above posed question. The result indicates that there is a critical threshold ρ ∗ such that when ρ ≤ ρ ∗ the sequence of mistakes possesses the following features: (1) low divergence ∆ from a random sequence, (2) low variance in algorithmic complexity. When ρ > ρ∗ , the characteristics of the mistake sequence changes sharply towards a high ∆ and high variance in algorithmic complexity. It is also shown that the quantity ρ is inversely proportional to k and the value of ρ ∗ corresponds to the value k ∗ . This is the point where the learner’s model becomes too simple and is unable to approximate the Bayes optimal decision. Here the characteristics of the mistake sequence change sharply, https://www.edusoft.ro/brain/index.php/brain/article/view/114/241
- Published
- 2010
- Full Text
- View/download PDF
273. Sequence prediction in realizable and non-realizable cases
- Author
-
Ryabko, Daniil, Sequential Learning (SEQUEL), Laboratoire d'Informatique Fondamentale de Lille (LIFL), Université de Lille, Sciences et Technologies-Institut National de Recherche en Informatique et en Automatique (Inria)-Université de Lille, Sciences Humaines et Sociales-Centre National de la Recherche Scientifique (CNRS)-Université de Lille, Sciences et Technologies-Institut National de Recherche en Informatique et en Automatique (Inria)-Université de Lille, Sciences Humaines et Sociales-Centre National de la Recherche Scientifique (CNRS)-Laboratoire d'Automatique, Génie Informatique et Signal (LAGIS), Université de Lille, Sciences et Technologies-Centrale Lille-Centre National de la Recherche Scientifique (CNRS)-Centrale Lille-Centre National de la Recherche Scientifique (CNRS)-Inria Lille - Nord Europe, Institut National de Recherche en Informatique et en Automatique (Inria), Université de Lille, Sciences et Technologies-Institut National de Recherche en Informatique et en Automatique (Inria)-Université de Lille, Sciences Humaines et Sociales-Centre National de la Recherche Scientifique (CNRS)-Université de Lille, Sciences et Technologies-Institut National de Recherche en Informatique et en Automatique (Inria)-Université de Lille, Sciences Humaines et Sociales-Centre National de la Recherche Scientifique (CNRS)-Inria Lille - Nord Europe, Institut National de Recherche en Informatique et en Automatique (Inria)-Laboratoire d'Automatique, Génie Informatique et Signal (LAGIS), and Université de Lille, Sciences et Technologies-Centrale Lille-Centre National de la Recherche Scientifique (CNRS)-Centrale Lille-Centre National de la Recherche Scientifique (CNRS)
- Subjects
[INFO.INFO-LG]Computer Science [cs]/Machine Learning [cs.LG] ,[MATH.MATH-ST]Mathematics [math]/Statistics [math.ST] ,[INFO.INFO-IT]Computer Science [cs]/Information Theory [cs.IT] ,[MATH.MATH-IT]Mathematics [math]/Information Theory [math.IT] ,[STAT.TH]Statistics [stat]/Statistics Theory [stat.TH] ,time series ,sequence prediction - Abstract
International audience; A sequence $x_1,\dots,x_n,\dots$ of discrete-valued observations is generated according to some unknown probabilistic law (measure) $\mu$. After observing each outcome, it is required to give the conditional probabilities of the next observation. The realizable case is when the measure $\mu$ belongs to an arbitrary but known class $C$ of process measures. The non-realizable case is when $\mu$ is completely arbitrary, but the prediction performance is measured with respect to a given set $C$ of process measures. We are interested in the relations between these problems and between their solutions, as well as in characterizing the cases when a solution exists, and finding these solutions. We show that if the quality of prediction is measured by total variation distance, then these problems coincide, while if it is measured by expected average KL divergence, then they are different. For some of the formalizations we also show that when a solution exists, it can be obtained as a Bayes mixture over a countable subset of $C$. As an illustration to the general results obtained, we show that a solution to the non-realizable case of the sequence prediction problem exists for the set of all finite-memory processes, but does not exist for the set of all stationary processes.
- Published
- 2010
274. Nonparametric sequential prediction of time series
- Author
-
Gérard Biau, László Györfi, Kevin Bleakley, György Ottucsák, Laboratoire de Probabilités et Modèles Aléatoires (LPMA), and Université Pierre et Marie Curie - Paris 6 (UPMC)-Université Paris Diderot - Paris 7 (UPD7)-Centre National de la Recherche Scientifique (CNRS)
- Subjects
Statistics and Probability ,FOS: Computer and information sciences ,Kernel density estimation ,02 engineering and technology ,Machine learning ,computer.software_genre ,01 natural sciences ,Field (computer science) ,Set (abstract data type) ,Methodology (stat.ME) ,010104 statistics & probability ,Sequence prediction ,[MATH.MATH-ST]Mathematics [math]/Statistics [math.ST] ,0202 electrical engineering, electronic engineering, information engineering ,Econometrics ,FOS: Mathematics ,0101 mathematics ,Time series ,Statistics - Methodology ,ComputingMilieux_MISCELLANEOUS ,Mathematics ,Series (mathematics) ,business.industry ,Probability (math.PR) ,Nonparametric statistics ,020206 networking & telecommunications ,[STAT.TH]Statistics [stat]/Statistics Theory [stat.TH] ,Universal consistency ,Artificial intelligence ,Statistics, Probability and Uncertainty ,business ,62G99 ,computer ,Mathematics - Probability - Abstract
Time series prediction covers a vast field of every-day statistical applications in medical, environmental and economic domains. In this paper we develop nonparametric prediction strategies based on the combination of a set of 'experts' and show the universal consistency of these strategies under a minimum of conditions. We perform an in-depth analysis of real-world data sets and show that these nonparametric strategies are more flexible, faster and generally outperform ARMA methods in terms of normalized cumulative prediction error., Comment: article + 2 figures
- Published
- 2010
- Full Text
- View/download PDF
275. Application of grey relational clustering and CGNN in analyzing stability control of surrounding rocks in deep entry of coal mine
- Author
-
Wanbin Yang and Zhiming Qu
- Subjects
Pattern clustering ,Artificial neural network ,Computer science ,business.industry ,Stability (learning theory) ,Coal mining ,computer.software_genre ,Trend prediction ,Electronic stability control ,Sequence prediction ,Data mining ,Cluster analysis ,business ,computer - Abstract
With combination of grey neural network (CGNN) and grey relational clustering, the models are constructed, which are used to solve the prediction and comparison of surrounding rocks stability controlling parameters in deep entry of coal mine. The results show that grey relational clustering is an effective way and CGNN has perfect ability to be studied in a short-term prediction. Combined grey neural network has the features of trend and fluctuation while combining with the time-dependent sequence prediction. It is concluded that great improvements compared with any methods of trend prediction and simple factor in combined grey neural network is stated and described in stably controlling the surrounding rocks in deep entry.
- Published
- 2009
- Full Text
- View/download PDF
276. Application of Grey Relation Clustering and CGNN in Gas Concentration Prediction in Top Corner of Coal Mine
- Author
-
Liang Xiaoying and Qu Zhiming
- Subjects
Relation (database) ,Artificial neural network ,business.industry ,Coal mining ,Gas concentration ,computer.software_genre ,Trend prediction ,Sequence prediction ,Coal ,Data mining ,business ,Cluster analysis ,computer ,Mathematics - Abstract
Using grey relation clustering and combined grey neural network (CGNN), the combined model is setup, which aims at solving the problems of predicting and comparing the gas concentration in top corner of coal mine. Through comparison and prediction, the results show that, in short-term prediction, grey relation clustering is an effective way and CGNN has perfect ability to study. CGNN has the dual properties of trend and fluctuation under the condition of combining with the time-dependent sequence prediction. It is concluded that great improvement comparing with any methods of trend prediction and simple factor in CGNN is stated and described in gas concentration in top corner of coal mine.
- Published
- 2009
- Full Text
- View/download PDF
277. Prediction of the Number of International Tourists in China Based on Gray Model (1,1)
- Author
-
Yang Lingbin, Wu Jin, and Zhang Xia
- Subjects
Computer science ,Sequence prediction ,High elevation ,Econometrics ,Operations management ,China ,Gray (horse) ,Tourism ,Small probability ,Data modeling - Abstract
A Gray Model for international tourists prediction in China is established based on Gray System Theory and related data since 1978. Tested by Posterior Check and Small Probability, the model is proved qualified to predict the number of international tourists with high elevation fitting precision. Gray System theory is a powerful tool to study tourist phenomena, because tourism system is belonging to Gray System. This paper will be a useful reference to the study on tourism sequence prediction.
- Published
- 2009
- Full Text
- View/download PDF
278. Sequential Prediction of Daily Groundwater Levels by a Neural Network Model Based on Weather Forecasts
- Author
-
C. A. S. Farias, A. Kadota, and Koichi Suzuki
- Subjects
Global Forecast System ,Water resources ,Trustworthiness ,Artificial neural network ,Meteorology ,Sequence prediction ,Sustainable management ,Environmental science ,Precipitation ,Groundwater - Abstract
This paper investigates the implementation of an Artificial Neural Network (ANN) model for sequential prediction of daily groundwater levels based on precipitation forecasts. The basic principle of the ANN-based procedure consists of relating previous daily groundwater levels and daily precipitation forecasts in order to predict daily groundwater levels up to seven days ahead. The daily precipitation values up to one week ahead are assumed to be deterministic since meteorological short-range forecasts are generally available. The methodology is applied to the groundwater system of Matsuyama City, Japan. Insufficiency of water is a periodical problem in this city and thus accurate predictions of groundwater levels are very important to improve the water resources management in the region. The excellent accuracy obtained by the ANN model indicates that it is very efficient for the multi-step-ahead prediction of daily groundwater levels. As conclusion, this methodology may provide trustworthy data for the application of models to the sustainable management of Matsuyama’s groundwater system.
- Published
- 2009
- Full Text
- View/download PDF
279. DirecTL
- Author
-
Grzegorz Kondrak, Qing Dou, Kenneth Dwyer, Aditya Bhargava, and Sittichai Jiampojamarn
- Subjects
Dynamic programming ,Training set ,Character (mathematics) ,Discriminative model ,Sequence modeling ,Computer science ,Sequence prediction ,Speech recognition ,Transliteration ,Segmentation - Abstract
We present DirecTL: an online discriminative sequence prediction model that employs a many-to-many alignment between target and source. Our system incorporates input segmentation, target character prediction, and sequence modeling in a unified dynamic programming framework. Experimental results suggest that DirecTL is able to independently discover many of the language-specific regularities in the training data.
- Published
- 2009
- Full Text
- View/download PDF
280. On Finding Predictors for Arbitrary Families of Processes
- Author
-
Ryabko, Daniil, Sequential Learning (SEQUEL), Laboratoire d'Informatique Fondamentale de Lille (LIFL), Université de Lille, Sciences et Technologies-Institut National de Recherche en Informatique et en Automatique (Inria)-Université de Lille, Sciences Humaines et Sociales-Centre National de la Recherche Scientifique (CNRS)-Université de Lille, Sciences et Technologies-Institut National de Recherche en Informatique et en Automatique (Inria)-Université de Lille, Sciences Humaines et Sociales-Centre National de la Recherche Scientifique (CNRS)-Laboratoire d'Automatique, Génie Informatique et Signal (LAGIS), Université de Lille, Sciences et Technologies-Centrale Lille-Centre National de la Recherche Scientifique (CNRS)-Centrale Lille-Centre National de la Recherche Scientifique (CNRS)-Inria Lille - Nord Europe, Institut National de Recherche en Informatique et en Automatique (Inria), Université de Lille, Sciences et Technologies-Institut National de Recherche en Informatique et en Automatique (Inria)-Université de Lille, Sciences Humaines et Sociales-Centre National de la Recherche Scientifique (CNRS)-Université de Lille, Sciences et Technologies-Institut National de Recherche en Informatique et en Automatique (Inria)-Université de Lille, Sciences Humaines et Sociales-Centre National de la Recherche Scientifique (CNRS)-Inria Lille - Nord Europe, Institut National de Recherche en Informatique et en Automatique (Inria)-Laboratoire d'Automatique, Génie Informatique et Signal (LAGIS), and Université de Lille, Sciences et Technologies-Centrale Lille-Centre National de la Recherche Scientifique (CNRS)-Centrale Lille-Centre National de la Recherche Scientifique (CNRS)
- Subjects
FOS: Computer and information sciences ,Time series ,Bayesian prediction ,Computer Science - Artificial Intelligence ,Computer Science - Information Theory ,Information Theory (cs.IT) ,[MATH.MATH-IT]Mathematics [math]/Information Theory [math.IT] ,Mathematics - Statistics Theory ,[STAT.TH]Statistics [stat]/Statistics Theory [stat.TH] ,Statistics Theory (math.ST) ,[INFO.INFO-AI]Computer Science [cs]/Artificial Intelligence [cs.AI] ,Machine Learning (cs.LG) ,Computer Science - Learning ,Online prediction ,Artificial Intelligence (cs.AI) ,[INFO.INFO-LG]Computer Science [cs]/Machine Learning [cs.LG] ,[INFO.INFO-IT]Computer Science [cs]/Information Theory [cs.IT] ,[MATH.MATH-ST]Mathematics [math]/Statistics [math.ST] ,FOS: Mathematics ,Sequence prediction ,ACM: I.: Computing Methodologies/I.2: ARTIFICIAL INTELLIGENCE/I.2.6: Learning - Abstract
International audience; The problem is sequence prediction in the following setting. A sequence $x_1,\dots,x_n,\dots$ of discrete-valued observations is generated according to some unknown probabilistic law (measure) $\mu$. After observing each outcome, it is required to give the conditional probabilities of the next observation. The measure $\mu$ belongs to an arbitrary but known class $C$ of stochastic process measures. We are interested in predictors $\rho$ whose conditional probabilities converge (in some sense) to the ``true'' $\mu$-conditional probabilities if any $\mu\in C$ is chosen to generate the sequence. The contribution of this work is in characterizing the families $C$ for which such predictors exist, and in providing a specific and simple form in which to look for a solution. We show that if any predictor works, then there exists a Bayesian predictor, whose prior is discrete, and which works too. We also find several sufficient and necessary conditions for the existence of a predictor, in terms of topological characterizations of the family $C$, as well as in terms of local behaviour of the measures in $C$, which in some cases lead to procedures for constructing such predictors. It should be emphasized that the framework is completely general: the stochastic processes considered are not required to be i.i.d., stationary, or to belong to any parametric or countable family.
- Published
- 2009
- Full Text
- View/download PDF
281. The Follow Perturbed Leader Algorithm Protected from Unbounded One-Step Losses
- Author
-
Vladimir V. V'yugin
- Subjects
Mathematical optimization ,Sequence prediction ,Expert advice ,Volume (computing) ,Algorithm ,Mathematics ,Zero (linguistics) - Abstract
In this paper the sequential prediction problem with expert advice is considered for the case when the losses of experts suffered at each step can be unbounded. We present some modification of Kalai and Vempala algorithm of following the perturbed leader where weights depend on past losses of the experts. New notions of a volume and a scaled fluctuation of a game are introduced. We present an algorithm protected from unrestrictedly large one-step losses. This algorithm has the optimal performance in the case when the scaled fluctuations of onestep losses of experts of the pool tend to zero.
- Published
- 2009
- Full Text
- View/download PDF
282. A Bayesian Approach for Identifying miRNA Targets by Combining Sequence Prediction and Expression Profiling
- Author
-
Hui Liu, Lin Zhang, Yufei Huang, and Shou-Jiang Gao
- Subjects
Bayesian probability ,Genomics ,Computational biology ,Biology ,Proteomics ,computer.software_genre ,Article ,Gene expression profiling ,Sequence prediction ,microRNA ,Identification (biology) ,Data mining ,computer ,Sequence (medicine) - Abstract
MicroRNAs (miRNAs) are single-stranded non-coding RNAs shown to plays important regulatory roles in a wide range of biological processes and diseases. The functions and regulatory mechanisms of most of miRNAs are still poorly understood in part because of the difficulty in identifying the miRNA regulatory targets. To this end, computational methods have evolved as important tools for genome-wide target screening. Although considerable work in the past few years has produced many target prediction algorithms, most of them are solely based on sequence, and their accuracy is still poor. In contrast, gene expression profiling from miRNA over-expression experiments can provide additional information about miRNA targets. This paper presents a Bayesian approach to integrate sequence level prediction result with expression profiling to improve the performance of miRNA target identification. The test on proteomic and IP pull-down data demonstrated better performance of the proposed approach.
- Published
- 2009
- Full Text
- View/download PDF
283. Algorithmic Probability: Theory and Applications
- Author
-
Ray J. Solomonoff
- Subjects
Bernoulli's principle ,Theoretical computer science ,Grammar ,Sequence prediction ,media_common.quotation_subject ,General problem ,Algorithmic probability ,Inductive reasoning ,Completeness (statistics) ,media_common ,Mathematics - Abstract
We flrst deflne Algorithmic Probability, an extremely powerful method of inductive inference. We discuss its completeness, incomputability, diversity and subjectivity and show that its incomputability in no way inhibits its use for practical prediction. Applications to Bernoulli sequence prediction and grammar discovery are described. We conclude with a note on its employment in a very strong AI system for very general problem solving.
- Published
- 2008
- Full Text
- View/download PDF
284. Some sufficient conditions on an arbitrary class of stochastic processes for the existence of a predictor
- Author
-
Daniil Ryabko, Sequential Learning (SEQUEL), Laboratoire d'Informatique Fondamentale de Lille (LIFL), Université de Lille, Sciences et Technologies-Institut National de Recherche en Informatique et en Automatique (Inria)-Université de Lille, Sciences Humaines et Sociales-Centre National de la Recherche Scientifique (CNRS)-Université de Lille, Sciences et Technologies-Institut National de Recherche en Informatique et en Automatique (Inria)-Université de Lille, Sciences Humaines et Sociales-Centre National de la Recherche Scientifique (CNRS)-Laboratoire d'Automatique, Génie Informatique et Signal (LAGIS), Université de Lille, Sciences et Technologies-Centrale Lille-Centre National de la Recherche Scientifique (CNRS)-Centrale Lille-Centre National de la Recherche Scientifique (CNRS)-Inria Lille - Nord Europe, Institut National de Recherche en Informatique et en Automatique (Inria), Freund, Y., Györfi, L., Turán, G., Zeugmann, Th., Université de Lille, Sciences et Technologies-Institut National de Recherche en Informatique et en Automatique (Inria)-Université de Lille, Sciences Humaines et Sociales-Centre National de la Recherche Scientifique (CNRS)-Université de Lille, Sciences et Technologies-Institut National de Recherche en Informatique et en Automatique (Inria)-Université de Lille, Sciences Humaines et Sociales-Centre National de la Recherche Scientifique (CNRS)-Inria Lille - Nord Europe, Institut National de Recherche en Informatique et en Automatique (Inria)-Laboratoire d'Automatique, Génie Informatique et Signal (LAGIS), and Université de Lille, Sciences et Technologies-Centrale Lille-Centre National de la Recherche Scientifique (CNRS)-Centrale Lille-Centre National de la Recherche Scientifique (CNRS)
- Subjects
Discrete mathematics ,Class (set theory) ,non-stationary processes ,Logarithm ,Stochastic process ,05 social sciences ,Conditional probability ,020206 networking & telecommunications ,02 engineering and technology ,[STAT.TH]Statistics [stat]/Statistics Theory [stat.TH] ,Term (time) ,[INFO.INFO-LG]Computer Science [cs]/Machine Learning [cs.LG] ,[MATH.MATH-ST]Mathematics [math]/Statistics [math.ST] ,0502 economics and business ,Convergence (routing) ,0202 electrical engineering, electronic engineering, information engineering ,Applied mathematics ,Sequence prediction ,Parametric family ,050205 econometrics ,Mathematics ,Probability measure - Abstract
We consider the problem of sequence prediction in a probabilistic setting. Let there be given a class $\mathcal C$ of stochastic processes (probability measures on the set of one-way infinite sequences). We are interested in the question of what are the conditions on $\mathcal C$ under which there exists a predictor (also a stochastic process) for which the predicted probabilities converge to the correct ones if any of the processes in $\mathcal C$ is chosen to generate the data. We find some sufficient conditions on $\mathcal C$ under which such a predictor exists. Some of the conditions are asymptotic in nature, while others are based on the local (truncated to first observations) behaviour of the processes. The conditions lead to constructions of the predictors. In some cases we obtain rates of convergence that are optimal up to an additive logarithmic term. We emphasize that the framework is completely general: the stochastic processes considered are not required to be i.i.d., stationary, or to belong to some parametric family.
- Published
- 2008
- Full Text
- View/download PDF
285. Predicting Non-Stationary Processes
- Author
-
Daniil Ryabko, Marcus Hutter, Sequential Learning (SEQUEL), Laboratoire d'Informatique Fondamentale de Lille (LIFL), Université de Lille, Sciences et Technologies-Institut National de Recherche en Informatique et en Automatique (Inria)-Université de Lille, Sciences Humaines et Sociales-Centre National de la Recherche Scientifique (CNRS)-Université de Lille, Sciences et Technologies-Institut National de Recherche en Informatique et en Automatique (Inria)-Université de Lille, Sciences Humaines et Sociales-Centre National de la Recherche Scientifique (CNRS)-Laboratoire d'Automatique, Génie Informatique et Signal (LAGIS), Université de Lille, Sciences et Technologies-Centrale Lille-Centre National de la Recherche Scientifique (CNRS)-Centrale Lille-Centre National de la Recherche Scientifique (CNRS)-Inria Lille - Nord Europe, Institut National de Recherche en Informatique et en Automatique (Inria), Australian National University - Department of engineering (ANU), Australian National University (ANU), This research was supported by the Swiss NSF grants 200020-107616 and 200021-113364., Université de Lille, Sciences et Technologies-Institut National de Recherche en Informatique et en Automatique (Inria)-Université de Lille, Sciences Humaines et Sociales-Centre National de la Recherche Scientifique (CNRS)-Université de Lille, Sciences et Technologies-Institut National de Recherche en Informatique et en Automatique (Inria)-Université de Lille, Sciences Humaines et Sociales-Centre National de la Recherche Scientifique (CNRS)-Inria Lille - Nord Europe, Institut National de Recherche en Informatique et en Automatique (Inria)-Laboratoire d'Automatique, Génie Informatique et Signal (LAGIS), and Université de Lille, Sciences et Technologies-Centrale Lille-Centre National de la Recherche Scientifique (CNRS)-Centrale Lille-Centre National de la Recherche Scientifique (CNRS)
- Subjects
Mathematical optimization ,Class (set theory) ,02 engineering and technology ,01 natural sciences ,Measure (mathematics) ,010104 statistics & probability ,Local absolute continuity ,Absolute/KL divergence ,Mixtures of measures ,[INFO.INFO-LG]Computer Science [cs]/Machine Learning [cs.LG] ,Non-stationary measures ,0202 electrical engineering, electronic engineering, information engineering ,0101 mathematics ,Average/expected criteria ,Probability measure ,Mathematics ,Sequence ,Applied Mathematics ,Conditional probability ,[MATH.MATH-IT]Mathematics [math]/Information Theory [math.IT] ,020206 networking & telecommunications ,Absolute continuity ,[INFO.INFO-IT]Computer Science [cs]/Information Theory [cs.IT] ,Probability distribution ,Sequence prediction ,Exponentially equivalent measures - Abstract
International audience; Suppose we are given two probability measures on the set of one-way infinite finite-alphabet sequences. Consider the question of when one of the measures predicts the other, that is, when conditional probabilities converge (in a certain sense), if one of the measures is chosen to generate the sequence. This question may be considered a refinement of the problem of sequence prediction in its most general formulation: for a given class of probability measures, does there exist a measure which predicts all of the measures in the class? To address this problem, we find some conditions on local absolute continuity which are sufficient for prediction and generalize several different notions that are known to be sufficient for prediction. We also formulate some open questions to outline a direction for finding the conditions on classes of measures for which prediction is possible.
- Published
- 2008
- Full Text
- View/download PDF
286. Combining sequence-based prediction methods and circular dichroism and infrared spectroscopic data to improve protein secondary structure determinations
- Author
-
Jonathan G. Lees and Robert W. Janes
- Subjects
Models, Molecular ,Circular dichroism ,Materials science ,Spectrophotometry, Infrared ,Infrared ,Molecular Sequence Data ,lcsh:Computer applications to medicine. Medical informatics ,Bioinformatics ,Protein secondary structure prediction ,Peptide Mapping ,Biochemistry ,Protein Structure, Secondary ,Sequence Analysis, Protein ,Structural Biology ,Sequence prediction ,Prediction methods ,Computer Simulation ,Amino Acid Sequence ,lcsh:QH301-705.5 ,Molecular Biology ,Protein secondary structure ,Peptide sequence ,Sequence (medicine) ,Circular Dichroism ,Applied Mathematics ,Proteins ,Computer Science Applications ,Systems Integration ,Models, Chemical ,lcsh:Biology (General) ,lcsh:R858-859.7 ,Biological system ,Algorithms ,Research Article - Abstract
Background A number of sequence-based methods exist for protein secondary structure prediction. Protein secondary structures can also be determined experimentally from circular dichroism, and infrared spectroscopic data using empirical analysis methods. It has been proposed that comparable accuracy can be obtained from sequence-based predictions as from these biophysical measurements. Here we have examined the secondary structure determination accuracies of sequence prediction methods with the empirically determined values from the spectroscopic data on datasets of proteins for which both crystal structures and spectroscopic data are available. Results In this study we show that the sequence prediction methods have accuracies nearly comparable to those of spectroscopic methods. However, we also demonstrate that combining the spectroscopic and sequences techniques produces significant overall improvements in secondary structure determinations. In addition, combining the extra information content available from synchrotron radiation circular dichroism data with sequence methods also shows improvements. Conclusion Combining sequence prediction with experimentally determined spectroscopic methods for protein secondary structure content significantly enhances the accuracy of the overall results obtained.
- Published
- 2008
- Full Text
- View/download PDF
287. One-sequence and two-sequence prediction for future Weibull records
- Author
-
Omar M. Bdair and Mohammad Z. Raqab
- Subjects
Statistics and Probability ,Applied Mathematics ,010103 numerical & computational mathematics ,01 natural sciences ,Computer Science Applications ,010104 statistics & probability ,Weibull distribution ,record values ,Bayes estimation ,Bayes prediction ,Monte Carlo samples ,Sequence prediction ,Statistics ,lcsh:Probabilities. Mathematical statistics ,0101 mathematics ,lcsh:QA273-280 ,Mathematics ,Sequence (medicine) - Abstract
Based on record data, prediction of the future records from the two-parameter Weibull distribution is studied. First we consider the sampling based procedure to compute the Bayes estimates and also to construct symmetric credible intervals. Secondly, we consider one-sequence and two-sequence Bayes prediction of the future records based on some observed records. The Monte Carlo algorithms are used to compute simulation consistent predictors and prediction intervals for future unobserved records. A numerical simulation study is conducted to compare the different methods and a real data set involving the annual rainfall recorded at Los Angeles Civic Center during 132 years is analyzed to illustrate the procedures developed here.
- Published
- 2016
- Full Text
- View/download PDF
288. On Universal Prediction and Bayesian Confirmation
- Author
-
Marcus Hutter
- Subjects
FOS: Computer and information sciences ,Bayes ,General Computer Science ,Occam’s razor ,Computer science ,Model classes ,Computer Science - Information Theory ,Bayesian probability ,Kolmogorov complexity ,Mathematics - Statistics Theory ,Machine Learning (stat.ML) ,Statistics Theory (math.ST) ,Solomonoff prior ,Theoretical Computer Science ,Machine Learning (cs.LG) ,Reparametrization invariance ,Bayes' theorem ,Statistics - Machine Learning ,FOS: Mathematics ,Symmetry principle ,Invariant (mathematics) ,Statistical hypothesis testing ,Mathematical model ,Estimation theory ,Information Theory (cs.IT) ,Confirmation theory ,Philosophical issues ,Contrast (statistics) ,Black raven paradox ,Inductive reasoning ,Solomonoff's theory of inductive inference ,(non)Computable environments ,Old-evidence/updating problem ,Computer Science - Learning ,Sequence prediction ,Prediction bounds ,Algorithm ,Computer Science(all) - Abstract
The Bayesian framework is a well-studied and successful framework for inductive reasoning, which includes hypothesis testing and confirmation, parameter estimation, sequence prediction, classification, and regression. But standard statistical guidelines for choosing the model class and prior are not always available or fail, in particular in complex situations. Solomonoff completed the Bayesian framework by providing a rigorous, unique, formal, and universal choice for the model class and the prior. We discuss in breadth how and in which sense universal (non-i.i.d.) sequence prediction solves various (philosophical) problems of traditional Bayesian sequence prediction. We show that Solomonoff's model possesses many desirable properties: Strong total and weak instantaneous bounds, and in contrast to most classical continuous prior densities has no zero p(oste)rior problem, i.e. can confirm universal hypotheses, is reparametrization and regrouping invariant, and avoids the old-evidence and updating problem. It even performs well (actually better) in non-computable environments., Comment: 24 pages
- Published
- 2007
- Full Text
- View/download PDF
289. Development of sequential prediction system for Large scale database-based online modeling
- Author
-
Harutoshi Ogai, Yichun Yeh, Ogawa Masatoshi, and Kenko Uchida
- Subjects
Development (topology) ,Database ,Scale (ratio) ,Industrial reactor ,Computer science ,Sequence prediction ,Process (computing) ,Data mining ,computer.software_genre ,computer - Abstract
This paper reports a sequential prediction system of "large scale database-based online modeling (LOM)". The sequential prediction system predicts time-series process variables repeating processing that predicts process variables of next step by using the predicted process variables of previous step and prepared manipulated variables. Furthermore, the system is applied to the industrial reactor; practical effectiveness of the system is verified. As the result, the system has predicted the process variables with satisfactory accuracy. The practical effectiveness has been confirmed.
- Published
- 2007
- Full Text
- View/download PDF
290. Strategies for Prediction Under Imperfect Monitoring
- Author
-
Gilles Stoltz, Gábor Lugosi, and Shie Mannor
- Subjects
Mathematical optimization ,Logarithm ,Constructive proof ,010102 general mathematics ,SIGNAL (programming language) ,01 natural sciences ,010104 statistics & probability ,Consistency (statistics) ,Sequence prediction ,Simple (abstract algebra) ,Convergence (routing) ,Imperfect ,0101 mathematics ,Mathematics - Abstract
We propose simple randomized strategies for sequential prediction under imperfect monitoring, that is, when the forecaster does not have access to the past outcomes but rather to a feedback signal. The proposed strategies are consistent in the sense that they achieve, asymptotically, the best possible average reward. It was Rustichini [11] who first proved the existence of such consistent predictors. The forecasters presented here offer the first constructive proof of consistency. Moreover, the proposed algorithms are computationally efficient. We also establish upper bounds for the rates of convergence. In the case of deterministic feedback, these rates are optimal up to logarithmic terms.
- Published
- 2007
- Full Text
- View/download PDF
291. A philosophical treatise of universal induction
- Author
-
Rathmanner, Samuel, Hutter, Marcus, Rathmanner, Samuel, and Hutter, Marcus
- Abstract
Understanding inductive reasoning is a problem that has engaged mankind for thousands of years. This problem is relevant to a wide range of fields and is integral to the philosophy of science. It has been tackled by many great minds ranging from philosophers to scientists to mathematicians, and more recently computer scientists. In this article we argue the case for Solomonoff Induction, a formal inductive framework which combines algorithmic information theory with the Bayesian framework. Although it achieves excellent theoretical results and is based on solid philosophical foundations, the requisite technical knowledge necessary for understanding this framework has caused it to remain largely unknown and unappreciated in the wider scientific community. The main contribution of this article is to convey Solomonoff induction and its related concepts in a generally accessible form with the aim of bridging this current technical gap. In the process we examine the major historical contributions that have led to the formulation of Solomonoff Induction as well as criticisms of Solomonoff and induction in general. In particular we examine how Solomonoff induction addresses many issues that have plagued other inductive systems, such as the black ravens paradox and the confirmation problem, and compare this approach with other recent approaches.
- Published
- 2011
292. Hannan Consistency in On-Line Learning in Case of Unbounded Losses Under Partial Monitoring
- Author
-
György Ottucsák, Chamy Allenberg, Peter Auer, and László Györfi
- Subjects
Class (set theory) ,Mathematical optimization ,Square root ,Computer science ,Sequence prediction ,Consistency (statistics) ,business.industry ,Expert advice ,Bounded function ,Line (geometry) ,Regret ,Artificial intelligence ,business - Abstract
In this paper the sequential prediction problem with expert advice is considered when the loss is unbounded under partial monitoring scenarios. We deal with a wide class of the partial monitoring problems: the combination of the label efficient and multi-armed bandit problem, that is, where the algorithm is only informed about the performance of the chosen expert with probability e≤1. For bounded losses an algorithm is given whose expected regret scales with the square root of the loss of the best expert. For unbounded losses we prove that Hannan consistency can be achieved, depending on the growth rate of the average squared losses of the experts.
- Published
- 2006
- Full Text
- View/download PDF
293. Sequence prediction for non-stationary processes
- Author
-
Ryabko, Daniil and Hutter, Marcus
- Subjects
Sequence prediction ,probability forecasting ,local absolute continuity - Abstract
We address the problem of sequence prediction for nonstationary stochastic processes. In particular, given two measures on the set of one-way infinite sequences over a finite alphabet, consider the question whether one of the measures predicts the other. We find some conditions on local absolute continuity under which prediction is possible.
- Published
- 2006
- Full Text
- View/download PDF
294. Error in Enumerable Sequence Prediction
- Author
-
Hay, Nick
- Subjects
Solomonoff induction ,Sequence prediction ,enumerable semimeasures - Abstract
We outline a method for quantifying the error of a sequence prediction. With sequence predictions represented by semimeasures $ u(x)$ we define their error to be $-log_2 u(x)$. We note that enumerable semimeasures are those which model the sequence as the output of a computable system given unknown input. Using this we define the simulation complexity of a computable system $C$ relative to another $U$ giving an emph{exact} bound on their difference in error. This error in turn gives an exact upper bound on the number of predictions $ u$ gets incorrect.
- Published
- 2006
- Full Text
- View/download PDF
295. Is There an Elegant Universal Theory of Prediction?
- Author
-
Shane Legg
- Subjects
Prediction algorithms ,Development (topology) ,Kolmogorov complexity ,Inductive logic programming ,Computer science ,Sequence prediction ,business.industry ,Intelligent decision support system ,Universal theory ,Artificial intelligence ,Gödel's incompleteness theorems ,business - Abstract
Solomonoff's inductive learning model is a powerful, universal and highly elegant theory of sequence prediction. Its critical flaw is that it is incomputable and thus cannot be used in practice. It is sometimes suggested that it may still be useful to help guide the development of very general and powerful theories of prediction which are computable. In this paper it is shown that although powerful algorithms exist, they are necessarily highly complex. This alone makes their theoretical analysis problematic, however it is further shown that beyond a moderate level of complexity the analysis runs into the deeper problem of Godel incompleteness. This limits the power of mathematics to analyse and study prediction algorithms, and indeed intelligent systems in general.
- Published
- 2006
- Full Text
- View/download PDF
296. Discrete MDL predicts in total variation
- Author
-
Hutter, Marcus and Hutter, Marcus
- Abstract
The Minimum Description Length (MDL) principle selects the model that has the shortest code for data plus model. We show that for a countable class of models, MDL predictions are close to the true distribution in a strong sense. The result is completely general. No independence, ergodicity, stationarity, identifiability, or other assumption on the model class need to be made. More formally, we show that for any countable class of models, the distributions selected by MDL (or MAP) asymptotically predict (merge with) the true measure in the class in total variation distance. Implications for non-i.i.d. domains like time-series forecasting, discriminative learning, and reinforcement learning are discussed.
- Published
- 2009
297. Application of Markovian principles in high frequency sequence prediction from log motif patterns in the Agbada formation, Niger Delta, Nigeria
- Author
-
SO Olabode and JA Adekoye
- Subjects
Niger delta ,Discrete mathematics ,symbols.namesake ,Geophysics ,Sequence prediction ,Statistics ,symbols ,Markov process ,Economic Geology ,Geology ,Motif (music) ,Water Science and Technology - Abstract
No Abstract.Journal of Mining and Geology 2005, Vol. 41(1): 57-79
- Published
- 2005
298. Rapid on-line temporal sequence prediction by an adaptive agent
- Author
-
Daniel Boley, Paul Schrater, Steven Jensen, and Maria Gini
- Subjects
Intelligent agent ,n-gram ,Sequence prediction ,business.industry ,Computer science ,Entropy (information theory) ,Markov decision process ,Artificial intelligence ,Machine learning ,computer.software_genre ,business ,computer - Abstract
Robust sequence prediction is an essential component of an intelligent agent acting in a dynamic world. We consider the case of near-future event prediction by an online learning agent operating in a non-stationary environment. The challenge for a learning agent under these conditions is to exploit the relevant experience from a limited environmental event history while preserving flexibility.We propose a novel time/space efficient method for learning temporal sequences and making short-term predictions. Our method operates on-line, requires few exemplars, and adapts easily and quickly to changes in the underlying stochastic world model. Using a short-term memory of recent observations, the method maintains a dynamic space of candidate hypotheses in which the growth of the space is systematically and dynamically pruned using an entropy measure over the observed predictive quality of each candidate hypothesis.The method compares well against Markov-chain predictions, and adapts faster than learned Markov-chain models to changes in the underlying distribution of events. We demonstrate the method using both synthetic data and empirical experience from a game-playing scenario with human opponents.
- Published
- 2005
- Full Text
- View/download PDF
299. Universal Sequence Prediction
- Author
-
Marcus Hutter
- Subjects
Computer science ,Sequence prediction ,Algorithm - Published
- 2005
- Full Text
- View/download PDF
300. Biological Sequence Prediction using General Fuzzy Automata
- Author
-
S.C. Kremer and M. Doostfatemeh
- Subjects
Fuzzy automata ,Formalism (philosophy of mathematics) ,Theoretical computer science ,Sequence analysis ,Sequence prediction ,Fuzzy set ,A protein ,Hidden Markov model ,Automaton ,Mathematics - Abstract
This paper shows how the newly developed paradigm of General Fuzzy Automata (GFA) can be used as a biological sequence predictor. We consider the positional correlations of amino acids in a protein family as the basic criteria for prediction and classification of unknown sequences. It will be shown how the GFA formalism can be used as an efficient tool for classification of protein sequences. The results show that this approach predicts the membership of an unknown sequence in a protein family better than profile Hidden Markov Models (HMMs) which are now a popular and putative approach in biological sequence analysis.
- Published
- 2005
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.