20 results on '"Stéphane Marchand"'
Search Results
2. Revised Mitigation of Systematic Errors in SMOS Sea Surface Salinity
- Author
-
Nicolas Kolodziejczyk, Nicolas Reul, Jacqueline Boutin, Stéphane Marchand, and Jean-Luc Vergely
- Subjects
Systematic error ,010504 meteorology & atmospheric sciences ,0211 other engineering and technologies ,Mesoscale meteorology ,02 engineering and technology ,01 natural sciences ,SSS ,Salinity ,Fresh water ,Climatology ,Environmental science ,Satellite ,Sea surface salinity ,Water content ,021101 geological & geomatics engineering ,0105 earth and related environmental sciences - Abstract
An important contribution of satellite Sea Surface Salinity (SSS) is the spatio-temporal monitoring of rivers fresh water plumes at mesoscale. In this paper, we detail a new correction for systematic errors in the Soil Moisture and Ocean Salinity (SMOS) measurements that is implemented in the Centre Aval de Traitement des Donnees SMOS (CATDS). With this new mitigation, the SMOS and Soil Moisture Active Passive (SMAP) SSS monitor very consistent features in most areas close to continents. The rms-difference between bi-weekly SMOS and SMAP SSS over 20 months and in selected coastal regions is about 0.3pss (once outliers are filtered out), rather consistent with the rms-difference between satellite and in situ SSS (on the order of 0.2pss). The coefficient of determination (r2) between SMOS and SMAP SSS is above than 0.8 in very fresh areas (river plumes). Over the open ocean, the rms difference between SMOS and ship SSS is 0.2pss.
- Published
- 2018
- Full Text
- View/download PDF
3. Indexing Mayan hieroglyphs with neural codes
- Author
-
Stéphane Marchand-Maillet and Edgar Roman-Rangel
- Subjects
Artificial neural network ,Computer science ,Binary image ,Search engine indexing ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,020206 networking & telecommunications ,02 engineering and technology ,Function (mathematics) ,computer.software_genre ,Visualization ,Computer Science::Computer Vision and Pattern Recognition ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Data mining ,computer ,Curse of dimensionality - Abstract
We present an approach for unsupervised computation of local shape descriptors, which relies on the use of linear autoencoders for characterizing local regions of complex shapes. The proposed approach responds to the need for a robust scheme to index binary images using local descriptors, which arises when only few examples of the complete images are available for training, thus making inaccurate the learning process of parameters of traditional neural networks schemes. Given the possibility of using linear operations during the encoding phase, the computation of the proposed local descriptor can be fast once the parameters of the encoding function are learned. After conducting a vast search, we identified the optimal dimensionality of the resulting local descriptor to be of only 128 dimensions, which allows for efficient further operations on them, such as the construction of bag representations with purposes of shape retrieval and classification. We evaluated the proposed approach indexing a collection of complex binary images, whose instances contain compounds of hieroglyphs from the ancient Maya civilization. Our retrieval experiments show that the proposed approach achieves competitive retrieval performance when compared with hand-crafted local descriptors.
- Published
- 2016
- Full Text
- View/download PDF
4. Permutation based indexing for high dimensional data on GPU architectures
- Author
-
Martin Kruliš, Stéphane Marchand-Maillet, and Hasmik Osipyan
- Subjects
Instruction set ,Clustering high-dimensional data ,Memory hierarchy ,Computer science ,Feature vector ,Search engine indexing ,Search problem ,Parallel computing ,General-purpose computing on graphics processing units ,Massively parallel - Abstract
Permutation-based indexing is one of the most popular techniques for the approximate nearest-neighbor search problem in high-dimensional spaces. Due to the exponential increase of multimedia data, the time required to index this data has become a serious constraint of the indexing techniques. One of the possible steps towards faster index construction is utilization of massively parallel platforms such as the GPGPU architectures. In this paper, we have analyzed the computational costs of individual steps of the permutation-based index construction in a high-dimensional feature space and proposed a hybrid solution, where computational power of GPU is utilized for distance computations whilst the host CPU performs the postprocessing and sorting steps. Despite the fact that computing the distances is a naturally data-parallel task, an efficient implementation is quite challenging due to various GPU limitations and complex memory hierarchy. We have tested possible approaches to work division and data caching to utilize the GPU to its best abilities. We summarize our empirical results and point out the optimal solution.
- Published
- 2015
- Full Text
- View/download PDF
5. Fast large-scale multimedia indexing and searching
- Author
-
Hasmik Osipyan, Stéphane Marchand-Maillet, and Hisham Mohamed
- Subjects
Permutation ,Theoretical computer science ,Search algorithm ,Computer science ,Encoding (memory) ,Search engine indexing ,Metric (mathematics) ,Multimedia database ,Table (database) ,Data structure - Abstract
Searching for digital images in large-scale multimedia database is a hard problem due to the rapid increase of the digital assets. Metric Permutation Table is an efficient data structure for large-scale multimedia indexing. This data structure is based on the Permutation-based indexing, that aims to predict the proximity between elements encoding their location with respect to their surrounding. The main constraint of the Metric Permutation Table is the indexing time. With the exponential increase of multimedia data, parallel computation is needed. Opening the GPUs to general purpose computation allows to perform parallel computation on a powerful platform. In this paper, we propose efficient indexing and searching algorithms for the Metric Permutation Table using GPU and multi-core CPU. We study the performance and efficiency of our algorithms on large-scale datasets of millions of images. Experimental results show a decrease of the indexing time while preserving the quality of the results.
- Published
- 2014
- Full Text
- View/download PDF
6. Rank Aggregation for QoS-Aware Web Service Selection and Composition
- Author
-
Stéphane Marchand-Maillet and Birgit Hofreiter
- Subjects
Information retrieval ,Database ,computer.internet_protocol ,Computer science ,Quality of service ,Service-oriented architecture ,computer.software_genre ,Sensor fusion ,Information fusion ,Empirical research ,Robustness (computer science) ,Data analysis ,Web service ,computer - Abstract
Service Composition is an essential part of Service Oriented Architectures (SOA). The number of available alternatives for web services offering similar functional characteristics opens opportunities to maximise compositions based on criteria embedded into the Quality of Service (QoS). These non-functional criteria may be numerous and diverse, and arise from various constraints, making composition an optimal information fusion problem. A number of solutions have been offered to resolve the selection problem for web service composition, essentially relying on exploiting and combining absolute performance values for measuring multi-dimensional QoS. In this paper, we investigate the use of Rank Aggregation methods for performing adequate information fusion for web service selection in view of composition. Our goal is to take advantage of these methods working on ranks rather than on the direct values for inferring global QoS. We formalise our approach and demonstrate its validity via an empirical study. We further evaluate our approach on publicly available real data gathered from actual web service characteristics.
- Published
- 2013
- Full Text
- View/download PDF
7. Bag-of-Visual-Phrases via Local Contexts
- Author
-
Stéphane Marchand-Maillet and Edgar Roman-Rangel
- Subjects
Series (mathematics) ,business.industry ,Computer science ,InformationSystems_INFORMATIONSTORAGEANDRETRIEVAL ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Representation (systemics) ,computer.software_genre ,Probabilistic description ,ComputingMethodologies_PATTERNRECOGNITION ,Image representation ,Statistical analysis ,Visual Word ,Artificial intelligence ,business ,computer ,Image retrieval ,Word (computer architecture) ,Natural language processing - Abstract
This paper extends the bag-of-visual-words representations to a bag-of-visual-phrases model. The introduced bag-of-visual-phrases representation is constructed upon a proposed method for probabilistic description of co-occurring visual words, which is adapted for each reference word. This bag-of-visual-phrases representation implicitly encodes spatial relationships among visual words, thus being a richer representation while remaining as compact as the bag-of-visual-words model. We demonstrate the effectiveness of our method with a series of statistical analysis and retrieval experiments, and show that it largely outperforms previous methods for construction of bag representations. Furthermore, our method allows to query traditional bag-of-words vs the proposed bag-of-phrases. We conducted retrieval experiments on a dataset of complex shapes, whose instances correspond to hieroglyphs of the pre-Columbian Maya culture from the ancient Americas.
- Published
- 2013
- Full Text
- View/download PDF
8. Multi-dimensional Information Ordering to Support Decision-Making Processes
- Author
-
Birgit Hofreiter and Stéphane Marchand-Maillet
- Subjects
Information management ,Visual analytics ,Computer science ,Digital data ,Scalability ,Information source ,ddc:025.063 ,Business case ,Data science ,Automatic summarization ,Visualization - Abstract
Massive amounts of textual and digital data are created daily from business or public activities. The organisation, mining and summarization of such a rich and large information source is required to capture the essential and critical knowledge it contains. Such a mining is of strategic importance in many domains including innovation (eg to mine technological reviews and scientific literature) and electronic commerce (eg to mine customer reviews). Information content generally bears several important aspects, mapped onto visualisation dimensions, whose number needs to be reduced to enable relevant interactive exploration. In this paper, we propose a novel strategy to mine and organise document sets, in order to present them in a consistent manner and to highlight interesting and relevant information patterns they contain. We base our method on the formulation of a global optimisation problem solved by using the Traveling Salesman Problem (TSP) approach. We show how this compact formulation opens interesting possibilities for the mining of document collections mapped onto multidimensional information sets. We discuss the issue of scalability and show that associated scalable solutions exist. We demonstrate the effectiveness of our method over several types of documents, embedded into real business cases.
- Published
- 2013
- Full Text
- View/download PDF
9. Distributed media indexing based on MPI and MapReduce
- Author
-
Stéphane Marchand-Maillet and Hisham Mohamed
- Subjects
Structure (mathematical logic) ,Theoretical computer science ,Speedup ,Computer Networks and Communications ,computer.internet_protocol ,Computer science ,Message passing ,Search engine indexing ,Message Passing Interface ,Parallel computing ,Hardware and Architecture ,Scalability ,Media Technology ,Programming paradigm ,ddc:025.063 ,computer ,Software ,XML ,Sequential algorithm - Abstract
Web-scale digital assets comprise millions or billions of documents. Due to such increase, sequential algorithms cannot cope with this data, and parallel and distributed computing become the solution of choice. MapReduce is a programming model proposed by Google for scalable data processing. MapReduce is mainly applicable for data intensive algorithms. In contrast, The message passing interface (MPI) is suitable for high performance algorithms. This paper proposes an adapted structure of MapReduce programming model using MPI for multimedia indexing. Experimental results on a large number of text (XML) excerpts related to images from the ImageNet corpus indicate that our implementation achieved good speedup compared to the sequential version and the earlier versions of MapReduce using MPI. Extensions to index large-scale multimedia collections are discussed.
- Published
- 2012
- Full Text
- View/download PDF
10. Enhancing MapReduce using MPI and an optimized data exchange policy
- Author
-
Stéphane Marchand-Maillet and Hisham Mohamed
- Subjects
Speedup ,Computer science ,Data exchange ,Pipeline (computing) ,Message passing ,Programming paradigm ,Message Passing Interface ,Parallel computing ,ddc:025.063 ,Data modeling - Abstract
MapReduce is a programming model proposed by Google to simplify large-scale data processing. In contrast, the message passing interface (MPI) standard is extensively used for algorithmic parallelization, as it accommodates an efficient communication infrastructure. In the original implementation of MapReduce, the reduce function can only start processing following termination of the map function. If the map function is slow for any reason, this will affect the whole running time. In this paper, we propose MapReduce overlapping using MPI, which is an adapted structure of the MapReduce programming model for fast intensive data processing. Our implementation is based on running the map and the reduce functions concurrently in parallel by exchanging partial intermediate data between them in a pipeline fashion using MPI. At the same time, we maintain the usability and the simplicity of MapReduce. Experimental results based on two different applications (Word Count and Distributed Inverted Indexing) show a good speedup compared to the earlier versions of MapReduce such as Hadoop and the available MPI-MapReduce implementations. For word count, we are able to achieve 1.9x and 5.3x speedup comparing to Hadoop and MPI-MapReduce respectively for 53Gb of data.
- Published
- 2012
11. SwiftLink: Serendipitous Navigation Strategy for Large-Scale Document Collections
- Author
-
Marc von Wyl and Stéphane Marchand-Maillet
- Subjects
Focus (computing) ,Information retrieval ,Computer science ,Search engine indexing ,Scalability ,Information processing ,Context (language use) ,Recommender system ,ddc:025.063 ,User requirements document ,PATH (variable) - Abstract
The multiplication of large-scale document collections has created the need for robust and adaptive access strategies in many applicative areas. In this paper, we depart from the traditional document search paradigm to move onto the construction of a collection navigation strategy. We thus detail a model where user clicks are taken as expression of interest rather than positive search feedback. In terms of the interpretation of user interaction, our model is close to the recommendation paradigm. However, it exploits user feedback in a navigational procedure to perform an approximation of interactive metric learning procedures. Our model is directly compatible with distributed data indexing and is therefore inherently built to cope with scalability issues. The evaluation of such models is discussed and implemented over the context of an image collection. Experiments are presented where the user browses a collection of images with changing center of focus in the course of the browsing operations. Results show that our model is effectively able to recommend the user a relevant path through the collection and to adapt according to variable expressed interests.
- Published
- 2012
12. Query log simulation for long-term learning in image retrieval
- Author
-
Donn Morrison, Eric Bruno, and Stéphane Marchand-Maillet
- Subjects
Web search query ,Information retrieval ,Computer science ,InformationSystems_INFORMATIONSTORAGEANDRETRIEVAL ,computer.software_genre ,Query language ,Query optimization ,Ranking (information retrieval) ,Query expansion ,Web query classification ,Sargable ,Data mining ,ddc:025.063 ,computer ,RDF query language ,computer.programming_language - Abstract
In this paper we formalise a query simulation framework for the evaluation of long-term learning systems for image retrieval. Long-term learning relies on historical queries and associated relevance judgements, usually stored in query logs, in order to improve search results presented to users of the retrieval system. Evaluation of long-term learning methods requires access to query logs, preferably in large quantity. However, real-world query logs are notoriously difficult to acquire due to legitimate efforts of safeguarding user privacy. Query log simulation provides a useful means of evaluating long-term learning approaches without the need for real-world data. We introduce a query log simulator that is based on a user model of long-term learning that explains the observed relevance judgements contained in query logs. We validate simulated queries against a real-world query log of an image retrieval system and demonstrate that for evaluation purposes, the simulator is accurate on a global level.
- Published
- 2011
- Full Text
- View/download PDF
13. Optimizing Strategies for the Exploration of Social-Networks and Associated Data Collections
- Author
-
Stéphane Marchand-Maillet, Eric Bruno, and Eniko Szekely
- Subjects
Information retrieval ,Document handling ,Social network ,Point (typography) ,business.industry ,Computer science ,Interface (Java) ,optimisation ,multimedia computing ,Context (language use) ,search engines ,World Wide Web ,Cultural heritage ,social networking (online) ,Lagrangian coherent structures ,information retrieval ,business ,• document handling ,Complement (set theory) - Abstract
Multimedia data collections immersed into social networks may be explored from the point of view of varying documents and users characteristics. In this paper, we develop a unified model to embed documents and users into coherent structures from which to extract optimal subsets. The result is the definition of guiding navigation strategies of both the user and document networks, as a complement to classical search operations. An initial interface that may materialize such browsing over documents is demonstrated in the context of Cultural Heritage.
- Published
- 2009
14. Exploiting document feature interactions for efficient information fusion in high dimensional spaces
- Author
-
Stéphane Marchand-Maillet, Jana Kludas, and Eric Bruno
- Subjects
business.industry ,Computer science ,Feature extraction ,Feature selection ,Mutual information ,Machine learning ,computer.software_genre ,Support vector machine ,Statistical classification ,Open research ,Feature (computer vision) ,Redundancy (engineering) ,Artificial intelligence ,business ,computer - Abstract
Information fusion, especially for high dimensional multimedia data, is still an open research problem. In this article, we present a new approach to target this problem. Feature information interaction is an information-theoretic dependence measure that can determine synergy and redundancy between attributes, which then can be exploited with feature selection and construction towards more efficient information fusion. This also leads to improved performances for algorithms that rely on information fusion like multimedia document classification. We show that synergetic and redundant feature pairs require different fusion strategies for optimal exploitation. The approach is compared to classical feature selection strategies based on correlation and mutual information.
- Published
- 2008
- Full Text
- View/download PDF
15. Semantic clustering of images using patterns of relevance feedback
- Author
-
D. Morrison, Stéphane Marchand-Maillet, and Eric Bruno
- Subjects
relevance feedback ,Information retrieval ,Probabilistic latent semantic analysis ,Computer science ,InformationSystems_INFORMATIONSTORAGEANDRETRIEVAL ,Relevance feedback ,Image clustering ,Semantic data model ,Semantic similarity ,Semantic equivalence ,Semantic computing ,Semantic technology ,latent semantic analysis ,Semantic integration ,longterm learning - Abstract
User-supplied data such as browsing logs, click-through data, and relevance feedback judgements are an important source of knowledge during semantic indexing of documents such as images and video. Low-level indexing and abstraction methods are limited in the manner with which semantic data can be dealt. In this paper and in the context of this semantic data, we apply latent semantic analysis on two forms of user-supplied data, real-world and artificially generated relevance feedback judgements in order to examine the validity of using artificially generated interaction data for the study of semantic image clustering.
- Published
- 2008
- Full Text
- View/download PDF
16. Countering the false positive projection effect in nonlinear asymmetric classification
- Author
-
Serhiy Kosinov, Stéphane Marchand-Maillet, and Thierry Pun
- Subjects
Nonlinear system ,Degree (graph theory) ,business.industry ,Pattern recognition ,Statistical analysis ,Artificial intelligence ,Extension (predicate logic) ,Kernel Fisher discriminant analysis ,Linear discriminant analysis ,business ,Projection (set theory) ,Mathematics - Abstract
This work concerns the problem of asymmetric classification and provides the following contributions. First, it introduces the method of KDDA - a kernelized extension of the distance-based discriminant analysis technique that treats data asymmetrically and naturally accommodates indefinite kernels. Second, it demonstrates that KDDA and other asymmetric nonlinear projective approaches, such as BiasMap and KFD are often prone to an adverse condition referred to as the false positive projection effect. Empirical evaluation on both synthetic and real-world data sets is carried out to assess the degree of performance degradation due to false positive projection effect, determine the viability of some schemes for its elimination, and compare the introduced KDDA method with state-of-the-art alternatives, achieving encouraging results
- Published
- 2006
- Full Text
- View/download PDF
17. Hierarchical ensemble learning for multimedia categorization and autoannotation
- Author
-
Stéphane Marchand-Maillet and S. Koisnov
- Subjects
Multimedia ,Computer science ,business.industry ,Contrast (statistics) ,Context (language use) ,Machine learning ,computer.software_genre ,Ensemble learning ,Set (abstract data type) ,Goodness of fit ,Categorization ,Relevance (information retrieval) ,Data mining ,Artificial intelligence ,business ,Baseline (configuration management) ,computer - Abstract
This paper presents a hierarchical ensemble learning method applied in the context of multimedia autoannotation. In contrast to the standard multiple-category classification setting that assumes independent, non-overlapping and exhaustive set of categories, the proposed approach models explicitly the hierarchical relationships among target classes and estimates their relevance to a query as a trade-off between the goodness of fit to a given category description and its inherent uncertainty. The promising results of the empirical evaluation confirm the viability of the proposed approach, validated in comparison to several techniques of ensemble learning, as well as with different type of baseline classifiers
- Published
- 2005
- Full Text
- View/download PDF
18. Visual Object Categorization using Distance-Based Discriminant Analysis
- Author
-
Stéphane Marchand-Maillet, Thierry Pun, and Serhiy Kosinov
- Subjects
Optimization problem ,Feature data ,Artificial neural network ,Computer science ,business.industry ,Dimensionality reduction ,Feature extraction ,Pattern recognition ,Linear discriminant analysis ,Boosting methods for object categorization ,Machine learning ,computer.software_genre ,Object detection ,Discriminative model ,Categorization ,Artificial intelligence ,business ,computer - Abstract
This paper formulates the problem of object categorization in the discriminant analysis framework focusing on transforming visual feature data so as to make it conform to the compactness hypothesis in order to improve categorization accuracy. The sought transformation, in turn, is found as a solution to an optimization problem formulated in terms of inter-observation distances only, using the technique of iterative majorization. The proposed approach is suitable for both binary and multiple-class categorization problems, and can be applied as a dimensionality reduction technique. In the latter case, the number of discriminative features is determined automatically since the process of feature extraction is fully embedded in the optimization procedure. Performance tests validate our method on a number of benchmark data sets from the UCI repository, while the experiments in the application of visual object and content-based image categorization demonstrate very competitive results, asserting the method's capability of producing semantically relevant matches that share the same or synonymous vocabulary with the query category and allowing multiple pertinent category assignment.
- Published
- 2005
- Full Text
- View/download PDF
19. Approximate Viterbi decoding for 2D-hidden Markov models
- Author
-
Bernard Merialdo, Stéphane Marchand-Maillet, and Benoit Huet
- Subjects
Iterative Viterbi decoding ,business.industry ,Computer science ,Computer Science::Computation and Language (Computational Linguistics and Natural Language and Speech Processing) ,Pattern recognition ,Sequential decoding ,Viterbi algorithm ,Markov model ,symbols.namesake ,Viterbi decoder ,symbols ,Artificial intelligence ,Forward algorithm ,business ,Hidden Markov model ,Algorithm ,Soft output Viterbi algorithm - Abstract
While one-dimensional hidden Markov models have been very successfully applied to numerous problems, their extension to two dimensions has been shown to be exponentially complex, and this has very much restricted their usage for problems such as image analysis. In this paper we propose a novel algorithm which is able to approximate the search for the best state path (Viterbi decoding) in a 2D HMM. This algorithm makes certain assumptions which lead to tractable computations, at a price of loss in full optimality. We detail our algorithm, its implementation, and present some experiments on handwritten character recognition. Because the Viterbi algorithm serves as a basis for many applications, and 1D HMMs have shown great flexibility in their usage, our approach has the potential to make 2D HMMs as useful for 2D data as 1D HMMs are for 1D data such as speech.
- Published
- 2002
- Full Text
- View/download PDF
20. 2019 International Conference on Content-Based Multimedia Indexing, CBMI 2019, Dublin, Ireland, September 4-6, 2019
- Author
-
Cathal Gurrin, Björn þór Jónsson 0001, Renaud Péteri, Stevan Rudinac, Stéphane Marchand-Maillet, Georges Quénot, Kevin McGuinness, Gylfi þór Guðmundsson, Suzanne Little, Marie Katsurai, and Graham Healy
- Published
- 2019
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.