80 results on '"Database Systems"'
Search Results
2. Validation tests of the W2020 energy levels of water vapor
- Author
-
Campargue, A., Mikhailenko, S.N., Kassi, S., Vasilchenko, S., LAsers, Molécules et Environnement (LAME-LIPhy ), Laboratoire Interdisciplinaire de Physique [Saint Martin d’Hères] (LIPhy ), and Centre National de la Recherche Scientifique (CNRS)-Université Grenoble Alpes (UGA)-Centre National de la Recherche Scientifique (CNRS)-Université Grenoble Alpes (UGA)
- Subjects
LIGHT MEASUREMENT ,rovibrational energy level ,SPECTRA'S ,absorption spectroscopy ,[PHYS.PHYS.PHYS-OPTICS]Physics [physics]/Physics [physics]/Optics [physics.optics] ,DATABASE ,WATER ABSORPTION ,WATER VAPOUR ,ABSORPTION SPECTROSCOPY ,ROVIBRATIONAL ENERGY LEVEL ,CAVITY RING DOWN SPECTROSCOPIES ,SPECTROSCOPIC DATABASE ,DATABASE SYSTEMS ,LINE LISTS ,ENERGY ,VALIDATION TEST ,RO-VIBRATIONAL ENERGIES ,water vapor ,WATER VAPOR ,H2O ,ATOMIC ABSORPTION SPECTROSCOPY ,spectroscopic database ,ComputingMilieux_MISCELLANEOUS ,MARVEL - Abstract
A decade ago, a task group of the International Union of Pure and Applied Chemistry performed an exhaustive collection and review of measured transitions, applied the MARVEL procedure, and derived recommended empirical energy levels for nine major water isotopologues. Very recently, using an improved methodology, the sets of empirical energy levels of H216O, H218O and H217O were updated, leading to the so-called W2020 energy levels and transition wavenumbers [Furtenbacher et al. J. Phys. Chem. Ref. Data 49 (2020) 043103; 10.1063/5.0030680]. Here we present validation tests of the W2020 line list of H216O against spectra recorded by cavity ring down spectroscopy (CRDS) referenced to a frequency comb (FC), newly obtained in the 8040-8630 cm−1 region. The recorded spectra are found in excellent agreement with previous high-quality studies available in the literature. While these literature sources were all incorporated in the transition database used to derive the W2020 energy levels, the direct superposition of the FC-CRDS spectra to the W2020 line list of H216O shows a number of large disagreements. Cases where deviations largely exceed the W2020 claimed uncertainty on the transition frequencies are noted. In the considered spectral region, the resulting W2020 list is thus less accurate than some of the published original sources used to derive the W2020 energy levels. We conclude that the sophisticated global procedure and algorithm elaborated to identify and adequately weight inaccurate line positions among the large W2020 transition database do not always prevent less accurate data from “spoiling” higher quality data sources. The W2020 list of H216O is also compared to newly recorded CRDS spectra in the 12970–13200 cm−1 region (corresponding to the region of the A-band of O2), where previous observations were very scarce. As in the previous region, substantial position deviations are evidenced, and in many cases, the W2020 error bars appear to be strongly underestimated. © 2021 Elsevier Ltd The support of the CNRS (France) in the frame of International Research Project SAMIA is acknowledged. SNM activity was also partly supported in the frame of the Russian Science Foundation, Grant No. 18-11-00024-Π. CRDS measurements near 760 nm were performed at IAO-Tomsk and funded by RFBR project 20-32-70054.
- Published
- 2021
- Full Text
- View/download PDF
3. SymbioConstruction: A Bibliography-Driven Dynamic Construction Industry Symbiosis Database
- Author
-
Olcay Genç, Mühendislik ve Doğa Bilimleri Fakültesi -- İnşaat Mühendisliği Bölümü, and Genç, Olcay
- Subjects
Optimization ,Resource (biology) ,Circular economy ,Parks ,Strategy and Management ,0211 other engineering and technologies ,02 engineering and technology ,010501 environmental sciences ,01 natural sciences ,Database ,Engineering ,IS implementation ,Eco-industrial Parks ,Sustainable development ,Industrial processs ,Industrial symbiosis ,Bibliography ,İnformation ,021108 energy ,Non-renewable resource ,Future ,Industrial Symbiosis ,0105 earth and related environmental sciences ,Civil and Structural Engineering ,Constriction industry ,Symbiotic relationship ,Building and Construction ,Environmental economics ,Construction industry ,Generalized method ,Eco-industrial park ,Sustainability ,Database systems ,Waste ,Methodological contributions ,Industrial relations ,Construction & Building Technology ,Business ,Networks ,Decision making ,Waste utilization ,Barriers ,Data streams - Abstract
The increasing depletion of nonrenewable resources has motivated society to find alternate sources and methods of industrial processes. This problem can only be solved by using the resources that are given by nature in such a way that they can be spontaneously renewed, namely through sustainability. Industrial symbiosis (IS), which is the substitution of raw materials used in an industrial process by another resource that would otherwise be discarded, is the key to sustainability in industry. IS is an alternative to a conventional linear economy (take-make-dispose) in which one keeps resources in use for as long as possible, extracts the maximum value from them during use, then reclaims and reproduces products and materials at the end of each service life. The initial step in developing a symbiotic relationship is to define possible relationships between production firms. Many studies have shown that lack of data on waste streams is one of the significant barriers to IS establishment. In this study, by screening IS case studies in the literature, a construction industry-related symbiosis database, namely SymbioConstruction, is developed as three modules in which the necessary data are provided to waste seekers to utilize/send a specific waste and to stakeholder seekers to establish symbiotic relationships by exchanging wastes. Then, the database is tested with two steps: (1) evaluation on the ecoindustrial parks, and (2) evaluation on the construction materials industry. The exploratory results of the evaluation show that the SymbioConstruction is able to identify the major amount of construction industry-related symbiotic relationships realized in ecoindustrial parks located around the world. Furthermore, it also shows that the main producers of construction materials have the opportunity of implementing symbiotic relationships in terms of sending and receiving wastes for utilization. The practical contribution of this study is in the form of a database in which the guidance is provided to practitioners for decision making in symbiotic relationship implementation with respect to the identification and selection of possible stakeholders. Two methodological contributions are also provided: (1) a generalized method for identifying sector-oriented potential IS implementations, and (2) a generalized method for storing those IS opportunities.
- Published
- 2021
- Full Text
- View/download PDF
4. Random Inspection Policies for a Database System.
- Author
-
Zhao, Xufeng, Qian, Cunhua, Nakamura, Syouji, and Nakagawa, Toshio
- Abstract
When failure occurs in a database, we execute the backup operations to the latest checking times. We formulate periodic and random inspection policies for a database. When the system is checked at the N th interval of working times, two total expected costs are obtained and optimal numbers N* which minimize them are derived analytically, when failure and random times are exponential. One modified backup model in which failures are detected only at checking times is proposed and two expected costs are obtained. [ABSTRACT FROM PUBLISHER]
- Published
- 2012
- Full Text
- View/download PDF
5. Overlay Indexes: Efficiently Supporting Aggregate Range Queries and Authenticated Data Structures in Off-the-Shelf Databases
- Author
-
Maurizio Pizzonia, Alessio Papi, Diego Pennino, Pennino, D., Pizzonia, M., and Papi, A.
- Subjects
FOS: Computer and information sciences ,Computer Science - Cryptography and Security ,computational efficiency ,General Computer Science ,Range query (data structures) ,Computer science ,Data security ,02 engineering and technology ,Overlay ,computer.software_genre ,01 natural sciences ,Computer Science - Databases ,020204 information systems ,Computer Science - Data Structures and Algorithms ,0202 electrical engineering, electronic engineering, information engineering ,General Materials Science ,Data Structures and Algorithms (cs.DS) ,0101 mathematics ,data security ,indexe ,Complement (set theory) ,Database system ,Database ,010102 general mathematics ,Aggregate (data warehouse) ,Search engine indexing ,General Engineering ,authenticated data structure ,Databases (cs.DB) ,Data structure ,indexes ,Index (publishing) ,Database systems ,tree data structures ,aggregated range queries ,lcsh:Electrical engineering. Electronics. Nuclear engineering ,computer ,Cryptography and Security (cs.CR) ,lcsh:TK1-9971 ,aggregated range querie - Abstract
Commercial off-the-shelf DataBase Management Systems (DBMSes) are highly optimized to process a wide range of queries by means of carefully designed indexing and query planning. However, many aggregate range queries are usually performed by DBMSes using sequential scans, and certain needs, like storing Authenticated Data Structures (ADS), are not supported at all. Theoretically, these needs could be efficiently fulfilled adopting specific kinds of indexing, which however are normally ruled-out in DBMSes design. We introduce the concept of overlay index: an index that is meant to be stored in a standard database, alongside regular data and managed by regular software, to complement DBMS capabilities. We show a data structure, that we call DB-tree, that realizes an overlay index to support a wide range of custom aggregate range queries as well as ADSes, efficiently. All DB-trees operations can be performed by executing a small number of queries to the DBMS, that can be issued in parallel in one or two query rounds, and involves a logarithmic amount of data. We experimentally evaluate the efficiency of DB-trees showing that our approach is effective, especially if data updates are limited.
- Published
- 2019
6. A Persistent Memory-Aware Buffer Pool Manager Simulator for Multi-Tenant Cloud Databases
- Author
-
Laurent d'Orazio, Taras Basiuk, Eleazar Leal, Le Gruenwald, Computer Science Department- University of Oklahoma, University of Oklahoma (OU), A Symbolic and Human-centric view of dAta MANagement (SHAMAN), GESTION DES DONNÉES ET DE LA CONNAISSANCE (IRISA-D7), Institut de Recherche en Informatique et Systèmes Aléatoires (IRISA), Université de Rennes (UR)-Institut National des Sciences Appliquées - Rennes (INSA Rennes), Institut National des Sciences Appliquées (INSA)-Institut National des Sciences Appliquées (INSA)-Université de Bretagne Sud (UBS)-École normale supérieure - Rennes (ENS Rennes)-Institut National de Recherche en Informatique et en Automatique (Inria)-CentraleSupélec-Centre National de la Recherche Scientifique (CNRS)-IMT Atlantique (IMT Atlantique), Institut Mines-Télécom [Paris] (IMT)-Institut Mines-Télécom [Paris] (IMT)-Université de Rennes (UR)-Institut National des Sciences Appliquées - Rennes (INSA Rennes), Institut Mines-Télécom [Paris] (IMT)-Institut Mines-Télécom [Paris] (IMT)-Institut de Recherche en Informatique et Systèmes Aléatoires (IRISA), Institut Mines-Télécom [Paris] (IMT)-Institut Mines-Télécom [Paris] (IMT), Université de Rennes 1 (UR1), Université de Rennes (UNIV-RENNES)-Université de Rennes (UNIV-RENNES)-Institut National des Sciences Appliquées - Rennes (INSA Rennes), Institut National des Sciences Appliquées (INSA)-Université de Rennes (UNIV-RENNES)-Institut National des Sciences Appliquées (INSA)-Université de Bretagne Sud (UBS)-École normale supérieure - Rennes (ENS Rennes)-Institut National de Recherche en Informatique et en Automatique (Inria)-CentraleSupélec-Centre National de la Recherche Scientifique (CNRS)-IMT Atlantique Bretagne-Pays de la Loire (IMT Atlantique), and Institut Mines-Télécom [Paris] (IMT)-Institut Mines-Télécom [Paris] (IMT)-Université de Rennes 1 (UR1)
- Subjects
050101 languages & linguistics ,nonvolatile memory ,Computer science ,Data management ,Cloud computing ,02 engineering and technology ,computer.software_genre ,Persistence (computer science) ,Service-level agreement ,Software ,0202 electrical engineering, electronic engineering, information engineering ,Cloud database ,0501 psychology and cognitive sciences ,Implementation ,Simulation ,database systems ,Multitenancy ,[INFO.INFO-DB]Computer Science [cs]/Databases [cs.DB] ,Database ,business.industry ,05 social sciences ,cloud computing ,simulation ,buffer storage ,Non-volatile memory ,020201 artificial intelligence & image processing ,business ,computer ,Data migration - Abstract
International audience; Non-Volatile Memory (NVM) is a promising development for Database Management Systems (DBMS), offering abundant and fast storage to complement traditional disk and main memory architectures. NVM introduces additional data migration possibilities to the traditional buffer pool (BP) managers used by the DBMS. Hence, the efficient use of this new technology requires a redesign of the BP manager. For the cloud Database-as-a-Service products, this need for a redesign is further complicated by the traditional cloud providers' goal to minimize the Service Level Agreement (SLA) violation penalties paid to their tenants. Unfortunately, current research in the area does not provide a comprehensive picture of the components constituting a multi-tenant persistent memory aware BP manager for a cloud database that makes use of NVM. Furthermore, researchers lack the software tools needed to quickly prototype and estimate the effectiveness of novel data management policies guiding those components. In this paper, we attempt to remedy both issues, first, by proposing a generalized framework that defines the purpose and the abstract interfaces of various multi-tenant persistent memory-aware BP manager components, and second, by developing and demonstrating a simulator algorithm that is shown to aid in quick testing of different implementations of those BP manager components.
- Published
- 2020
7. Exclusive Region Design for Spatial Grid-Based Spectrum Database: A Stochastic Geometry Approach
- Author
-
Masahiro Morikura, Shota Yamashita, Koji Yamamoto, and Takayuki Nishio
- Subjects
Optimization problem ,General Computer Science ,Database ,Computer science ,Cognitive radio ,General Engineering ,020206 networking & telecommunications ,02 engineering and technology ,Grid ,computer.software_genre ,wireless communication ,Reduction (complexity) ,spatial grid ,Transmission (telecommunications) ,0202 electrical engineering, electronic engineering, information engineering ,General Materials Science ,stochastic processes ,lcsh:Electrical engineering. Electronics. Nuclear engineering ,Stochastic geometry ,computer ,lcsh:TK1-9971 ,spectrum sharing ,database systems - Abstract
This paper presents the stochastic-geometry analysis and designs a primary exclusive region (PER) for a spatial grid-based spectrum database system. The purpose of the spatial grid is to utilize information, such as the primary receiver (PR) antenna pattern, and the secondary transmitter (ST) density and transmission power, in each divided region. This paper introduces polar and square grids. For these spatial grids, the cumulants of the aggregate interference at a PR, from the STs, are derived, where the probability generating functional for the Poisson point process (PPP) is used on the assumption that the distribution of the STs in each divided region follows an inhomogeneous PPP. In addition, by introducing the allowable transmission probability of STs in each divided region, the PER optimization problem can be formulated as a continuous optimization problem. Numerical results demonstrate that a PER, corresponding to the information, is successfully designed and that more detailed information, due to a reduction in the area of each divided region, leads to a smaller and complex-shaped PER.
- Published
- 2018
8. Consistent normalization approach for Life Cycle Assessment based on inventory databases
- Author
-
Matthias Finkbeiner, Antoine Esnouf, Arnaud Helias, Laboratoire de Biotechnologie de l'Environnement [Narbonne] (LBE), Institut national d’études supérieures agronomiques de Montpellier (Montpellier SupAgro), Institut national d'enseignement supérieur pour l'agriculture, l'alimentation et l'environnement (Institut Agro)-Institut national d'enseignement supérieur pour l'agriculture, l'alimentation et l'environnement (Institut Agro)-Institut National de Recherche pour l’Agriculture, l’Alimentation et l’Environnement (INRAE), Information – Technologies – Analyse Environnementale – Procédés Agricoles (UMR ITAP), and Technische Universität Berlin (TU)
- Subjects
Normalization (statistics) ,Environmental Engineering ,010504 meteorology & atmospheric sciences ,Computer science ,BOTTOM UP APPROACH ,010501 environmental sciences ,computer.software_genre ,01 natural sciences ,Goods and services ,Environmental Chemistry ,IMPACT ASSESSMENTS ,Waste Management and Disposal ,Life-cycle assessment ,0105 earth and related environmental sciences ,Database ,Impact assessment ,GEOMETRIC MEAN ,BOTTOM-UP REFERENCE ,INVENTORY DATABASE ,Pollution ,DATABASE SYSTEMS ,LIFE CYCLE ASSESSMENT (LCA) ,Reference values ,[SDE]Environmental Sciences ,TOP DOWN APPROACHES ,computer - Abstract
International audience; The process-based life cycle assessments (LCA) of goods and services are calculated using a bottom-up approach related to a functional unit. However, this does not provide any information regarding the scale of the environmental impacts. Therefore, the normalization allows to relate the impacts to a reference system (specific countries, regions or even the whole world). These references are usually obtained from top-down approach. The different data sources introduce inconsistencies on results and raise doubts on their adequacy and representativity. This paper proposes a novel approach for determining the data for the reference in order to ensure consistency about boundaries, data sources and modelling hypotheses describing the system. For this purpose, normalization is applied as an expression of the result relative to the average component of the reference system, instead of the sum of all the components. The reference values are determined from the geometric means of the datasets of the inventory database, used for assessing the studied systems. The exemplary application to the ecoinvent databases provides normalization references for 878 versions of the impacts categories listed by ecoinvent and for the 2077 involved substances. For eight impact assessment methods, the results are compared with 16 normalization sets from the literature and point out highly significant correlations.
- Published
- 2019
- Full Text
- View/download PDF
9. Database-Assisted Spectrum Sharing in Satellite Communications: A Survey
- Author
-
Jonathan Gardey, Marko Hoyhtya, Xianfu Chen, Jean-Christophe Dunat, Janne Janhunen, Aarne Mammela, and Ari Hulkkonen
- Subjects
General Computer Science ,Cover (telecommunications) ,millimetre-wave communications ,millimetre-wave communication ,Computer science ,050801 communication & media studies ,Satellite system ,02 engineering and technology ,Interference (wave propagation) ,computer.software_genre ,Radio spectrum ,0508 media and communications ,0202 electrical engineering, electronic engineering, information engineering ,General Materials Science ,Physics::Atmospheric and Oceanic Physics ,database systems ,ta113 ,Dynamic spectrum access ,Database ,ta213 ,05 social sciences ,General Engineering ,020206 networking & telecommunications ,Orbit ,dynamic spectrum access ,Physics::Space Physics ,Communications satellite ,Orbit (dynamics) ,Geostationary orbit ,Satellite ,lcsh:Electrical engineering. Electronics. Nuclear engineering ,computer ,lcsh:TK1-9971 - Abstract
This survey paper discusses the feasibility of sharing the spectrum between satellite telecommunication networks and terrestrial and other satellite networks on the basis of a comprehensive study carried out as part of the European Space Agency's (ESA) Advanced Research in Telecommunications Systems (ARTES) programme. The main area of investigation is the use of spectrum databases to enable a controlled sharing environment. Future satellite systems can largely benefit from the ability to access spectrum bands other than the dedicated licensed spectrum band. Potential spectrum sharing scenarios are classified as: a) secondary use of the satellite spectrum by terrestrial systems, b) satellite system as a secondary user of spectrum, c) extension of a terrestrial network by using the satellite network, and d) two satellite systems sharing the same spectrum. We define practical use cases for each scenario and identify suitable techniques. The proposed scenarios and use cases cover several frequency bands and satellite orbits. Out of all the scenarios reviewed, owing to the announcement of many different mega-constellation satellite networks, we focus on analysing the feasibility of spectrum sharing between geostationary orbit (GSO) and non-geostationary orbit (NGSO) satellite systems. The performance is primarily analysed on the basis of widely accepted recommendations of the Radiocommunications Sector of the International Telecommunications Union (ITU-R). Finally, future research directions are identified.
- Published
- 2017
- Full Text
- View/download PDF
10. Kümelemeye dayalı otomatik dizin seçim aracı gerçekleştirimi
- Author
-
Yanatma, Mehmet Akif, Kalay, Mustafa Utku, and Bilgisayar Mühendisliği Anabilim Dalı
- Subjects
Database ,Database systems ,Cluster technics ,Index selection ,Indexing ,Cluster method ,Database design ,Database management system ,Computer Engineering and Computer Science and Control ,Clustering ,Bilgisayar Mühendisliği Bilimleri-Bilgisayar ve Kontrol ,Index - Abstract
Günümüzde sayısal veri miktarının artmasıyla birlikte veri tabanı sistemlerinin kullanımı da artmıştır. Bu kadar çok veri arasından istenilen veriye hızlı bir şekilde ulaşmada ise veri tabanlarının fiziksel tasarımı önemli rol oynamaktadır. Fiziksel veri tabanı tasarımının önemli bir kısmını dizin seçimi oluşturmaktadır. Dizinler, tüm tablonun taranmasına gerek duyulmadan belirli seçim şartlarına uyan satırlara ulaşmayı sağlayan yardımcı yapılardır. Yüzlerce tablo ve binlerce sütun içeren ilişkisel bir veri tabanı için milyarlarca dizin oluşturabilme ihtimali vardır. Ancak bir veri tabanı için dizin sayısı sınırsız bir şekilde arttırılamaz. Çünkü veri tabanında yapılan bir güncellemeden sonra yeni veri tabanı ile tutarlı kalmak için dizinlerin de güncellenmesi gerekir, dolayısıyla dizinlerin bir bakım maliyeti vardır. Diğer taraftan dizinler de birer dosya olduklarından depolama maliyetleri vardır. Bu esaslar göz önünde bulundurulduğunda dizin seçim problemi, NP-tam bir problem olarak karşımıza çıkmaktadır. Böyle bir problem karşısında veri tabanı yöneticisinin bir veri tabanı için dizin seçimi yapması çok zordur.Bu tez kapsamında dizin seçim problemini otomatik olarak çözmek için dizin seçim aracı geliştirilmiştir. Dizin seçim aracı geliştirilirken kümeleme tekniklerinden yararlanılmıştır. Geliştirilen dizin seçim aracı ile bir iş yükü için uygun dizin kümesi seçerek, veri tabanı yöneticisi üzerindeki iş yükünü hafifletmek ve iş yükünü oluşturan sorguların cevaplanması için gereken toplam zamanın ve toplam disk erişim sayısının azaltılması hedeflenmiştir.Dizin seçim aracının çalışma prensibi, sorgulardan oluşan bir iş yükünü, birbirine benzer sorgular bir araya gelecek şekilde kümelere ayırmaya çalışmaktır. Benzer sorguları biraraya getirmedeki amaç ise birbirine benzer sorguların ortak nitelikleri üzerinde dizin oluşturma önerisinde bulunmaktır. Böylece oluşturulacak dizinlerin birden çok sorgu için faydalı olması beklenmektedir.Yapılan performans değerlendirmesi testlerinde iş yükünü oluşturan sorgular ardışık ve eş zamanlı olmak üzere 2 farklı şekilde çalıştırılmıştır. Testler ile iş yükünün işlenmesi için gereken toplam disk erişimi sayısı ve toplam cevap süresi ölçülmüştür. İş yükünün işlenmesi için gereken toplam disk erişimi sayısı ve toplam cevap süresinde, kümeleme yöntemi ile elde edilen dizin önerilerinin kullanılması ile büyük oranda iyileşme olduğu görülmüştür. Nowadays, with the increase in the number of digital data, the use of database systems has increased. The physical design of the databases plays an important role in reaching the requested data quickly among such a large number of data. An important part of the physical database design is the index selection. Indexes are auxiliary structures that allow access to rows that match specific selection conditions without having to scan the entire table. There is a possibility of creating billions of indexes for a relational database containing hundreds of tables and thousands of columns. However, the number of indexes for a database cannot be increased limitlessly. Because after an update to the database, the indexes need to be updated to remain consistent with the new database, so the indexes have a maintenance cost. On the other hand, indexes are stored in ordinary files like other data files, so they have storage cost. When these criteria are taken into consideration, the problem of index selection is an NP-complete problem. In the face of such a problem, it is very difficult for the database administrator to select the appropriate indexes for a database.In this thesis, an index selection tool has been developed to automatically solve the index selection problem. Clustering techniques were used to develop the index selection tool. By selecting the appropriate index set for a workload with the developed index selection tool, it is aimed to lighten of responsibility of database administrator and to reduce the total time and total disk access count required to answer the queries that make up the workload.The working principle of the index selection tool is to divide a workload into clusters such that similar queries come together. The purpose of grouping similar queries is to suggest indexing on the common attributes of similar queries. By this way, the indexes to be created are expected to be useful for multiple queries.In the performance evaluation tests, the queries that constitute the workload were executed in two different ways as consecutive and concurrent. The total number of disk access and total response time required to process the workload were measured with these tests. The total number of disk access count and the total response time required to process the workload has been greatly reduced by the use of index recommendations obtained by the clustering method. 112
- Published
- 2019
11. Speciation of uranium: Compilation of a thermodynamic database and its experimental evaluation using different analytical techniques
- Author
-
Frank Wagner, Clemens Walther, and Elena L. Mühr-Ebert
- Subjects
Dewey Decimal Classification::500 | Naturwissenschaften::550 | Geowissenschaften ,Calcium and magnesiums ,Water purification technologies ,010501 environmental sciences ,010502 geochemistry & geophysics ,01 natural sciences ,Specific ion interaction theories ,Chemicals removal (water treatment) ,radionuclide migration ,chemistry.chemical_compound ,Ionic strength ,ddc:550 ,Surface water resources ,Fluorescence spectroscopy ,Groundwater ,Purification ,database ,Fluorescence microscopy ,Surface waters ,Uranium ,Uranyl ,Pollution ,speciation (chemistry) ,Specific ion interaction theory ,Environmental chemistry ,radioactivity ,Chemical speciation ,Uranium compounds ,Experimental investigations ,Experimental evaluation ,chemistry.chemical_element ,Portable water purification ,experimental study ,uranium ,Time-resolved laser-induced fluorescence spectroscopy ,thermodynamics ,Geochemistry and Petrology ,Genetic algorithm ,Environmental Chemistry ,chemical composition ,Magnesium compounds ,Groundwater resources ,0105 earth and related environmental sciences ,Thermodynamic database ,Radionuclide ,chemistry ,Database systems ,pollution incidence ,Environmental science ,ionic composition - Abstract
Environmental hazards are caused by uranium mining legacies and enhanced radioactivity in utilized groundwater and surface water resources. Knowledge of uranium speciation in these waters is essential for predicting radionuclide migration and for installing effective water purification technology. The validity of the thermodynamic data for the environmental media affected by uranium mining legacies is of utmost importance. Therefore, a comprehensive and consistent database was established according to current knowledge. The uranium data included in the database is based on the NEA TDB (Guillaumont et al., 2003) and is modified or supplemented as necessary e.g. for calcium and magnesium uranyl carbonates. The specific ion interaction theory ( Bronsted, 1922 ) is used to estimate activity constants, which is sufficient for the considered low ionic strengths. The success of this approach was evaluated by comparative experimental investigations and model calculations (PHREEQC (Parkhurst and Appelo, 1999)) for several model systems. The waters differ in pH (2.7–9.8), uranium concentration (10−9-10−4 mol/L) and ionic strength (0.002–0.2 mol/L). We used chemical extraction experiments, ESI-Orbitrap-MS and time-resolved laser-induced fluorescence spectroscopy (TRLFS) to measure the uranium speciation. The latter method is nonintrusive and therefore does not change the chemical composition of the investigated waters. This is very important, because any change of the system under study may also change the speciation.
- Published
- 2019
- Full Text
- View/download PDF
12. A small vocabulary database of ultrasound image sequences of vocal tract dynamics
- Author
-
Margareth Castillo, Dagoberto Porras, Sonia H. Contreras-Ortiz, Alexander Sepulveda, and Felipe Rubio
- Subjects
Speech production ,Vocabulary ,medicine.medical_specialty ,Microphone ,Computer science ,Vision ,media_common.quotation_subject ,Articulation ,computer.software_genre ,Tongue ,Image processing ,Ultrasound ,medicine ,Speech ,Acoustic data ,Ultrasonics ,media_common ,Ultrasound videos ,Database ,Data visualization ,Ultrasound image sequences ,Acoustic speech ,Speech processing ,Visualization ,Database systems ,Cardioid ,Speech-Language Pathology ,computer ,Speech pathology ,Vocal tract ,Articulatory data - Abstract
This paper presents a new database consisting of concurrent articulatory and acoustic speech data. The articulatory data correspond to ultrasound videos of the vocal tract dynamics, which allow the visualization of the tongue upper contour during the speech production process. Acoustic data is composed of 30 short sentences that were acquired by a directional cardioid microphone. This database includes data from 17 young subjects (8 male and 9 female) from the Santander region in Colombia, who reported not having any speech pathology. © 2019 IEEE. IEEE Colombia Section;IEEE Signal Processing Society Colombia Chapter;Universidad Industrial de Santander
- Published
- 2019
13. The Development of Information Educational Environment
- Author
-
Valentina A. Chupina, Tatiana Lavina, Nikita Topolskiy, Elena A. Zvereva, Olga A. Fedorenko, and Igor Matyushchenko
- Subjects
Reflection (computer programming) ,Knowledge management ,Process (engineering) ,Computer science ,DATABASE ,APPLICATION PROGRAMS ,CRITICAL THINKING ,STUDENTS ,PROFESSIONAL PRACTICES ,Development (topology) ,Software ,DESIGN ,FUNCTIONAL ANALYSIS ,ICT COMPETENCES ,ComputingMilieux_COMPUTERSANDEDUCATION ,ICT-COMPETENCE ,EDUCATIONAL ENVIRONMENT ,business.industry ,INFORMATION-EDUCATIONAL ENVIRONMENT ,WEB APPLICATION ,Information technology ,Functional requirement ,COMMUNICATIVE REFLECTION ,BIOMEDICAL ENGINEERING ,WEB-APPLICATION ,DATABASE SYSTEMS ,Critical thinking ,PROFESSIONAL PRACTICE ,business - Abstract
Within the framework of the research, the information-educational environment 'Practice' has been developed: functional requirements have been identified, software and development tools have been selected, a database has been designed, a client-server web-application for its access has been implemented. The created information educational environment will effectively organize the process of students' practical training in terms of increasing the level of formation of the most important 'XXI century skills': bachelors' ICT-competence, communicative reflection, critical thinking. © 2019 IEEE.
- Published
- 2019
14. A fast and flexible instance selection algorithm adapted to non-trivial database sizes
- Author
-
Rachid Harba, Frédéric Ros, Serge Guillaume, Marco Pintore, Laboratoire Pluridisciplinaire de Recherche en Ingénierie des Systèmes, Mécanique et Energétique (PRISME), Université d'Orléans (UO)-Ecole Nationale Supérieure d'Ingénieurs de Bourges (ENSI Bourges), aucun, PILA, Information – Technologies – Analyse Environnementale – Procédés Agricoles (UMR ITAP), Institut national de recherche en sciences et technologies pour l'environnement et l'agriculture (IRSTEA)-Institut national d’études supérieures agronomiques de Montpellier (Montpellier SupAgro), Institut national d'enseignement supérieur pour l'agriculture, l'alimentation et l'environnement (Institut Agro)-Institut national d'enseignement supérieur pour l'agriculture, l'alimentation et l'environnement (Institut Agro), and Institut national d’études supérieures agronomiques de Montpellier (Montpellier SupAgro)-Institut national de recherche en sciences et technologies pour l'environnement et l'agriculture (IRSTEA)
- Subjects
K-NEAREST NEIGHBORS ,SYNTHETIC AND REAL DATA ,SUPERVISED CLASSIFICATION ,Computer science ,Iterative method ,Nearest neighbor search ,NEAREST NEIGHBOR SEARCH ,Boundary (topology) ,Context (language use) ,02 engineering and technology ,computer.software_genre ,CLASSIFICATION ,Theoretical Computer Science ,Set (abstract data type) ,CLASSIFICATION (OF INFORMATION) ,DIGITAL STORAGE ,ALGORITHME ,Artificial Intelligence ,020204 information systems ,INSTANCE SELECTION ,0202 electrical engineering, electronic engineering, information engineering ,CLUSTERING ALGORITHM ,Cluster analysis ,ITERATIVE PROCESS ,Iterative and incremental development ,Database ,ITERATIVE METHODS ,ALGORITHMS ,Process (computing) ,DIFFERENT MECHANISMS ,BASE DE DONNEES ,CLUSTERING ALGORITHMS ,DATABASE SYSTEMS ,INFORMATIQUE ,STORAGE REQUIREMENTS ,[SDE]Environmental Sciences ,020201 artificial intelligence & image processing ,Computer Vision and Pattern Recognition ,Data mining ,FILTERING PROCESS ,computer ,Algorithm - Abstract
International audience; In this paper, a new instance selection algorithm is proposed in the context of classification to manage non-trivial database sizes. The algorithm is hybrid and runs with only a few parameters that directly control the balance between the three objectives of classification, i.e. errors, storage requirements and runtime. It comprises different mechanisms involving neighborhood and stratification algorithms that specifically speed up the runtime without significantly degrading efficiency. Instead of applying an IS (Instance Selection) algorithm to the whole database, IS is applied to strata deriving from the regions, each region representing a set of patterns selected from the original training set. The application of IS is conditioned by the purity of each region (i.e. the extent to which different categories of patterns are mixed in the region) and the stratification strategy is adapted to the region components. For each region, the number of delivered instances is firstly limited via the use of an iterative process that takes into account the boundary complexity, and secondly optimized by removing the superfluous ones. The sets of instances determined from all the regions are put together to provide an intermediate instance set that undergoes a dedicated filtering process to deliver the final set. Experiments performed with various synthetic and real data sets demonstrate the advantages of the proposed approach.
- Published
- 2015
- Full Text
- View/download PDF
15. Data deduplication on database: Atatürk University student information system case
- Author
-
Bayoğlu, Yakup, Özdemir, Abdulkadir, and İşletme Ana Bilim Dalı
- Subjects
Database ,Student information system ,Database systems ,Universities ,Data deduplication ,İşletme ,Information systems ,Management information systems ,Business Administration - Abstract
Veri kavramı hayatımızın her alanında kullanılmaktadır. Bilişim sistemleri söz konusu olduğunda; veri tabanı, veri güvenliği, veri tutarlılığı, veri kirliliği kavramlarında olduğu gibi yanına yeni sözcükler alarak karşımıza çıkmaktadır.Atatürk Üniversitesi'nde kullanılan birçok sistem, çok fazla kişi tarafından kullanılmakta ve değişik şekillerde veriler girilmektedir. Girilen veriler aynı şeyi ifade etmesine rağmen kullanıcıların değişik bakış açılarından dolayı; bazen farkında olmadan, bazen de kasıtlı olarak farklı şekillerde girilebilmektedir. Bu farklılıklar sebebi ile tekrar tekrar eklenmiş olan veriler sistemin sağlıklı çalışmasını etkileyecek derecede veri kirliliğine sebep olabilmektedir. Sistemlerdeki bu tarz veri kirliliğinin önüne geçmenin çeşitli yöntemleri bulunmaktadır. Bu yöntemler kullanıcıların sistemdeki verilere müdahalesini kısıtlamak/engellemek ve yeni kayıt girilirken çeşitli kontroller yapıldıktan sonra eklemek şeklindedir. Atatürk Üniversitesi oldukça büyük ve köklü bir geçmişe sahip olmasından dolayı Öğrenci Bilgi Sistemine (ÖBS) ait veri tabanındaki veriler çok eski tarihlere uzanmaktadır. Daha önceleri yukarıda bahsedilen tedbirler uygulanmadığından dolayı, sistemde eskiden kalma kirli veriler mevcuttur. Ders adlarındaki veri kirliliği, ÖBS de en çok karşılaşılan ve en çok soruna sebep olan kirliliktir. Bu çalışmamızda aynı isim ve krediye sahip derslerin tek kayıtta birleştirilmesi ve aslında aynı olup farklı şekilde yazılmış olan derslerin tespit edilerek bunların da tek bir kayıtta birleştirilmesi amaçlanmıştır. Bu çalışma kapsamında ders adlarında bazı düzeltmeler yapılmış ve akabinde aynı isimli dersler birleştirilmiştir. Dizge karşılaştırma algoritmalarından Damerau-Levenshtein algoritması kullanılarak benzer isimli dersler tespit edilmiş ve bunlardan da birleştirilmesi uygun görülen dersler birleştirilmiştir. The concept of data is used in all areas of our lives. When it comes to information systems, the word data is used with another words alongside. Such as database, data safety, data integrity, data pollution etc. The systems used in Atatürk University, are used by many people and data is entered in different ways. Although the input data indicate the same thing, due to the different approaches of the users; sometimes unwittingly, sometimes intentionally, it could be entered in different forms. Data which is repeatedly added due to these different approaches, could cause data corruption that will affect the healthy operation of the system. There are various methods of preventing such data pollution in systems. These methods are to restrict / prevent users from interfering with the data in the system and to make various checks while entering the new record. Since Atatürk University has a very large and deep-rooted history, the data in the database of the Student Information System (SIS) goes back to very old dates. Since measures mentioned above had not been applied in earlier time, there are some old dirty data in the system. The data pollution in the course names is the most encountered pollution and causes the most problems in SIS. In this study, it is aimed to combine the courses with the same name and credits in a single record. It is also aimed to determine the courses that are actually the same but typed differently, and combine them in a single record. Within the scope of this study, some corrections have been made in the course names and then the courses with the same name were combined. Similarly named courses were determined by using Damerau-Levenshtein Algorithm which is one of string comparison algorithms. Similarly named courses which are agreed to be same ones, were combined in a single record. 126
- Published
- 2018
16. Türkiye sağlık araştırmasının veri madenciliği teknikleri ile incelenmesi
- Author
-
Can, Onur, Senger, Ötüken, and İşletme Anabilim Dalı
- Subjects
Health sector ,Database ,Health management ,Database systems ,İşletme ,Health ,Health surveys ,Sağlık Kurumları Yönetimi ,Algorithmic solutions ,Health Care Management ,Data mining ,Algorithms ,Business Administration - Abstract
Bu çalışmanın temel amacı, bireylerin genel sağlık durumunu etkileyen faktörleri veri madenciliği teknikleri ile belirlemektir. Bu amaç doğrultusunda, 2014 yılında Türkiye İstatistik Kurumu tarafından yapılan Türkiye Sağlık Araştırmasına ait B grubu mikro veri seti temin edilmiştir. Daha sonra çeşitli manipülasyon yöntemleriyle veri madenciliği analizlerine uygun hale getirilen veri seti üzerinde CHAID, QUEST ve C&R Tree algoritmaları uygulanmıştır. Analiz sonuçlarına göre kullanılan üç algoritma da genel sağlık durumu üzerinde en çok etkiye sahip faktörlerin, 6 ay ya da daha uzun süren veya sürmesi beklenen sağlık problemi ve herhangi bir sağlık problemi nedeniyle günlük faaliyetlerin 6 ay ve daha uzun süre kısıtlanma durumu olduğunu tespit etmiştir. Karar ağacı algoritmaları birbirleri ile kıyaslandığında ise CHAID algoritmasının hem QUEST hem de C&R Tree algoritmasına göre daha iyi sonuçlar verdiği gözlenmiştir. The main purpose of this study is to determine the factors affecting general state of health of individuals by data mining techniques. In accordance with this purpose, B group micro data set of the Turkish Health Research conducted by the Turkish Statistical Institute in 2014 was provided. Afterwards, CHAID, QUEST and C&R Tree algorithms were applied on the data set that was made suitable for data mining analysis with data pre-processing methods. The three algorithms used according to the results of the analysis found that the factors with the greatest impact on the general state of health, the health problem which is expected to last for 6 months or longer, and the daily activities are restricted for 6 months or longer due to any health problem. When the decision tree algorithms are compared with each other, it is observed that the CHAID algorithm gives better results than both the QUEST and C&R Tree algorithms. 152
- Published
- 2017
17. Data Encryption on GPU for High-Performance Database Systems
- Author
-
Seung-Tae Hong, Heeseung Jo, Jae-Woo Chang, and Dong Hoon Choi
- Subjects
Outsourced databases ,Database ,business.industry ,Computer science ,Computation ,GPU ,Encryption ,Parallel computing ,computer.software_genre ,CUDA ,Database systems ,General Earth and Planetary Sciences ,business ,computer ,General Environmental Science - Abstract
Graphics processing units have proved its capability for general purpose computing in many research areas. In this paper, we propose the mechanism and implementation of a database system that encrypts and decrypts data by using GPU. The proposed mechanism is mainly designed for a database system that requires data encryption and decryption to support high security level or ODBS. By exploiting the computation capability of GPU, we achieve not only a fast encryption and decryption time per operation but also a higher overall performance of a database system by offioading computation to GPU. Moreover, the proposed system includes a mechanism which can decide whether to offload computation to GPU or not for more performance gain. We implemented the AES algorithm based on CUDA framework and integrate with MySQL, a commodity database system. Our evaluation demonstrates that the encryption and decryption on GPU shows eight times better performance compared that on CPU when the data size is 16 MB. We also show that the proposed system alleviates the utilization of CPU and the overall performance of database system is improved by offioading heavy encrypting and decrypting computation to GPU.
- Published
- 2013
- Full Text
- View/download PDF
18. Implications of non-volatile memory as primary storage for database management systems
- Author
-
Adrian Cristal, Osman Unsal, Ozcan Ozturk, Naveed Ul Mustafa, Adria Armejach, Barcelona Supercomputing Center, and Universitat Politècnica de Catalunya. Departament d'Arquitectura de Computadors
- Subjects
Emulation platform ,Digital storage ,Jerarquia de memòria (Informàtica) ,Computer science ,Embedded systems ,Storage capacity ,Memory hierarchy (Computer science) ,02 engineering and technology ,Relational data ,computer.software_genre ,01 natural sciences ,Memory architecture ,Data_FILES ,0202 electrical engineering, electronic engineering, information engineering ,Relational database systems ,Power constraints ,010302 applied physics ,Sequential access memory ,Database ,Dynamic random access storage ,NVM Express ,Enginyeria electrònica [Àrees temàtiques de la UPC] ,Random access memory ,020202 computer hardware & architecture ,Data storage equipment ,Query execution time ,Computer data storage ,Management information systems ,Primary storages ,Memory management ,Nonvolatile storage ,Query processing ,Bases de dades--Gestió ,Information management ,Database management ,Hardware ,In-Memory Processing ,0103 physical sciences ,Computer architecture ,Virtual storage ,Computer memory ,Hardware_MEMORYSTRUCTURES ,business.industry ,Nonvolatile memory ,Semiconductor memory ,Non-volatile memory ,Database systems ,Intermediate storage ,Operating system ,business ,computer ,Hard disks - Abstract
Traditional Database Management System (DBMS) software relies on hard disks for storing relational data. Hard disks are cheap, persistent, and offer huge storage capacities. However, data retrieval latency for hard disks is extremely high. To hide this latency, DRAM is used as an intermediate storage. DRAM is significantly faster than disk, but deployed in smaller capacities due to cost and power constraints, and without the necessary persistency feature that disks have. Non-Volatile Memory (NVM) is an emerging storage class technology which promises the best of both worlds. It can offer large storage capacities, due to better scaling and cost metrics than DRAM, and is non-volatile (persistent) like hard disks. At the same time, its data retrieval time is much lower than that of hard disks and it is also byte-addressable like DRAM. In this paper, we explore the implications of employing NVM as primary storage for DBMS. In other words, we investigate the modifications necessary to be applied on a traditional relational DBMS to take advantage of NVM features. As a case study, we have modified the storage engine (SE) of PostgreSQL enabling efficient use of NVM hardware. We detail the necessary changes and challenges such modifications entail and evaluate them using a comprehensive emulation platform. Results indicate that our modified SE reduces query execution time by up to 40% and 14.4% when compared to disk and NVM storage, with average reductions of 20.5% and 4.5%, respectively. The research leading to these results has received funding from the European Union’s 7th Framework Programme under grant agreement number 318633, the Ministry of Science and Technology of Spain under contract TIN2015-65316-P, and a HiPEAC collaboration grant awarded to Naveed Ul Mustafa.
- Published
- 2016
- Full Text
- View/download PDF
19. Dynamic execution of secure queries over homomorphic encrypted databases
- Author
-
Renaud Sirdey, Oana Stan, Sergiu Carpov, Département d'Architectures, Conception et Logiciels Embarqués-LIST (DACLE-LIST), Laboratoire d'Intégration des Systèmes et des Technologies (LIST (CEA)), Direction de Recherche Technologique (CEA) (DRT (CEA)), Commissariat à l'énergie atomique et aux énergies alternatives (CEA)-Commissariat à l'énergie atomique et aux énergies alternatives (CEA)-Direction de Recherche Technologique (CEA) (DRT (CEA)), Commissariat à l'énergie atomique et aux énergies alternatives (CEA)-Commissariat à l'énergie atomique et aux énergies alternatives (CEA), and Laboratoire d'Intégration des Systèmes et des Technologies (LIST)
- Subjects
Homomorphic secret sharing ,Query processing ,Computer science ,02 engineering and technology ,computer.software_genre ,Encryption ,Multiple encryption ,Filesystem-level encryption ,0202 electrical engineering, electronic engineering, information engineering ,Cloud computing ,[INFO]Computer Science [cs] ,Run-time information ,Boolean circuit ,Ho-momorphic encryptions ,060201 languages & linguistics ,Database ,business.industry ,Cloud computing securities ,Fully homomorphic encryption ,06 humanities and the arts ,Protection mechanisms ,Reconfigurable hardware ,Bus encryption ,Database systems ,Communication overheads ,Homomorphic Encryption Schemes ,Probabilistic encryption ,0602 languages and literature ,Query languages ,Cryptography ,020201 artificial intelligence & image processing ,Link encryption ,On-the-fly encryption ,business ,computer ,Data privacy ,Logic circuits - Abstract
Conference of 4th ACM International Workshop on Security in Cloud Computing, SCC 2016 ; Conference Date: 30 May 2016; Conference Code:121977; International audience; The wide use of cloud computing and of data outsourcing rises important concerns with regards to data security resulting thus in the necessity of protection mechanisms such as encryption of sensitive data. The recent major theoretical breakthrough of finding the Holy Grail of encryption, i.e. fully homomorphic encryption guarantees the privacy of queries and their results on encrypted data. However, there are only a few studies proposing a practical performance evaluation of the use of homomorphic encryption schemes in order to perform database queries. In this paper, we propose and analyse in the context of a secure framework for a generic database query interpreter two different methods in which client requests are dynamically executed on homomorphically encrypted data. Dynamic compilation of the requests allows to take advantage of the different optimizations performed during an off-line step on an intermediate code representation, taking the form of boolean circuits, and, moreover, to specialize the execution using runtime information. Also, for the returned encrypted results, we assess the complexity and the efficiency of the different protocols proposed in the literature in terms of overall execution time, accuracy and communication overhead.
- Published
- 2016
- Full Text
- View/download PDF
20. Geometrical codification for clustering mixed categorical and numerical databases
- Author
-
José Luis Díez and Fátima Barceló-Rico
- Subjects
General method ,Clustering algorithms ,Computer Networks and Communications ,Computer science ,k-means ,Mixed data ,Categorical attributes ,Mixed database ,computer.software_genre ,Computer Science::Digital Libraries ,Clustering ,Data conversion ,Matrix (mathematics) ,Spherical coordinates ,Artificial Intelligence ,Cluster (physics) ,k-Means algorithm ,Cluster analysis ,Categorical variable ,Database ,k-means clustering ,Spherical coordinate system ,computer.file_format ,INGENIERIA DE SISTEMAS Y AUTOMATICA ,Input matrices ,Benchmarking ,Database systems ,Hardware and Architecture ,Data mining ,computer ,Software ,Codification error ,Information Systems - Abstract
[EN] This paper presents an alternative to cluster mixed databases. The main idea is to propose a general method to cluster mixed data sets, which is not very complex and still can reach similar levels of performance of some good algorithms. The proposed approach is based on codifying the categorical attributes and use a numerical clustering algorithm on the resulting database. The codification proposed is based on polar or spherical coordinates, it is easy to understand and to apply, the increment in the length of the input matrix is not excessively large, and the codification error can be determined for each case. The proposed codification combined with the well known k-means algorithm showed a very good performance in different benchmarks and has been compared with both, other codifications and other mixed clustering algorithms, showing a better or comparable performance in all cases., The authors acknowledge the partial funding of this work by the National projects DPI2007-66728-C02-01 and DPI2008-06737-C02-01.
- Published
- 2011
- Full Text
- View/download PDF
21. A feature extraction software tool for agricultural object-based image analysis
- Author
-
A. Fernández-Sarría, Txomin Hermosilla, Luis A. Ruiz, and J. A. Recio
- Subjects
Semivariogram ,Computer science ,Semivariograms ,Transform ,Shapefile ,computer.software_genre ,Software testing ,Image analysis ,Input output programs ,Hough transform ,law.invention ,Co-occurrence-matrix ,Digital image ,Numeric values ,Software ,Spatial arrangements ,Feature extraction algorithms ,law ,Agricultural land ,Grey levels ,Input datas ,Object based image analysis ,Spectral information ,Land use/land cover ,Classification (of information) ,Contextual image classification ,Dynamic tools ,Extraction procedure ,Spatial database ,Forestry ,computer.file_format ,Remote sensing ,Intensity values ,Computer Science Applications ,Data processing ,Algorithm ,Processing time ,Agricultural database updating ,Feature extraction ,Data sets ,Data mining ,Land cover ,Spectral resolution ,Geo-spatial database ,Image classification ,Object shape ,High resolution digital images ,Software tool ,Numerical method ,Horticulture ,Image band ,Descriptors ,Database ,Wavelet transforms ,Texture ,business.industry ,Parcel-based ,Hough transforms ,Multi-spectral ,Software applications ,Co-occurrence matrix ,Database systems ,Land use ,INGENIERIA CARTOGRAFICA, GEODESIA Y FOTOGRAMETRIA ,business ,Agronomy and Crop Science ,computer ,Structural feature - Abstract
A software application for automatic descriptive feature extraction from image-objects, FETEX 2.0, is presented and described in this paper. The input data include a multispectral high resolution digital image and a vector file in shapefile format containing the polygons or objects, usually extracted from a geospatial database. The design of the available descriptive features or attributes has been mainly focused on the description of agricultural parcels, providing a variety of information: spectral information from the different image bands; textural descriptors of the distribution of the intensity values based on the grey level co-occurrence matrix, the wavelet transform and a factor of edgeness; structural features describing the spatial arrangement of the elements inside the objects, based on the semivariogram curve and the Hough transform; and several descriptors of the object shape. The output file is a table that can be produced in four alternative formats, containing a vector of features for every object processed. This table of numeric values describing the objects from different points of view can be externally used as input data for any classification software. Additionally, several types of graphs and images describing the feature extraction procedure are produced, useful for interpretation and understanding the process. A test of the processing times is included, as well as an application of the program in a real parcel-based classification problem, providing some results and analyzing the applicability, the future improvement of the methodologies, and the use of additional types of data sets. This software is intended to be a dynamic tool, integrating further data and feature extraction algorithms for the progressive improvement of land use/land cover database classification and agricultural database updating processes. © 2011 Elsevier B.V., The authors appreciate the financial support provided by the Spanish Ministerio de Ciencia e Innovacion and the FEDER in the framework of the Project CGL2009-14220 and CGL2010-19591/BTE, the Spanish Institut Geografico Nacional (IGN), Institut Cartografico Valenciano (ICV), Institut Murciano de Investigacion y Desarrollo Agrario y Alimentario (IMIDA) and Banco de Terras de Galicia (Bantegal).
- Published
- 2011
- Full Text
- View/download PDF
22. A Join Algorithm for Large Databases: A Quadtrees Structure Approach
- Author
-
Hatim Aboalsamh
- Subjects
Hash join ,Recursive join ,General Computer Science ,Database ,Computer science ,Relational database ,Sort-merge join ,Block nested loop ,computer.software_genre ,Relational Databases ,Quadtrees ,Join Based Queries ,Quadtree ,Join (sigma algebra) ,Database Systems ,Hash Join ,computer ,Algorithm ,Time complexity - Abstract
Enhancing the performance of large database systems depends heavily on the cost of performing join operations. When two very large tables are joined, optimizing such operation is considered one of the interesting research topics to many researchers, especially when both tables, to be joined, are very large to fit in main memory. In such case, join is usually performed by any other method than hash Join algorithms. In this paper, a novel join algorithm that is based on the use of quadtrees, is introduced. Applying the proposed algorithm on two very large tables, that are too large to fit in main memory, is proven to be fast and efficient. In the proposed new algorithm, both tables are represented by a storage efficient quadtree that is designed to handle one-dimensional arrays (1-D arrays). The algorithm works on the two 1-D arrays of the two tables to perform join operations. For the new algorithm, time and space complexities are studied. Experimental studies show the efficiency and superiority of this algorithm. The proposed join algorithm requires minimum number of I/O operations and operates in main memory with O(n log (n/k)) time complexity, where k is number of key groups with same first letter, and (n/k) is much smaller than n.
- Published
- 2010
- Full Text
- View/download PDF
23. Structure, Components, and Interfaces of the Airborne Prism Experiment (APEX) Processing and Archiving Facility
- Author
-
Stefan Adriaensen, Daniel Odermatt, Klaus I Itten, S. Kempenaers, Jan Biesemans, F. Dell'Endice, Andreas Hueni, Jens Nieke, Koen Meuleman, Daniel Schläpfer, M. Kneubuehler, University of Zurich, and Hueni, A
- Subjects
Information management ,parallel processing ,Relational database ,Computer science ,system architecture ,computer.software_genre ,Software ,910 Geography & travel ,Electrical and Electronic Engineering ,Adaptation (computer science) ,hyperspectral data calibration ,demand processing ,Database ,business.industry ,2208 Electrical and Electronic Engineering ,1900 General Earth and Planetary Sciences ,Hyperspectral imaging ,on ,10122 Institute of Geography ,Database systems ,Parallel processing (DSP implementation) ,Systems architecture ,General Earth and Planetary Sciences ,business ,computer ,Computer hardware - Abstract
The product generation from hyperspectral sensor data has high requirements on the processing infrastructure, both hardware and software. The Airborne Prism Experiment (APEX) processing and archiving facility has been set up to provide for the automated generation of level-1 calibrated data and user-configurable on-demand product generation for higher processing levels. The system offers full reproducibility of user orders and processing parameters by employing a relational database. The flexible workflow software allows for the quick integration of novel algorithms or the definition of new processing sequences. Reprocessing of data is supported by the archiving approach. Configuration management based on the database enables the control over different versions of processing modules to be applied. The system is described with a focus on the APEX instrument; however, its generic design allows adaptation to other sensor systems.
- Published
- 2009
- Full Text
- View/download PDF
24. Managing Records and Information for Transparent, Accountable, and Inclusive Governance in the Digital Environment : Lessons from Nordic Countries
- Author
-
Thurston, Anne
- Subjects
COMMUNICATIONS ,DEVICES ,DIGITAL SIGNATURES ,INFORMATION ,DIGITAL SIGNATURE ,DIGITAL FORMAT ,SOFTWARE ,COMMUNICATION ,DATABASES ,ELECTRONIC IDENTITY ,DATA SECURITY ,DATA FORMATS ,INFORMATION ARCHITECTURE ,DIGITIZATION ,WEBSITES ,RETENTION ,MONITORING ,INFORMATION TECHNOLOGY ,CODES ,ACCESS RIGHTS ,INTERNATIONAL STANDARDS ,USABILITY ,COMPETITIVENESS ,RELATIONAL DATABASE ,TECHNICAL SUPPORT ,SERVICE PROVIDERS ,BUSINESS ,TRANSACTIONS ,FRAUD ,INFORMATION POLICY ,PUBLIC KEY INFRASTRUCTURE ,RELIABILITY ,PROCUREMENT ,INSTITUTIONS ,VIRUS ,TECHNOLOGIES ,REAL TIME ,E-MAIL ,METADATA ,PHOTOS ,INFORMATION SERVICES ,USERS ,INTERFACES ,HARDWARE ,STORAGE ,TECHNOLOGICAL CHANGE ,COMPUTER ,WORD PROCESSING ,PERSONAL DATA ,IMAGES ,DIGITAL DOCUMENTS ,MARKUP ,E-BUSINESS ,BASIC ,SYSTEM DESIGN ,INFORMATION SYSTEMS ,PROTECTION OF DATA ,PROFIT ,LINKS ,DEVELOPMENT OF INFORMATION ,MEDIA ,ADVANCED SEARCH ,RELATIONAL DATABASES ,SILOS ,PUBLIC UTILITIES ,SEARCH ,TRACEABILITY ,INTEROPERABILITY ,DIGITAL ,E-GOVERNANCE ,INFORMATION TRANSFER ,POLICY SUPPORT ,INFORMATION SYSTEM ,COMPONENTS ,SERVERS ,DIGITAL SYSTEMS ,SECURITY MANAGEMENT ,XML ,PERFORMANCE ,DATA PROCESSING ,INFORMATION FLOW ,DATABASE SYSTEMS ,ENCRYPTION ,GOVERNMENT OFFICES ,IT ,MANAGEMENT SYSTEMS ,LEGAL ISSUES ,ADMINISTRATION PROCESSES ,INFORMATION NETWORKS ,AT ,ELECTRONIC DOCUMENT ,WEB SERVICES ,ECONOMIC DEVELOPMENT ,TELEPHONE ,DATA ,DOCUMENT MANAGEMENT ,INNOVATION ,DATA STRUCTURES ,ELECTRONIC FORMS ,HACKING ,COMMUNICATIONS TECHNOLOGY ,CONTENT MANAGEMENT ,EMAIL ,DOMAIN ,ACTION PLAN ,BUSINESS SYSTEMS ,DATA MODELS ,CUSTOMER SERVICE ,DIGITAL SERVICES ,NETWORK ,TECHNOLOGY DEVELOPMENT ,IMAGING ,RESULT ,PUBLIC ADMINISTRATION ,SECURITY ,PRIVACY ,KNOWLEDGE ECONOMY ,ACCESS TO INFORMATION ,ALGORITHMS ,POLICY FORMULATION ,USES ,USER ,WEB ,NETWORKS ,TECHNOLOGICAL DEVELOPMENTS ,ACCESS TO SERVICES ,INTERFACE ,PUBLIC KEY ,SERVICES TO CITIZENS ,EQUIPMENT ,PUBLIC ACCESS ,SOCIAL DEVELOPMENT ,TECHNOLOGICAL DEVELOPMENT ,PRIVATE SECTOR ,AUTHENTICATION ,DIGITAL ARCHIVE ,USER INTERFACE ,FILES ,DATABASE ,USER INTERFACES ,WEBSITE ,E-SERVICES ,MATERIALS ,ENCRYPTION KEY ,DATA MANAGEMENT ,ADMINISTRATION ,INTEGRATED SOFTWARE ,INNOVATIONS ,IT INFRASTRUCTURE ,ADMINISTRATIVE PROCESSES ,SEARCHES ,DIGITAL STORAGE ,REGISTERS ,SYSTEMS ,PERSONAL DATA PROTECTION ,TECHNOLOGY ,INFORMATION SOCIETY ,MATERIAL ,INTERNET ,INNOVATION POLICY ,CUSTOMER ,INFORMATION STRUCTURES ,RESULTS ,LEGAL FRAMEWORK ,PUBLISHING ,LEGAL EXPERTS ,STANDARDIZATION ,INSPECTION ,DIGITAL ERA ,ELECTRONIC STORAGE ,REGISTRY ,ICT ,LINK ,DIGITAL INFORMATION ,DATABASE MANAGEMENT ,COMMUNITIES ,SYSTEM ,INFORMATION SECURITY ,SECURE DATA ,FUNCTIONALITY - Abstract
This set of three case studies explores the intersection of openness, digital governance, andhigh quality information in Estonia,1 Finland, and Norway with the aim of identifying lessonsthat will support the same objectives in lower resource countries. Openness, a key aspect ofthe international agenda for increasing transparency and accountability, for reducing public sectorcorruption, and for strengthening economic performance, rests on the principle that citizenshave a right to know what their governments are doing and to benefit from using governmentinformation. Goals for open, accountable, and inclusive governance rest on the assumption thattrustworthy information is available and can be shared meaningfully through strategies for digitalgovernance. This assumption needs to be examined. Does reliable and complete informationexist across lower resource countries? Can it be accessed readily? Will it survive through time?
- Published
- 2015
25. Fluorescence measurements for the identification of biological agents features for the construction of a spectra database
- Author
-
Roberto Pizzoferrato, Michela Gelfusa, D. Scarpellini, Andrea Malizia, F D'Amico, Pasquale Gaudio, M Carestia, and O Cenciarelli
- Subjects
Early detection ,Optical emission spectra ,computer.software_genre ,Fluorescence ,Spectral line ,law.invention ,law ,Biological warfare agents ,Standoff detection ,Biological warfare ,Biological weapons ,Chemical warfare ,Database systems ,Emission spectroscopy, Biological warfare agents ,Different boundary condition ,Fluorescence features ,Fluorescence measurements ,Identification techniques ,UV-LIF, Fluorescence ,Laser-induced fluorescence ,Remote sensing ,Database ,Chemistry ,Settore FIS/01 - Fisica Sperimentale ,Emission spectroscopy ,Laser ,Settore FIS/07 - Fisica Applicata(Beni Culturali, Ambientali, Biol.e Medicin) ,Spectral database ,Identification (information) ,UV-LIF ,computer - Abstract
The use of biological weapons represents a great concern both from a military and civilian point of view. The early detection of biological warfare agents (BWAs) in atmosphere is a huge challenge that could be addressed through UV-LIF (Ultra Violet Laser Induced Fluorescence) techniques. Fluorescence measurements of aerosol particles can provide gross discrimination between bio-agents and atmospheric background particles, In this work we intend to investigate the capability of discriminating among different biological warfare agents (BWA) through the analysis of the optical emission spectra. To accomplish this task, a deep knowledge of fluorescence features with different boundary conditions is required, in order to create a database of comparable spectral fingerprints. Preliminary results, obtained through a laboratory setup with a standard UV lamp source, showed that significant differences can be appreciated among BWAs simulants' spectra. This represents a first step towards the implementation of a spectral database and a laser-based biological stand-off identification technique.
- Published
- 2014
- Full Text
- View/download PDF
26. Optimization of signature file parameters for databases with varying record lengths
- Author
-
Jon M. Patton, Fazli Can, and Seyit Kocberber
- Subjects
Optimization ,General Computer Science ,Database ,Accurate estimation ,Computer science ,business.industry ,Query Languages ,Signature file ,Response time ,computer.software_genre ,Signature (logic) ,Formal proof ,File Organization ,Record Lengths ,Computer data storage ,Signature File ,Data Storage Equipment ,Overhead (computing) ,Database Systems ,business ,computer ,Algorithm - Abstract
For signature files we propose a new false drop estimation method for databases with varying record lengths. Our approach provides more accurate estimation of the number of false drops by considering the lengths of individual records instead of using the average number of terms per record. In signature file processing, accurate estimation of the number of false drops is essential to obtain a more accurate signature file and therefore to obtain a better (query) response time. With a formal proof we show that under certain conditions the number of false drops estimated by considering the average record length is less than or equal to the precise 'expected' estimation which is based on the individual record lengths. The experiments with real data show that the proposed method accurately estimates the number of false drops and the actual response time. Depending on the space overhead, our approach obtains up to 33% and 20% response time improvements for the conventional sequential and new efficient multiframe signature file methods, respectively.
- Published
- 1999
- Full Text
- View/download PDF
27. Visual Query Systems for Databases: A Survey
- Author
-
Maria Francesca Costabile, Tiziana Catarci, Stefano Levialdi, and Carlo Batini
- Subjects
Information retrieval ,Web search query ,Database ,Computer science ,Query languages ,Database systems ,Semantic Web ,Query optimization ,computer.software_genre ,Query language ,Language and Linguistics ,Computer Science Applications ,Human-Computer Interaction ,Query expansion ,Web query classification ,Sargable ,Query by Example ,computer ,RDF query language ,computer.programming_language - Abstract
Visual query systems (VQSs) are query systems for databases that use visual representations to depict the domain of interest and express related requests. VQSs can be seen as an evolution of query languages adopted into database management systems; they are designed to improve the effectiveness of the human?computer communication. Thus, their most important features are those that determine the nature of the human?computer dialogue. In order to survey and compare existing VQSs used for querying traditional databases, we first introduce a classification based on such features, namely the adopted visual representations and the interaction strategies. We then identify several user types and match the VQS classes against them, in order to understand which kind of system may be suitable for each kind of user. We also report usability experiments which support our claims. Finally, some of the most important open problems in the VQS area are described.
- Published
- 1997
- Full Text
- View/download PDF
28. Simulation of hierarchical multiprocessor database systems
- Author
-
Leonid B. Sokolinsky and P. S. Kostenetskii
- Subjects
Physical data model ,Computer science ,Simulation and analysis ,Parallel computing ,Simulation studies ,computer.software_genre ,Database design ,Database tuning ,System model ,Computer Science::Hardware Architecture ,Computer Science::Operating Systems ,Multi processor systems ,Computer Science::Databases ,Database model ,Hardware_MEMORYSTRUCTURES ,Database ,Multiprocessing systems ,Parallel database ,Comparative analysis ,InformationSystems_DATABASEMANAGEMENT ,Parallel database systems ,Database applications ,Database systems ,Online transaction processing ,Database theory ,computer ,Software - Abstract
The paper is dedicated to issues concerning simulation and analysis of hierarchical multiprocessor systems oriented to database applications. Requirements for a parallel database system model are given. A survey and comparative analysis of known parallel database system models are presented. A new multiprocessor database system model is introduced. This model allows us to simulate and evaluate arbitrary hierarchical multiprocessor configurations in the context of the OLTP class database applications. Examples of using the database multiprocessor model for simulation study of multiprocessor database systems are presented. © Pleiades Publishing, Ltd., 2013.
- Published
- 2013
29. Space optimization in deductive databases
- Author
-
Jeffrey F. Naughton, Raghu Ramakrishnan, Divesh Srivastava, and Sundararajarao Sudarshan
- Subjects
Soundness ,Theoretical computer science ,Logic Programming ,Database ,Computer science ,Query Languages ,Deductive database ,Recursive Functions ,Monotonic function ,Space optimization ,Active database ,Synchronization ,computer.software_genre ,Query language ,Redundancy (engineering) ,Database Systems ,computer ,Logic programming ,Information Systems - Abstract
In the bottom-up evaluation of logic programs and recursively defined views on databases, all generated facts are usually assumed to be stored until the end of the evaluation. Discarding facts during the evaluation, however, can considerably improve the efficiency of the evaluation: the space needed to evaluate the program, the I/O costs, the costs of maintaining and accessing indices, and the cost of eliminating duplicates may all be reduced. Given an evaluation method that is sound, complete, and does not repeat derivation steps, we consider how facts can be discarded during the evaluation without compromising these properties. We show that every such space optimization method has certain components, the first to ensure soundness and completeness, the second to avoid redundancy (i.e., repetition of derivations), and the third to reduce “fact lifetimes” (i.e., the time period for which each fact must be retained during evaluation). We present new techniques based on providing bounds on the number of derivations and uses of facts, and using monotonicity constraints for each of the first two components, and provide novel synchronization techniques for the third component of a space optimization method. We describe how techniques for each of the three components can be combined in practice to obtain a space optimization method for a program. Our results are also of importance in applications such as sequence querying, and in active databases where triggers are defined over multiple “events.”, © ACM
- Published
- 1995
- Full Text
- View/download PDF
30. Completely automated multiresolution edge snapper-A new technique for an accurate carotid ultrasound IMT measurement: Clinical validation and benchmarking on a multi-institutional database
- Author
-
Molinari, F., Pattichis, Constantinos S., Zeng, G., Saba, L., Acharya, U. R., Sanfilippo, R., Nicolaïdes, Andrew N., Suri, J. S., and Pattichis, Constantinos S. [0000-0003-1271-8151]
- Subjects
Societies and institutions ,Databases, Factual ,Computer science ,Carotid arteries ,Intima-media thickness ,Diseases ,Carotid imt ,computer.software_genre ,Automation ,ultrasound imaging ,Disease management ,Computer vision ,Segmentation ,Ultrasonography ,Image segmentation ,Database ,Contextual image classification ,Ultrasonic imaging ,article ,methodology ,Two stage ,Computer Graphics and Computer-Aided Design ,Manual segmentation ,Multi-resolutions ,Automated techniques ,Automated segmentation ,Carotid Arteries ,validation study ,tunica media ,factual database ,first-order Gaussian derivative ,Ultrasound imaging ,Data sets ,Integrated approach ,Tunica Media ,Algorithms ,Low-complexity ,First-order ,Feature-based ,Feature extraction ,Cardio-vascular disease ,Edge detection ,Image Interpretation, Computer-Assisted ,Mean distances ,Segmentation error ,Humans ,Standard deviation ,human ,image enhancement ,edge detection ,algorithm ,Scale-space ,business.industry ,segmentation ,echography ,Statistical classification ,computer assisted diagnosis ,intima–media thickness (IMT) ,Image Enhancement ,Atherosclerosis ,Database systems ,intima-media thickness (IMT) ,Clinical tools ,first-order absolute moment ,Artificial intelligence ,business ,Tunica Intima ,computer ,Software ,Carotid artery ,Measurement bias ,intima - Abstract
The aim of this paper is to describe a novel and completely automated technique for carotid artery (CA) recognition, far (distal) wall segmentation, and intima-media thickness (IMT) measurement, which is a strong clinical tool for risk assessment for cardiovascular diseases. The architecture of completely automated multiresolution edge snapper (CAMES) consists of the following two stages: 1) automated CA recognition based on a combination of scale-space and statistical classification in a multiresolution framework and 2) automated segmentation of lumen-intima (LI) and media-adventitia (MA) interfaces for the far (distal) wall and IMT measurement. Our database of 365 B-mode longitudinal carotid images is taken from four different institutions covering different ethnic backgrounds. The ground-truth (GT) database was the average manual segmentation from three clinical experts. The mean distance ± standard deviation of CAMES with respect to GT profiles for LI and MA interfaces were 0.081 ± 0.099 and 0.082 ± 0.197 mm, respectively. The IMT measurement error between CAMES and GT was 0.078 ± 0.112 mm. CAMES was benchmarked against a previously developed automated technique based on an integrated approach using feature-based extraction and classifier (CALEX). Although CAMES underestimated the IMT value, it had shown a strong improvement in segmentation errors against CALEX for LI and MA interfaces by 8% and 42%, respectively. The overall IMT measurement bias for CAMES improved by 36% against CALEX. Finally, this paper demonstrated that the figure-of-merit of CAMES was 95.8% compared with 87.4% for CALEX. The combination of multiresolution CA recognition and far-wall segmentation led to an automated, low-complexity, real-time, and accurate technique for carotid IMT measurement. Validation on a multiethnic/multi-institutional data set demonstrated the robustness of the technique, which can constitute a clinically valid IMT measurement for assistance in atherosclerosis disease management. © 2011 IEEE. 21 3 1211 1222 Cited By :37
- Published
- 2012
31. Türkiye'de bilim dallarında karşılaştırmalı bibliyometrik analiz çalışması
- Author
-
Zan, Burcu Umut, Gökkurt Demirtel, Saime Özlem, Karasözen, Bülent, and Bilgi ve Belge Yönetimi Ana Bilim Dalı
- Subjects
Bilgi ve Belge Yönetimi ,Bibliometry ,Comparative analysis ,Classification ,Information access ,Librarionship ,Scinentific disciplines ,Database ,Citation analysis ,Database systems ,Information centers ,Information ,Information access system ,Information and Records Management - Abstract
Hızla gelişen dünyada varlığını sürdürmek isteyen ülkeler bilginin üretilmesine ve üretilen bilginin kullanılarak yeni uygulama alanlarına taşınmasına önem vermeye başlamıştır. Yeni fikirlerin üretilmesinde farklı bilgi ve beceriler ile donanımlı bilim insanlarının birbirlerini tamamlayarak bilgi paylaşımında bulunması modern araştırma sistemlerinde işbirliği çalışmaları ile teşvik edilmektedir.Son yıllarda araştırma işbirliklerinin incelenmesi matematiksel ve istatistiksel yöntemlerin kullanıldığı bibliyometrik çalışmaların odak noktası haline gelmiştir. Matematiksel ve istatistiksel yöntemlerin kullanıldığı bibliyometrik çalışmalara ait analizler, atıf veri tabanlarının kullanılması ile elde edilmektedir. 50 yılı aşkın süredir bibliyometrik çalışmaların yapılabildiği kapsamlı veri tabanı olma özelliğini taşıyan Bilimsel Bilgi Enstitüsü (Institute for Scientific Information, ISI) bu özelliğini 2004 yılına kadar korumuştur. Ancak, 2004 yılında Scopus'un kurulması ile beraber atıf indeksleyen veri tabanları arasında bir rekabet ortamı oluşmuştur. Bu bağlamda her iki veri tabanında Türkiye'nin profilini çıkarmak, sosyal ve beşeri bilimler ile temel bilimler alanlarında ortak yazarlı çalışmaları karşılaştırıp, disiplinler arası farklılıkları tanımlamaya ihtiyaç duyulmuştur. Bu ihtiyaç doğrultusunda tezin konusu; 2001-2010 yılları arasında ortak yazarlı çalışmalarda temel bilimler alanında matematik konu başlığında üretilen yayınları Orta Doğu Teknik Üniversitesi (ODTÜ) ve Bilkent Üniversitesi örnekleri üzerinden ülkeler, kurumlar ve yazarlar arası işbirliği çalışmalarını değerlendirerek, kullanılan dergileri yine işbirliği çerçevesinde karşılaştırmaktır.2001-2010 yılları arasında Türkiye adresli makaleler temel alınarak karşılaştırma yapıldığında, WoS ve Scopus veri tabanlarının yapılarından kaynaklanan farklılıklar bulunduğu saptanmıştır. Ayrıca her iki veri tabanında çıkan sonuçlar değerlendirildiğinde Türkiye'nin uluslararası ortak yazarlı çalışma oranının dünya ortalamasının altında bulunduğu tespit edilmiştir. In order to continue their existence in a rapidly developing world, countries have given importance to the production of knowledge along with its transmission to new fields of application. During the production of new ideas, sharing of knowledge between scientists with alternative knowledge and skills has been encouraged within the modern research systems.Recently, the analysis of collaboration has been the main focus point of bibliometrical studies, where mathematical and statistical methods are used. The analyses, which belong to these studies, are obtained through the utilization of citation databases. The Institute for Scientific Information, having been a comprehensive database where bibliometrical studies could be done, continued its reputation until 2004. However, a competitive environment has been created among the databases indexing citations with the establishment of Scopus in 2004. Therefore, a need arose for revealing the profile of Turkey in both databases along with comparing the studies with co-authors in the fields of social and human sciences as well as physical sciences in order to define interdisciplinary differences.In the light of this requirement, the topic of the thesis is to compare and evaluate the collaborated publications among countries, institutions and authors through the samplings of Middle East Technical University and Bilkent University in terms of studies with co authors listed under the title of mathematics in the field of basic sciences between 2001 and 2010.From the comparison of articles submitted from Turkey between 2001 and 2010, differences have been determined depending on the structures of WoS and Scopus databases. Additionally, the results obtained from both databases concluded that the percentage of articles with international co authors have found to be below the world average. 264
- Published
- 2012
32. Database structure of book sales at eletronic media, an application example
- Author
-
Özgür, Serkan, Tıngöy, Özhan, İletişim Anabilim Dalı Bilişim Bilim Dalı, and İletişim Bilimleri Anabilim Dalı
- Subjects
Internet ,Veri Tabanı Yönetimi ,İletişim Bilimleri ,Books ,Electronic commerce ,Data models ,Sale ,Veri Tabanları ,Computer Engineering and Computer Science and Control ,Database ,Database systems ,Internet use ,Bilişim ,Bilişim Sistemleri ,Booksellers ,Communication Sciences ,Bilgisayar Mühendisliği Bilimleri-Bilgisayar ve Kontrol - Abstract
Konu: Elektronik Ortamda Kitap Satışı Veritabanı Yapısı Örnek Bir UygulamaÖZETElektronik ticaret günlük hayatımızdaki yerini almış bulunmaktadır. Çok çeşitli ürün grupları internet ortamında satılmaktadır. Dünyada olduğu gibi, ülkemizdede internetde en çok satılan ürün kitaptır. Bu gün ciro bazında dünyanın en büyük alışveriş siteleri kitap satış siteleridir. İnternet ortamında kitap satışı hedeflendiğinde, iyi bir site ve veritabanı tasarımına ihtiyaç bulunmaktadır. Site tasarımı ile anlatılmak istenen, görsellik kadar sitenin içeriği ve fonksiyonları ile birlikte ihtiyaçlara cevap verebilmesidir. Sitenin veritabanı yapısının kitap satışına uygun dizayn edilmesi gerekmektedir. Tablolar ve bunlar arasındaki ilişkiler özel olarak tasarlanmalıdır. Bu tezde, bilişim ve iletişim teknolojileri açıklanarak başlanmakta, ülkemizde kitabın basımından satışına kalan olan süreçler ortaya konulmakta, ideal bir web sitesinin özellikleri tanımlanmakta ve örnek bir sitenin veritabanı yapısı tablolar ve ilişkilerle birlikte incelenmektedir.Subject: Database Structure of Book Sales at Eletronic Media, An application ExampleABSTRACTElectronic commerce has taken its place in our daily lives. Various product groups sold on the internet. As in the world, the best-selling product in our country in internet is book. These days the world's biggest shopping sites on the basis of book sales turnover sites. Book sales targeted in internet environment, there is a need of a good site and database design. Site design means, with the site's content and functions of visual ability to answer to the needs. Database structure of the site should be designed in accordance with the sale of the book. tables and relations between them specially designed. In this thesis, information and communication technologies is beginning to be explained, the publication of the book examines the processes of the sale of our country, described the ideal characteristics of a web site and tables in the database structure and relationships are examined with a sample site.
- Published
- 2012
33. NoSQL databases
- Author
-
Günzl, Richard, Palovská, Helena, and Strossa, Petr
- Subjects
InformationSystems_DATABASEMANAGEMENT ,NoSQL databáze ,DBMS ,CAP theorem ,Databáze ,database systems ,Database ,CAP teorém ,SŘBD ,Apache Cassandra ,NoSQL databases ,databázové systémy - Abstract
This thesis deals with database systems referred to as NoSQL databases. In the second chapter, I explain basic terms and the theory of database systems. A short explanation is dedicated to database systems based on the relational data model and the SQL standardized query language. Chapter Three explains the concept and history of the NoSQL databases, and also presents database models, major features and the use of NoSQL databases in comparison with traditional database systems. In the fourth chapter I focus on the various representatives of NoSQL databases, in particular the ones that are most frequently used. In the next chapter, I have taken a practical look at a NoSQL database, specifically Apache Cassandra. I briefly describe the steps required to launch Apache Cassandra and its administration tools. In this practically-oriented chapter, I also show basic operations performed over a sample database using Cassandra CLI, its interactive command line interface. The purpose of this chapter is to make the reader familiar with the method of working with the Apache Cassandra database system and to point out some of its specific aspects. The primary objective of this thesis is to acquaint readers with the most important features and representatives of NoSQL databases and the potential for their practical use.
- Published
- 2012
34. Veritabanı uygulamalarında performans ve güvenlik sorunları ve önerilen çözümler
- Author
-
Kaya, Halil, Köymen, Kemal, Maltepe Üniversitesi, Fen Bilimleri Enstitüsü, Kaya, Halil, and Bilgisayar Mühendisliği Ana Bilim Dalı
- Subjects
Oracle architecture ,Veritabanı türleri ,Veritabanı güvenliği ,Database security ,Computer Engineering and Computer Science and Control ,Oracle veritabanı yapısı ,Database ,Database performance ,Database systems ,Uygulama güvenliği ,Veritabanı performansı ,Security ,Oracle database ,Type of database ,Oracle veritabanı ,DBA ,Bilgisayar Mühendisliği Bilimleri-Bilgisayar ve Kontrol ,Application security - Abstract
Bu tez çalışmasında veritabanında bahsedilmiş, veritabanı türleri ve günümüzde yaygın kullanılan ticari veritabanları incelenmiştir. En çok kullanılan oracle veritabanı üzerinden çalışan uygulamaların güvenlik ve performans sorunlarının tespit edilmesi ve çözüm önerilerinin yapan bir yazılım geliştirilmiştir. Geliştirilen yazılım, oracle veritabanı kullanan kullanıcıların gelecekle ilgili donanım ve kaynak planlamasının yapılmasına yardımcı olmaktır. Yazılımın içindeki örnek uygulamalar ve sorun çözme yöntemleriyle veritabanı yöneticilerinin bilgilerinin gelişmesine katkı sağlamaktadır. Aynı zamanda veritabanı ve uygulamalarının sağlıklı ve güvenli çalışmasına yardımcı olmaktadır. Tez 2011 yılında yapılmıştır ve 128 sayfadan oluşmaktadır., In the scope of this thesis, database concept and the types of databases are mentioned, widespreadly used commercial databases are examined. Since oracle database is the most used and widespreadly known database, a software application has been developed that discover, determine, make suggestions in order to solve the security and performance issues in the oracle databases. Developed software helps oracle database users with hardware and resource planning in the future. The best practices and samples in the software contributes significantly to expand database administrator's knowledge and ability to manage the database. This software application ensures database and their applications running healty and secure at the same time. This thesis has been completed in 2011 and it has 128 pages.
- Published
- 2011
35. Progress in chemistry modelling for vapour and aerosol transport analyses
- Author
-
M.H. Kaye, Martin Kissane, Paul K. Mason, Institut de Radioprotection et de Sûreté Nucléaire (IRSN), University of Ontario Institute of Technology (UOIT), and Thermo-Calc Software AB
- Subjects
Reactive gas ,Materials science ,Aerosol transport ,020209 energy ,Nuclear engineering ,Chemical phenomenas ,Mechanical engineering ,02 engineering and technology ,01 natural sciences ,010305 fluids & plasmas ,Database ,Nuclear reactors ,Vapour-phase ,Physical phenomena ,Germany ,Chemical reactions ,0103 physical sciences ,0202 electrical engineering, electronic engineering, information engineering ,Materials Chemistry ,Radioactive sources ,Chemical analysis ,Physical and Theoretical Chemistry ,Computer codes ,Atmospheric movements ,[PHYS]Physics [physics] ,Chemistry ,Metals and Alloys ,Data verification ,Condensed Matter Physics ,Atmospheric aerosols ,Aerosol ,Transfer ,Radioactivity ,Database systems ,13. Climate action ,Accidents ,Nuclear propulsion ,Systematic identification - Abstract
The ASTEC/Sophaeros computer code has been developed to model transport of reactive gases and aerosols in pipe flows with extensive coverage of chemical and physical phenomena. As part of the ASTEC severe-accident code co-developed by IRSN in France and GRS (Gesellschaft für Anlagen- und Reaktorsicherheit) in Germany, it is principally used to evaluate so-called radioactive source terms released to the environment in the event of potential accidents affecting water-cooled nuclear reactors. A major uncertainty in the past concerned the impact of chemistry on radioactivity transfer where the treatment of chemical phenomena was recognized as inadequate. Work described here, involving mainly systematic identification of relevant species and thermodynamic- data verification, has vastly improved this situation leading in particular to creation of a comprehensive database with fully-referenced sources of information. Further prospective improvements in this area are described.
- Published
- 2010
- Full Text
- View/download PDF
36. The BioPAX community standard for pathway data sharing
- Author
-
Margot Sunshine, Frank Schacherer, Nigam H. Shah, Akhilesh Pandey, Harsha Rajasimha, Andrew Finney, Rebecca Tang, Martijn P. van Iersel, Kumaran Kandasamy, Kei-Hoi Cheung, Martina Kutmon, Geeta Joshi-Tope, Matthias Samwald, Dean Ravenscroft, Mustafa H Syed, Vincent Schächter, Michael L. Blinov, Chris Sander, Liya Ren, Guanming Wu, Christian Lemer, Zhenjun Hu, Peter Hornbeck, Andrey Rzhetsky, Nicolas Le Novère, Emek Demir, Shiva Krupa, Michelle Whirl-Carrillo, Ken Fukuda, Alejandra López-Fuentes, Michael P. Cary, Erik Brauner, David Merberg, Julie Leonard, Imran Shah, David W. Kane, Alexander R. Pico, Shannon K. McWeeney, Michael Hucka, Peter D. Karp, Nadia Anwar, Andrea Splendiani, Peter D'Eustachio, Olivier Hubaut, Ugur Dogrusoz, Julio Collado-Vides, Gary D. Bader, Jeremy Zucker, Carl F. Schaefer, Keith Allen, Kam D. Dahlquist, Oliver Reubenacker, Paul Thomas, Mirit I. Aladjem, Victoria Petri, Verónica Jiménez-Jacinto, Igor Rodchenkov, Edgar Wingender, Gopal R. Gopinath, Imre Vastrik, Stan Letovksy, Susumu Goto, Ryan Whaley, Frank Gibbons, Natalia Maltsev, Özgün Babur, Ranjani Ramakrishnan, Robin Haw, Elgar Pichler, Burk Braun, Sylva L. Donaldson, Suzanne M. Paley, Huaiyu Mi, Sarala M. Wimalaratne, Elizabeth M. Glass, Sasha Tkachev, Irma Martínez-Flores, Augustin Luna, Joanne S. Luciano, Debbie Marks, Marc Gillespie, Michael Honig, Ewan Birney, Dan Corwin, Bruno S. Sobral, Kenneth H. Buetow, Li Gong, Eric K. Neumann, Robert N. Goldberg, Peter Murray-Rust, Demir, Emek, Babur, Özgün, Doğrusöz, Uğur, Bioinformatica, and RS: NUTRIM - R4 - Gene-environment interaction
- Subjects
Signaling pathways ,interaction network ,representation ,Molecular biology ,WikiPathways : Pathways for the people ,Computer science ,Biological pathways ,Review ,Signal transduction ,Biological pathway exchange ,Bioinformatics ,Applied Microbiology and Biotechnology ,information ,Computational biology ,ConsensusPathDB ,information system ,0302 clinical medicine ,Databases as topic ,pathway database ,ontology ,Visualization ,Priority journal ,0303 health sciences ,Genetic interaction ,Messenger RNA ,software environment ,Systems Biology Graphical Notation ,Promoter region ,systems biology ,Molecular interaction ,Programming languages ,Semantics ,Enzyme substrate ,Protein modification ,Databases as Topic ,030220 oncology & carcinogenesis ,Cellular levels ,Molecular Medicine ,Metabolic Networks and Pathways ,Signal Transduction ,standard exchange format ,Biotechnology ,Biomedical Engineering ,Bioengineering ,pathway data integration ,Molecular dynamics ,Structure analysis ,Article ,Database ,Biological pathway ,03 medical and health sciences ,Data visualization ,Protein kinase B ,Fragmentation reaction ,BioPAX : Biological Pathways Exchange ,biological pathways ,Rapid growth ,Computational tools ,Gene regulation network ,030304 developmental biology ,Electronic data interchange ,Protein DNA interaction ,Copy number variation ,Information Dissemination ,business.industry ,Information dissemination ,Computational Biology ,knowledgebase ,Single nucleotide polymorphism ,Data sharing ,Metabolism ,Community standards ,Database systems ,Metabolic networks and pathways ,Protein protein interaction ,Protein structure ,Molecular evolution ,Protein expression ,cellular pathways ,Biological discoveries ,Programming Languages ,collaborative construction ,business ,Software - Abstract
BioPAX (Biological Pathway Exchange) is a standard language to represent biological pathways at the molecular and cellular level. Its major use is to facilitate the exchange of pathway data (http://www.biopax.org). Pathway data captures our understanding of biological processes, but its rapid growth necessitates development of databases and computational tools to aid interpretation. However, the current fragmentation of pathway information across many databases with incompatible formats presents barriers to its effective use. BioPAX solves this problem by making pathway data substantially easier to collect, index, interpret and share. BioPAX can represent metabolic and signaling pathways, molecular and genetic interactions and gene regulation networks. BioPAX was created through a community process. Through BioPAX, millions of interactions organized into thousands of pathways across many organisms, from a growing number of sources, are available. Thus, large amounts of pathway data are available in a computable form to support visualization, analysis and biological discovery.
- Published
- 2010
37. Sayısal görüntü arşivi bilgi sistemi
- Author
-
Demirel, Ahmet Şamil, Yıldız, Ferruh, Enstitüler, Fen Bilimleri Enstitüsü, Jeodezi ve Fotogrametri Mühendisliği Ana Bilim Dalı, and Jeodezi ve Fotogrametri Mühendisliği Anabilim Dalı
- Subjects
Veri tabanı sistemleri ,Database ,Database management ,Veri tabanı yönetimi ,Database systems ,Jeodezi ve Fotogrametri ,Archiving system ,Veri tabanı ,Image ,Arşivleme sistemi ,Görüntü ,Geodesy and Photogrammetry - Abstract
Gelişen teknolojiye paralel olarak sayısal görüntülerin üretimi ve kullanımında büyük bir artış olmuştur. Bu artış verilerin uygun ortamlarda arşivlenmesi, yönetimi ve ihtiyaç duyulduğunda kısa surede ulaşılarak kullanıma sunulması ile ilgili sorunları da beraberinde getirmiştir. Bu sorunların en kolay yoldan çözümü bu amaç için ticari olarak üretilmiş arşiv ve yönetim yazılımları kullanmaktır. Ancak bu yöntemin maliyeti de bu oranda yüksek olmaktadır. Sorunun bir diğer çözümü ise doğrudan amaca yönelik olarak ve ücretsiz, açık kaynak kodlu yazılımlar kullanarak geliştirilecek yeni bir yazılım üretmektir. Bu çalışmada bu yöntem denenmiş ve sayısal görüntülerin bir düzen içerisinde bir veri tabanında bilgileri tutularak, gerektiğinde bu verilere kolayca ulaşılması sağlanmıştır. Geliştirilen yazılım tamamen ücretsiz programlama dili yazılımları ve veri tabanı yönetim yazılımları kullanılarak ortaya çıkarılmıştır. Temel olarak hava fotoğrafları, uydu görüntüleri, sayısal Ortofoto ve Ortofoto haritaları, Sayısal Arazi Modellerini ve sayısal haritaları ele almıştır. Bu veriler ile ilgili bilgilerin girilmesi, düzenlenmesi ve sorgulanması ile ilgili bölümler oluşturulmuştur. Ortaya çıkan yazılım, sıfır maliyette, milli, ve gelişime açık bir yazılımdır. Ayrıca benzer özellik taşıyan birçok ticari yazılımın ortak özelliklerini taşımaktadır., In parallel to developing technology, production and use of digital images has become a big increase. Along with this increase, data archiving, management and use problems were observed. . The easiest way to solve these problems is to use archive and management software that commercially produced for this purpose. However, the cost of this method is a little higher. Another solution is using free, open source software. In this study we tested this method. Information will be kept in a database belonging to the digital images, easy access to this data to be provided when necessary. This software is developed with free programming language, and free data base management software was used. Basically, aerial photographs, satellite images, digital Orthophoto and Orthophoto maps, digital maps, Digital Terrain Model were used. Data entering, editing and querying sections have been set up. Program has been developed with zero cost, and is fully national. Also includes common features of many commercial software.
- Published
- 2010
38. Machine learning proposed approach for detecting database intrusions in RBAC enabled databases
- Author
-
Dhiren Patel, Udai Pratap Rao, G. J. Sahani, and Presented at the 2010 International Conference on Computing Communication and Networking Technologies (ICCCNT)
- Subjects
Database ,Computer science ,InformationSystems_DATABASEMANAGEMENT ,Intrusion detection system ,Computer security ,computer.software_genre ,Asset (computer security) ,Social Security number ,Database testing ,Correlation ,Credit card ,Database systems ,Role-based access control ,Intrusion detection ,Database security ,Data mining ,Database transaction ,computer ,Probability - Abstract
Information is valuable asset of any organization which is stored in databases. Data in such databases may contain credit card numbers, social security number or personal medical records etc. Failing to protect these databases from intrusions will result in loss of customer's confidence and might even result in lawsuits. Traditional database security mechanism does not design to detect anomalous behavior of database users. There are number of approaches to detect intrusions in network. But they cannot detect intrusions in database. There have been very few ID mechanisms specifically tailored to database systems. We propose transaction level approach to detect malicious behavior in database systems enabled with Role Based Access Control (RBAC) mechanism., by Udai Pratap Rao, G. J. Sahani and Dhiren R. Patel
- Published
- 2010
- Full Text
- View/download PDF
39. PROM-OOGLE
- Author
-
Dean Cheng, Marcelo Marcet-Palacios, Osmar R. Zaïane, and John Sheldon
- Subjects
Focus (computing) ,Text mining ,Database ,Computer science ,business.industry ,Process (engineering) ,Promoter ,computer.software_genre ,Medical research ,Data science ,Task (project management) ,Resource (project management) ,Database systems ,Gene promoter ,Web application ,Data integration ,Transcriptor factor ,Data mining ,Line (text file) ,business ,computer ,Gene - Abstract
Technical report TR08-13. The vast number of on-line biological and medical databases available can be a great resource for medical researchers. However, the different types of data and interfaces available can be overwhelming for many medical researchers to learn. Moreover, the available resources lack needed integration. Here we focus on an important task in medical research: to provide researchers with promoter analysis for a given gene. PROM-OOGLE is a web based data mining tool that provides a means for researchers to take a gene name of interest and obtain its promoter sequence in return after automatic integration of text databases. Additionally, the program is capable of returning multiple promoters from different genes allowing researchers to study how promoters regulate genes. This tool facilitates the process of acquiring information on a promoter and may lead to interesting discoveries. | TRID-ID TR08-13
- Published
- 2010
- Full Text
- View/download PDF
40. Dağıtık veri tabanı sistemlerinde optimizasyon süreçleri ve uygulamaları
- Author
-
Şengül, Ayşe Yasemin, Orkan, Ahmet Lütfi, and İletişim Bilimleri Anabilim Dalı
- Subjects
Optimization ,Database ,Information management ,Information society ,Database systems ,Distributed database systems ,İletişim Bilimleri ,Communication Sciences ,Computer Engineering and Computer Science and Control ,Bilgisayar Mühendisliği Bilimleri-Bilgisayar ve Kontrol - Abstract
Çalışmada dağıtık veri tabanı sistemleri gelişim süreci ve bu gelişimi etkileyen faktörler incelenmiştir. Dağıtık veri tabanı sistemleri sadece bilgi ve iletişim teknolojileri ile değil sosyal ve kültürel birçok faktör ile etkileşim halinde olmuştur. Bilgi toplumuna geçilmesi, bununla beraber küreselleşme eğiliminin başlaması bilgi ve iletişim teknolojilerindeki birçok gelişmeyi tetiklemiştir. Bu süreçte öne çıkan diğer kavram ise bilgi yönetimidir. Bunların gelişmelerin veri tabanı teknolojileri kullanılarak veriden bilgi çıkarımı ve elde edilen bu bilgilerin doğru zamanda doğru yerde kullanılması organizasyonların rekabet güçlerini arttırmada büyük rol oynamaya başlamıştır.Dağıtık veri tabanı sistemleri ve optimizasyonu için gerekli süreçler tespit edilip incelendikten sonra sağlık sektöründe veri tabanı sistemleri ve dağıtık veri tabanı sistemleri ile ilgili bir araştırma çalışması gerçekleştirilmiştir. Araştırmanın sonucunda sağlık sektöründe kullanılan bilgi ve iletişim teknolojilerinin diğer sektörlerde olduğundan daha geç geliştiği gözlemlenmiştir. Bunun yanında sürekli artan insan nüfusu ile hizmet verebilecek doktor sayısının arasındaki farkın büyümesi sonucu bilgi paylaşımının daha da büyük önem taşımaya başladığı ve bu sebeple araştırmacılar farklı branşlarda sağlık sektöründe tutulan verilerin entegre edilmesiyle ilgilenmişlerdir.Bu çalışmanın amacı araştırmanın yapıldığı merkezde kurulması planlanan veri tabanı sisteminin merkezi veya dağıtık olma durumunda sağlayacağı avantajlar ve dezavantajların belirlenmesidir. Uygulama sonucunda literatürde daha önce dağıtık veri tabanı sistemleri optimizasyonu ile ilgili gerçekleştirilmemiş olan merkezi ve dağıtık veri tabanı sistemi performans karşılaştırması elde edilmiştir. Çalışmada yapılan uygulamada bir tüp bebek merkezi için dağıtık veri tabanı sistemi oluşturulmuş ve bu sistem merkezi veri tabanı sistemiyle karşılaştırılmıştır. Karşılaştırma merkezde kullanılan uygulamalara alınan yanıt süreleri arasında yapılmıştır. In this study development of distributed database systems and the factors efecting this development has been analyzed. Distributed database systems are not only related with information and communication technologies but they are also in interaction with varying social and cultural factors. Going towards information society and beginning of globalization paradigm triggered varying developments in information and communication technologies. Another important concept among these is knowledge management. As a result of these knowledge discovery form stored data, and using these knowledge at the right time at the right place using database technologies, increased companies competitive advantage.After identifying the necessary processes for optimization a survey has been conducted in healthcare industry. As a result survey it has been observed that information and communication technologies have been developed slowly in healthcare industry then the other industries. Also, because of increasing difference between the population and the medical doctors to serve patients, researchers have been interested in integration of data stored in healthcare industry.The aim of this study is to compare the advantages and disadvantages of distributed and central data base systems with an application in the center where the survey has been conducted. As a result the performance comparison between distributed and centralized data base systems, which has not been measured before have been achieved. In this application a distributed data base system for an infertility clinic have been designed and compared with a centralized one. The comparison has been done between response times of applications used. 261
- Published
- 2010
41. Cimbrian as a Test Case for Synchronic and Diachronic Language Variation: a Conceptual Approach for the Information Space
- Author
-
Maristella Agosti, Alber, B., Giorgio Maria Di Nunzio, Dussin, M., Rabanus, S., and Tomaselli, A.
- Subjects
Linguistic Analysis ,Part of Speech Tagging ,Database ,Dialettologia ,Cimbro ,Informatica ,Multilingual Information Retrieval ,Database Systems ,User Interfaces - Published
- 2010
42. Consumer Databases, Neoliberalism, and the Commercial Mediation of Identity: A Medium Theory Analysis
- Author
-
Sandra Smeltzer and Vincent Manzerolle
- Subjects
Surveillance ,Database ,Consumption ,media_common.quotation_subject ,Communication ,Medium Theory ,Neoliberalism ,Capitalism ,computer.software_genre ,Political Economy ,Urban Studies ,Sovereignty ,Mediation ,Consumer sovereignty ,Harold Innis ,Consumer capitalism ,Sociology ,Neutrality ,Free market ,Database Systems ,Safety Research ,computer ,media_common - Abstract
This paper argues that the systemic nature of contemporary consumer surveillance undermines the most fundamental principle of free market economics: consumer sovereignty. Specifically, this paper argues that the rise of an ‘information’ or ‘knowledge’ society in conjunction with neoliberal capitalism has entrenched routine forms of surveillance within commercial strategies by employing networked databases as a primary medium for the articulation of consumer sovereignty (choice/demand). The communicative relationship between consumers and producers within the market involves effectively ‘listening’ (and then responding) to consumer needs and wants in a timely manner. Surveillance is therefore not only necessary for the operation of globalized consumer capitalism, it is also the primary means by which consumer communicate their sovereignty within the marketplace. By turning to the work of Harold Innis and the intellectual tradition known as medium theory, this paper will theorize how in linking the actions of individual consumers to the decision-making capacities of trans-national corporations (TNC), the prevalence of consumer databases violates the fundamental neutrality of the market, and thus sovereignty, of individual consumers. In sum, by treating the database as a distinct communication medium, this paper will highlight how the commercial mediation of identity under neoliberalism can conceal the potential for the uneven geographic development, the marginalization of ‘less valuable’ consumer segments, and the exploitation of individual vulnerabilities through behavior and profile modeling.
- Published
- 2010
43. Query evaluation techniques for cluster database systems
- Author
-
Andrey V. Lepikhov and Leonid B. Sokolinsky
- Subjects
Query processing ,Parallel Database ,Computer science ,View ,Data allocation ,Load balancing methods ,Computer supported cooperative work ,computer.software_genre ,Query optimization ,Query language ,Database tuning ,Query expansion ,Information systems ,Query by Example ,Computer Science::Databases ,computer.programming_language ,Database ,Parallel database ,Parallel processing systems ,Parallel query processing ,Online aggregation ,Query evaluation ,Data skew ,Database systems ,Data mining ,Management information systems ,computer ,Cluster systems ,Cluster computing - Abstract
The paper is dedicated to a problem of effective query processing in cluster database systems. An original approach to data allocation and replication at nodes of a cluster system is presented. On the basis of this approach the load balancing method is developed. Also, we propose a new method for parallel query processing on the cluster systems. All described methods have been implemented in "Omega" parallel database management system prototype. Our experiments show that "Omega" system demonstrates nearly linear scalability even in presence of data skew. © 2010 Springer-Verlag.
- Published
- 2010
44. Automated SQL tuning through trial and (sometimes) error
- Author
-
Shivnath Babu and Herodotos Herodotou
- Subjects
Autocommit ,Database ,Computer science ,Data definition language ,InformationSystems_DATABASEMANAGEMENT ,Electrical Engineering - Electronic Engineering - Information Engineering ,PL/SQL ,computer.software_genre ,Medical and Health Sciences ,Database tuning ,Query plan ,Database systems ,SQL tuning ,Query by Example ,Stored procedure ,SQL/PSM ,computer ,database query ,computer.programming_language - Abstract
SQL tuning - the attempt to improve a poorly-performing execution plan produced by the database query optimizer - is a critical aspect of database performance tuning. Ironically, as commercial databases strive to improve on the manageability front, SQL tuning is becoming more of a black art. It requires a high level of expertise in areas like (i) query optimization, run-time execution of query plan operators, configuration parameter settings, and other database internals; (ii) identification of missing indexes and other access structures; (iii) statistics maintained about the data; and (iv) characteristics of the underlying storage system. Since database systems, their workloads, and the data that they manage are not getting any simpler, database users and administrators often rely on trial and error for SQL tuning. In this paper, we take the position that the trial-and-error (or, experiment-driven) process of SQL tuning can be automated by the database system in an e cient manner; freeing the user or administrator from this burden in most cases. A number of current approaches to SQL tuning indeed take an experiment-driven approach. We are prototyping a tool, called zTuned, that automates experiment-driven SQL tuning. This paper describes the design choices in zTuned to address three nontrivial issues: (i) how is the SQL tuning logic integrated with the regular query optimizer, (ii) how to plan the experiments to conduct so that a satisfactory (new) plan can be found quickly, and (iii) how to conduct experiments with minimal impact on the user-facing production workload. We conclude with a preliminary empirical evaluation and outline promising new directions in automated SQL tuning. Copyright 2009 ACM.
- Published
- 2009
- Full Text
- View/download PDF
45. An association rule-based recommendation engine for online dating sites
- Author
-
Özseyhan, Yağiz Civan, Badur, Bertan, Darcan, Osman, and Yönetim Bilişim Sistemleri Anabilim Dalı
- Subjects
Database ,Internet ,Bilim ve Teknoloji ,Database systems ,İşletme ,Friendship ,Message ,Science and Technology ,Data mining ,Social networks ,Business Administration - Abstract
Bu çalışmada Türkiye'den bir arkadaşlık sitesinin veritabanını incelenerek, site üzerinden tanışmış çiftlerin sahip olduğu ortak özellikler araştırılmıştır.Veri Madenciliği araçlarından biri olan ?Association Rule Mining? `ten yararlanılarak, mesajlaşma ve kullanıcı verilerinden bir kural seti türetilmiş; bu kurallar da kullanıcının kendisine en uygun adayları listeleyebileceği bir öneri sisteminin geliştirilmesinde kullanılmıştır.Öneri sisteminin üretilen sonuçların kalitesinde yaptığı artış istatistik araçlarıyla test edilmiş, verilerin anlamlı olduğu sonucuna varılmıştır. In this study, the database of a Turkish online dating site is analyzed to reveal patterns in the personal features of matching couples.By using a functionality of data Mining, the Association Rule mining, a set of rules is extracted from the available messaging and user data. The rules are used for the development of a recommendation engine which is capable to give list of best potential matches to the site user.The performance of the recommendation engine is tested with statistical tools to find whether the increase in the quality of results is significant. 104
- Published
- 2009
46. A model of authorization for next-generation database systems
- Author
-
Fausto Rabitti, Darrell Woelk, Elisa Bertino, and Won Kim
- Subjects
Object-oriented programming ,Database ,Computer science ,Systems ,Authorization ,Semantic data modeling ,Next-Generation Database Systems ,computer.software_genre ,Key issues ,Data modeling ,Object oriented databases ,Next-Generation ,Database Systems ,User interface ,computer ,Model ,Information Systems - Abstract
The conventional models of authorization have been designed for database systems supporting the hierarchical, network, and relational models of data. However, these models are not adequate for next-generation database systems that support richer data models that include object-oriented concepts and semantic data modeling concepts. Rabitti, Woelk, and Kim [14] presented a preliminary model of authorization for use as the basis of an authorization mechanism in such database systems. In this paper we present a fuller model of authorization that fills a few major gaps that the conventional models of authorization cannot fill for next-generation database systems. We also further formalize the notion of implicit authorization and refine the application of the notion of implicit authorization to object-oriented and semantic modeling concepts. We also describe a user interface for using the model of authorization and consider key issues in implementing the authorization model. © 1991, ACM. All rights reserved.
- Published
- 1991
- Full Text
- View/download PDF
47. Robust Runtime Optimization of Data Transfer in Queries over Web Services
- Author
-
Gounaris, A., Yfoulis, C., Sakellariou, R., Dikaiakos, Marios D., and Dikaiakos, Marios D. [0000-0002-4350-6058]
- Subjects
Optimization ,Technology ,Computer science ,Distributed computing ,Real environments ,Robust control ,Run-time ,Distributed applications ,On-line system identification ,Computer systems ,Data engineering ,computer.software_genre ,Application software ,Bottleneck ,Control theory ,Self-managing ,Robustness (computer science) ,Data transfer cost ,Information services ,Run-time optimization ,Web services ,Distributed database ,Database ,Database querying ,Data transfer ,World Wide Web ,Solutions ,Block sizes ,Database systems ,Database community ,Web service ,International conferences ,computer ,Static solutions - Abstract
Self-managing solutions have recently attracted a lot of interest from the database community. The need for self-* properties is more evident in distributed applications comprising heterogeneous and autonomous databases and functionality providers. Such resources are typically exposed as Web Services (WSs), which encapsulate remote DBMSs and functions called from within database queries. In this setting, database queries are over WSs, and the data transfer cost becomes the main bottleneck. To reduce this cost, data is shipped to and from WSs in chunks however the optimum chunk size is volatile, depending on both the resources' runtime properties and the query. In this paper we propose a robust control theoretical solution to the problem of optimizing the data transfer in queries over WSs, by continuously tuning at runtime the block size and thus tracking the optimum point. Also, we develop online system identification mechanisms that are capable of estimating the optimum block size analytically. Both contributions are evaluated via both empirical experimentation in a real environment and simulations, and have been proved to be more effective and efficient than static solutions. © 2008 IEEE. 596 605 Conference code: 73145 Cited By :23
- Published
- 2008
- Full Text
- View/download PDF
48. Safespot local dynamic maps : context-dependent view generation of a platform's state & environment
- Subjects
Informatics ,Industrial Innovation ,Advanced applications ,Traffic control ,Safety systems ,Context awareness ,Context- awareness ,Spatio-temporal scale ,Automobile drivers ,Cooperative systems ,Co-operative systems ,Database ,Driver assistance ,Database systems ,Security systems ,Safety applications ,Context dependent ,Digital maps ,Digital map - Abstract
Increased availability of sensing, communication and computing equipment in modern vehicles presents opportunities for a generational advance in driver assistance and safety functions. However, the growth in the volume and complexity of data also presents challenges for its collection, maintenance and distribution. The SAFESPOT Local Dynamic Map (LDM) contains a structured model of the world. It acts as an interface between sensing modules and the safety applications. It enables applications to extract context-dependent views of the environment and the platform's state. These unique views differ in spatio-temporal scale, as well as the objects, attributes and relationships they contain. The ability to extract various views from a LDM is seen as an important tool in the development of multiple, advanced applications on a single platform.
- Published
- 2008
49. A clustering-based prefetching scheme on a Web cache environment
- Author
-
Pallis, George C., Vakali, Athena I., Pokorný, Jaroslav, Pallis, George C. [0000-0003-1815-5468], and Vakali, Athena I. [0000-0002-0666-6984]
- Subjects
Users' access patterns ,Web server ,User interfaces ,General Computer Science ,Clustering algorithms ,Computer science ,Network management ,Spatial locality ,computer.software_genre ,Web cache ,Web page ,Web navigation ,Access control ,Web caching ,Electrical and Electronic Engineering ,Data mining ,Data Web ,Web services ,Database ,Web data clustering ,Clustering based prefetching ,Users access patterns ,World Wide Web ,Graph theory ,Web prefetching ,Web mining ,Database systems ,Control and Systems Engineering ,Web infrastructure ,Web mapping ,Web service ,computer - Abstract
Web prefetching is an attractive solution to reduce the network resources consumed by Web services as well as the access latencies perceived by Web users. Unlike Web caching, which exploits the temporal locality, Web prefetching utilizes the spatial locality of Web objects. Specifically, Web prefetching fetches objects that are likely to be accessed in the near future and stores them in advance. In this context, a sophisticated combination of these two techniques may cause significant improvements on the performance of the Web infrastructure. Considering that there have been several caching policies proposed in the past, the challenge is to extend them by using data mining techniques. In this paper, we present a clustering-based prefetching scheme where a graph-based clustering algorithm identifies clusters of "correlated" Web pages based on the users' access patterns. This scheme can be integrated easily into a Web proxy server, improving its performance. Through a simulation environment, using a real data set, we show that the proposed integrated framework is robust and effective in improving the performance of the Web caching environment. © 2007 Elsevier Ltd. All rights reserved. 34 4 309 323 Cited By :52
- Published
- 2008
50. Automatic Detection Of Salient Objects And Spatial Relations In Videos For A Video Database System
- Author
-
Tarkan Sevilmiş, Muhammet Bastan, Özgür Ulusoy, Uğur Güdükbay, Ulusoy, Özgür, and Güdükbay, Uğur
- Subjects
Computer science ,Feature extraction ,Storage ,Knowledge-base Construction, Spatio-temporal Queries ,Extraction ,computer.software_genre ,Information retrieval ,Information systems ,Video databases ,Computer vision ,Data storage ,Salient Object Detection And Tracking ,Camera Focus Estimation ,Database ,Spatio-temporal queries ,Labels ,business.industry ,Search engine indexing ,Frame (networking) ,Process (computing) ,Video processing ,Cameras ,Object (computer science) ,Management ,Video signal processing ,Data processing ,Video recording ,Spatial relation ,Data storage equipment ,Video database management ,Administrative data processing ,Database systems ,Knowledge-base construction ,Video tracking ,Multimedia Databases ,Signal Processing ,Automatic detection ,Indexing ,Management information systems ,Computer Vision and Pattern Recognition ,Artificial intelligence ,Object Labeling ,business ,computer ,Spatio-temporal relations - Abstract
Cataloged from PDF version of article. Multimedia databases have gained popularity due to rapidly growing quantities of multimedia data and the need to perform efficient indexing, retrieval and analysis of this data. One downside of multimedia databases is the necessity to process the data for feature extraction and labeling prior to storage and querying. Huge amount of data makes it impossible to complete this task manually. We propose a tool for the automatic detection and tracking of salient objects, and derivation of spatio-temporal relations between them in video. Our system aims to reduce the work for manual selection and labeling of objects significantly by detecting and tracking the salient objects, and hence, requiring to enter the label for each object only once within each shot instead of specifying the labels for each object in every frame they appear. This is also required as a first step in a fully-automatic video database management system in which the labeling should also be done automatically. The proposed framework covers a scalable architecture for video processing and stages of shot boundary detection, salient object detection and tracking, and knowledge-base construction for effective spatio-temporal object querying. (c) 2008 Elsevier B.V. All rights reserved.
- Published
- 2008
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.