106 results on '"parallel data processing"'
Search Results
2. libxtc: an efficient library for reading XTC-compressed MD trajectory data
- Author
-
Nikolay A. Krylov and Roman G. Efremov
- Subjects
Molecular dynamics ,Biomolecular simulations ,Parallel data processing ,Efficiency of MD trajectories reading ,Medicine ,Biology (General) ,QH301-705.5 ,Science (General) ,Q1-390 - Abstract
Abstract Objective The purpose of this work is to optimize the processing of molecular dynamics (MD) trajectory data obtained for large biomolecular systems. Two popular software tools were chosen as the reference: the tng and the xdrfile libraries. Current implementation of tng algorithms and library is either fast or storage efficient and xdrfile is storage efficient but slow. Our aim was to combine speed and storage efficiency through the xdrfile’s code modification. Results Here we present libxtc, a ready-to-use library for reading MD trajectory files in xtc format. The effectiveness of libxtc is demonstrated for several biomolecular systems of various sizes (~ 2 × 104 to ~ 2 × 105 atoms). In sequential mode, the performance of libxtc is up to 1.8 times higher and 1.4 times lower than xdrfile and tng, respectively. In parallel mode, libxtc is about 3 and 1.3 times faster than xdrfile and tng. At the same time, MD data stored in the xtc format require about 1.3 times less disk space than those treated with the tng algorithm in the fastest reading mode, which is a noticeable saving especially when the MD trajectory is long and the number of atoms is large—this applies to most biologically relevant systems.
- Published
- 2021
- Full Text
- View/download PDF
3. Threshold-Variation-Tolerant Coupling-Gate α-IGZO Synaptic Transistor for More Reliably Controllable Hardware Neuromorphic System
- Author
-
Dongyeon Kang, Jun Tae Jang, Shinyoung Park, Md. Hasan Raza Ansari, Jong-Ho Bae, Sung-Jin Choi, Dong Myong Kim, Changwook Kim, Seongjae Cho, and Dae Hwan Kim
- Subjects
Hardware-oriented neuromorphic computing ,parallel data processing ,energy efficiency ,synaptic transistor ,amorphous indium-gallium-zinc-oxide ,extended gate ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Hardware-oriented neuromorphic computing is gaining great deal of interest for highly parallel data processing and superb energy efficiency, as the candidate for replacement of conventional von Neumann computing. In this work, a novel synaptic transistor constructing the neuromorphic system is proposed, fabricated, and characterized. Amorphous indium-gallium-zinc-oxide ( $\alpha $ -IGZO) and Al2O3 are introduced as the channel and gate dielectric materials, respectively. Along with the high functionality and low-temperature processing viability, geometric peculiarity featuring extended gate structure improves the performances of the proposed transistor as synaptic component in the neuromorphic system. The insight into the substantial effect of optimal device structure design on energy efficiency is highlighted.
- Published
- 2021
- Full Text
- View/download PDF
4. libxtc: an efficient library for reading XTC-compressed MD trajectory data.
- Author
-
Krylov, Nikolay A. and Efremov, Roman G.
- Subjects
- *
MOLECULAR dynamics , *SOFTWARE development tools , *READING , *LIBRARIES - Abstract
Objective: The purpose of this work is to optimize the processing of molecular dynamics (MD) trajectory data obtained for large biomolecular systems. Two popular software tools were chosen as the reference: the tng and the xdrfile libraries. Current implementation of tng algorithms and library is either fast or storage efficient and xdrfile is storage efficient but slow. Our aim was to combine speed and storage efficiency through the xdrfile's code modification. Results: Here we present libxtc, a ready-to-use library for reading MD trajectory files in xtc format. The effectiveness of libxtc is demonstrated for several biomolecular systems of various sizes (~ 2 × 104 to ~ 2 × 105 atoms). In sequential mode, the performance of libxtc is up to 1.8 times higher and 1.4 times lower than xdrfile and tng, respectively. In parallel mode, libxtc is about 3 and 1.3 times faster than xdrfile and tng. At the same time, MD data stored in the xtc format require about 1.3 times less disk space than those treated with the tng algorithm in the fastest reading mode, which is a noticeable saving especially when the MD trajectory is long and the number of atoms is large—this applies to most biologically relevant systems. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
5. Technology of Real-World Analyzers (TAUR) and its practical application.
- Author
-
Jung, J., Kiełbik, R., Hałagan, K., Polanowski, P., and Sikorski, A.
- Subjects
ADAPTIVE computing systems ,FIELD programmable gate arrays ,DATA transmission systems ,MOLECULAR dynamics ,SCALABILITY - Abstract
The article describes the most important details of the project for reconfigurable construction of dedicated electronic machines intended for performing analyses of phenomena that occur in multi-component systems containing at least several million mutually interacting elements. Devices built in the presented technology can be characterized by the use of reconfigurable integrated circuits, spatial construction ensuring scalability, a redundant panel system as well as specially developed data transmission and work control systems. Machines work in a parallel manner and can solve problems in various fields of science and technology by competing with the speed of data processing with the latest supercomputing systems. As an example, we present details of the ARUZ machine containing 26,000 FPGAs, which was made using this technology. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
6. Experimental Realization of a Passive Gigahertz Frequency‐Division Demultiplexer for Magnonic Logic Networks.
- Author
-
Heussner, Frank, Talmelli, Giacomo, Geilen, Moritz, Heinz, Björn, Brächer, Thomas, Meyer, Thomas, Ciubotaru, Florin, Adelmann, Christoph, Yamamoto, Kei, Serga, Alexander A., Hillebrands, Burkard, and Pirro, Philipp
- Subjects
- *
MAGNONS , *SPIN waves , *SIGNAL separation , *MAGNETIC films , *PARALLEL processing , *LOGIC , *ELECTRONIC data processing - Abstract
The emerging field of magnonics uses spin waves and their quanta, magnons, to implement wave‐based computing on the micro‐ and nanoscale. Multifrequency magnon networks would allow for parallel data processing within single logic elements, whereas this is not the case with conventional transistor‐based electronic logic. However, a lack of experimentally proven solutions to efficiently combine and separate magnons of different frequencies has impeded the intensive use of this concept. Herein, the experimental realization of a spin‐wave demultiplexer enabling frequency‐dependent separation of magnonic signals in the gigahertz range is demonstrated. The device is based on 2D magnon transport in the form of spin‐wave beams in unpatterned magnetic films. The intrinsic frequency dependence of the beam direction is exploited to realize a passive functioning obviating an external control and additional power consumption. This approach paves the way to magnonic multiplexing circuits enabling simultaneous information transport and processing. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
7. P4: Portable Parallel Processing Pipelines for Interactive Information Visualization.
- Author
-
Li, Jianping Kelvin and Ma, Kwan-Liu
- Subjects
DATA visualization ,PARALLEL processing ,GRAPHICS processing units ,ELECTRONIC data processing ,PIPELINES ,DATA modeling - Abstract
We present P4, an information visualization toolkit that combines declarative design specification and GPU computing for building high-performance interactive systems. Most of the existing information visualization toolkits do not harness the power of parallel processors in today's mainstream computers. P4 leverages GPU computing to accelerate both data processing and visualization rendering for interactive visualization applications. P4's programming interface offers a declarative visualization grammar for rapid specifications of data transformations, visual encodings, and interactions. By simplifying the development of GPU-accelerated visualization systems while supporting a high degree of flexibility and customization for design specification, P4 narrows the gap between expressiveness and scalability in information visualization toolkits. Through a range of examples and benchmark tests, we demonstrate that P4 provides high efficiency for creating interactive visualizations and offers drastic performance improvement over current state-of-the-art toolkits. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
8. Neuroprocessor Automatic Control System of the Module SEMS
- Author
-
Tarasova, I. L., Gorodetskiy, A. E., Kurbanov, V. G., Kacprzyk, Janusz, Series editor, and Gorodetskiy, Andrey E., editor
- Published
- 2016
- Full Text
- View/download PDF
9. Improvising and Optimizing Resource Utilization in Big Data Processing
- Author
-
Praveen Kumar, Rathore, Vijay Singh, Kacprzyk, Janusz, Series editor, Pant, Millie, editor, Deep, Kusum, editor, Bansal, Jagdish Chand, editor, Nagar, Atulya, editor, and Das, Kedar Nath, editor
- Published
- 2016
- Full Text
- View/download PDF
10. Data Flow Processing Framework for Multimodal Data Environment Software
- Author
-
Janiak, Mateusz, Kulbacki, Marek, Knieć, Wojciech, Nowacki, Jerzy Paweł, Drabik, Aldona, Kacprzyk, Janusz, Series editor, Barbucha, Dariusz, editor, Nguyen, Ngoc Thanh, editor, and Batubara, John, editor
- Published
- 2015
- Full Text
- View/download PDF
11. Optimized and Parallel Query Processing in Similarity-Based Databases
- Author
-
Krajča, Petr, Goebel, Randy, Series editor, Tanaka, Yuzuru, Series editor, Wahlster, Wolfgang, Series editor, Torra, Vicenc, editor, and Narukawa, Torra, editor
- Published
- 2015
- Full Text
- View/download PDF
12. The Problem with Data
- Author
-
Frampton, Michael and Frampton, Michael
- Published
- 2015
- Full Text
- View/download PDF
13. Increasing the performance of the software framework for implementing the algorithms of the group method of data handling
- Author
-
A. A. Orlov
- Subjects
software framework ,object-oriented analysis and design ,parallel data processing ,efficiency of paralleling ,group method of data handling ,Engineering geology. Rock mechanics. Soil mechanics. Underground construction ,TA703-712 - Abstract
In previous works the author has proposed a universal software framework that allows implementing the known algorithms of group method of data handling, model bases, training methods and model selection criteria. This paper introduces the solution of a topical problem of increasing the performance of the framework. Based on the review of existing computing architectures for parallel data processing and software systems for inductive modeling supporting parallel computations the author has worked out the requirements for the subsystems of parallel computing and memory management of the software framework. Using the methodology of object-oriented analysis and design the author developed the object-oriented structure of these subsystems and introduced the specifics of their operation on each of the mentioned computing architectures. The performance of the parallel implementation of the combinatorial group method of data handling algorithm on basis of the software framework was evaluated experimentally for multi-core processors.
- Published
- 2019
14. An Energy-Efficient, Parallel Neighborhood and Adaptation Functions for Hardware Implemented Self-Organizing Maps Applied in Smart Grid
- Author
-
Marta Kolasa
- Subjects
smart grid ,intelligent sensors ,artificial neural networks ,parallel data processing ,asic ,cmos technology ,Technology - Abstract
Smart Grids (SGs) can be successfully supported by Wireless Sensor Networks (WSNs), especially through these consisting of intelligent sensors, which are able to efficiently process the still growing amount of data. We propose a contribution to the development of such intelligent sensors, which in an advanced version can be equipped with embedded low-power artificial neural networks (ANNs), supporting the analysis and the classification of collected data. This approach allows to reduce the energy consumed by particular sensors during the communication with other nodes of a larger WSN. This in turn, facilitates the maintenance of a net of such sensors, which is a paramount feature in case of their application in SG devices distributed over a large area. In this work, we focus on a novel, energy-efficient neighborhood mechanism (NM) with the neighborhood function (NF). This mechanism belongs to main components of self learning ANNs. We propose a realization of this component as a specialized chip in the CMOS technology and its optimization in terms of the circuit complexity and the consumed energy. The circuit was realized as a prototype chip in the CMOS 130 nm technology, and verified by means of transistor level simulations and measurements.
- Published
- 2020
- Full Text
- View/download PDF
15. BoxBroker: A Policy-Driven Framework for Optimizing Storage Service Federation.
- Author
-
Heinsen, Rene, Lopez, Cindy, and Eui-Nam Huh
- Subjects
INFORMATION storage & retrieval systems ,DATA distribution ,COMPUTER algorithms ,COMPUTER network architectures ,SCALABILITY - Abstract
Storage services integration can be done for achieving high availability, improving data access performance and scalability while preventing vendor lock-in. However, multiple services environment management and interoperability have become a critical issue as a result of service architectures and communication interfaces heterogeneity. Storage federation model provides the integration of multiple heterogeneous and self-sufficient storage systems with a single control point and automated decision making about data distribution. In order to integrate diverse heterogeneous storage services into a single storage pool, we are proposing a storage service federation framework named BoxBroker. Moreover, an automated decision model based on a policy-driven data distribution algorithm and a service evaluation method is proposed enabling BoxBroker to make optimal decisions. Finally, a demonstration of our proposal capabilities is presented and discussed. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
16. Gilbert-Multiplier-Based Parallel 1-D and 2-D Analog FIR Filters for Medical Diagnostics
- Author
-
Długosz, Rafał, Gaudet, Vincent, Wojtyna, Ryszard, Kacprzyk, Janusz, editor, Kącki, Edward, editor, Rudnicki, Marek, editor, and Stempczyńska, Joanna, editor
- Published
- 2009
- Full Text
- View/download PDF
17. Od algorytmu dynamicznej cieczy sieciowej do dedykowanego komputera równoległego II – maszyna mDLL.
- Author
-
JUNG, Jarosław, POLANOWSKI, Piotr, KIEŁBIK, Rafał, and RUDNICKI, Kamil
- Abstract
Copyright of Przegląd Elektrotechniczny is the property of Przeglad Elektrotechniczny and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2017
- Full Text
- View/download PDF
18. Design of Sensor Data Management System using Amazon Web Service
- Author
-
福山大学工学部情報工学科
- Subjects
センサーデータ ,並列データ処理 ,aggregation ,data warehouse ,sensor data ,データウエアハウス ,AWS ,parallel data processing ,grouping ,グループ化 ,Amazon Web Service ,集計集約 ,relational database management system ,リレーショ ナルデータベース管理システム - Published
- 2021
19. Components of Artificial Neural Networks Realized in CMOS Technology to be Used in Intelligent Sensors in Wireless Sensor Networks
- Author
-
Tomasz Talaśka
- Subjects
wireless sensor networks ,intelligent sensors ,artificial neural networks ,ASIC ,CMOS technology ,low power solutions ,parallel data processing ,Chemical technology ,TP1-1185 - Abstract
The article presents novel hardware solutions for new intelligent sensors that can be used in wireless sensor networks (WSN). A substantial reduction of the amount of data sent by the sensor to the base station in the WSN may extend the possible sensor working time. Miniature integrated artificial neural networks (ANN) applied directly in the sensor can take over the analysis of data collected from the environment, thus reducing amount of data sent over the RF communication block. A prototype specialized chip with components of the ANN was designed in the CMOS 130 nm technology. An adaptation mechanism and a programmable multi-phase clock generator—components of the ANN—are described in more detail. Both simulation and measurement results of selected blocks are presented to demonstrate the correctness of the design.
- Published
- 2018
- Full Text
- View/download PDF
20. Improved Parallel Algorithms for Path Expression Query Processing of Semi-Structured Data
- Author
-
Sun, Wenjun, Lü, Kevin J., Wong, Kam Fai, Chan, Alvin T. S., editor, Chan, Stephen C. F., editor, Leong, Hong Va, editor, and Ng, Vincent T. Y., editor
- Published
- 2003
- Full Text
- View/download PDF
21. An Effective Data Placement Strategy for XML Documents
- Author
-
Zhu, Yuanling, Lü, Kevin, Goos, Gerhard, editor, Hartmanis, Juris, editor, van Leeuwen, Jan, editor, and Read, Brian, editor
- Published
- 2001
- Full Text
- View/download PDF
22. Fast iterative circuits and RAM-based mergers to accelerate data sort in software/hardware systems.
- Author
-
Sklyarov, Valery, Skliarova, Iouliia, Rjabov, Artjom, and Sudnitson, Alexander
- Subjects
- *
CELLULAR automata , *RANDOM access memory , *SORTING (Electronic computers) , *FIELD programmable analog arrays , *PCI bus (Computer bus) , *COMPUTER architecture - Abstract
The paper suggests and describes two architectures for parallel data sort. The first architecture is applicable to large data sets and it combines three stages of data processing: data sorting in hardware (in a Field-Programmable Gate Arrays - FPGA), merging preliminary sorted blocks in hardware (in the FPGA), and merging large subsets received from the FPGA in general-purpose software. Data exchange between the FPGA and a general-purpose computer is organized through a fast Peripheral Component Interconnect (PCI) express bus. The second architecture is applicable to small data sets and it enables sorting to be done at the time of data acquisition, i.e. as soon as the last data item is received, the sorted items can be transferred immediately. The results of experiments clearly demonstrate the advantages of the proposed architectures that permit the reduction of the required hardware resources and increasing throughput compared to the results reported in publications and software functions targeted to data sorting. [ABSTRACT FROM AUTHOR]
- Published
- 2017
- Full Text
- View/download PDF
23. Development of a system for analyzing the quality of marine seismic data
- Subjects
parallel data processing ,паÑаллелÑÐ½Ð°Ñ Ð¾Ð±ÑабоÑка даннÑÑ ,data quality analysis ,marine seismic exploration ,обÑабоÑка ÑиÑÑовÑÑ Ñигналов ,digital signal processing ,ÑейÑмоÑазведка ,анализ каÑеÑÑва даннÑÑ ,seismic exploration ,моÑÑÐºÐ°Ñ ÑейÑмоÑазведка - Abstract
Тема вÑпÑÑкной квалиÑикаÑионной ÑабоÑÑ: «РазÑабоÑка ÑиÑÑÐµÐ¼Ñ Ð°Ð½Ð°Ð»Ð¸Ð·Ð° каÑеÑÑва даннÑÑ Ð¼Ð¾ÑÑкой ÑейÑмоÑазведки».ÐагиÑÑеÑÑÐºÐ°Ñ ÑабоÑа поÑвÑÑена иÑÑÐ»ÐµÐ´Ð¾Ð²Ð°Ð½Ð¸Ñ Ð¼Ð¾ÑÑкой ÑейÑмоÑазведки как ÑÐ°Ð·Ð´ÐµÐ»Ñ ÑазведоÑной геоÑизики. Ð ÑабоÑе изÑÑаÑÑÑÑ Ð°Ð»Ð³Ð¾ÑиÑÐ¼Ñ Ð¸ ÑÑнкÑии, коÑоÑÑе пÑименÑÑÑÑÑ Ð¿Ñи анализе каÑеÑÑва даннÑÑ , полÑÑеннÑÑ Ð² Ñ Ð¾Ð´Ðµ иÑпÑÑаний моÑÑкой ÑейÑмоÑазведки. Также в данной ÑабоÑе ÑаÑÑмаÑÑиваÑÑÑÑ Ð¿Ð¾Ð´Ñ Ð¾Ð´Ñ Ðº Ð°Ð½Ð°Ð»Ð¸Ð·Ñ ÐºÐ°ÑеÑÑва даннÑÑ Ð¼Ð¾ÑÑкой ÑейÑмоÑазведки, а Ñакже алгоÑиÑÐ¼Ñ Ð¿Ð¾ ÑлÑÑÑÐµÐ½Ð¸Ñ ÑÑого каÑеÑÑва.Ð Ñ Ð¾Ð´Ðµ ÑабоÑÑ Ð¸Ð·ÑÑаÑÑÑÑ Ð±Ð¸Ð±Ð»Ð¸Ð¾Ñеки ÑзÑка ÑазÑабоÑки Python, Ñ Ð¿Ð¾Ð¼Ð¾ÑÑÑ ÐºÐ¾ÑоÑÑÑ Ð¼Ð¾Ð¶Ð½Ð¾ ÑеализоваÑÑ Ð²Ñе изÑÑеннÑе Ñанее алгоÑиÑмÑ, а Ñакже визÑализиÑоваÑÑ Ð¸Ñ ÑезÑлÑÑаÑÑ Ð² виде инÑеÑакÑивнÑÑ Ð³ÑаÑиков.Ð ÑезÑлÑÑаÑе пÑоделанной ÑабоÑÑ Ñеализована ÑиÑÑема анализа каÑеÑÑва даннÑÑ Ð¼Ð¾ÑÑкой ÑейÑмоÑазведки на ÑзÑке Python. ÐаÑем пÑоведено ÐµÑ Ð¼Ð¾Ð´ÑлÑное и ÑиÑÑемное ÑеÑÑиÑование, опиÑÐ°Ð½Ñ ÐµÑ Ð¿ÑеимÑÑеÑÑва, недоÑÑаÑки и напÑÐ°Ð²Ð»ÐµÐ½Ð¸Ñ Ð´Ð°Ð»ÑнейÑего ÐµÑ ÑазвиÑиÑ., The topic of the final qualifying work: "Development of a system for analyzing the quality of marine seismic data".The master's thesis is devoted to the study of marine seismic exploration as a section of exploration geophysics. The paper studies algorithms and functions that are used in analyzing the quality of data obtained during tests of marine seismic exploration. This paper also discusses approaches to analyzing the quality of marine seismic data, as well as algorithms to improve this quality.In the course of the work, Python development language libraries are being studied, with the help of which it is possible to implement all previously studied algorithms, as well as visualize their results in the form of interactive graphs.As a result of the work done, a system for analyzing the quality of marine seismic data in Python has been implemented. Then its modular and system testing was carried out, its advantages, disadvantages and directions of its further development are described.
- Published
- 2022
- Full Text
- View/download PDF
24. QCLab: a framework for query compilation on modern hardware platforms
- Author
-
Funke, Henning, Teubner, Jens, and Neumann, Thomas
- Subjects
Query compilation ,Abfragesprache ,Modern hardware ,Parallel data processing ,Parallelverarbeitung - Abstract
As modern in-memory database systems achieve higher and higher processing speeds, the performance of memory becomes an increasingly limiting factor. Although there has been significant progress, the bottleneck only has shifted. While earlier systems were optimized for memory latencies, current systems are rather affected by the limited memory bandwidth. Query compilation is a proven technique to address bandwidth limitations. It translates queries via Just-In-Time compilation to native programs for the target hardware. The compiled queries execute with very high efficiency and only with a bare minimum of communication via memory. Despite these important improvements, the benefit of query compilation in certain scenarios is limited. On the one hand query compilers typically use standard compiler technology with relatively long compilation times. Therefore the overall execution time can be prolonged by the additional compilation time. On the other hand, not all emerging database technology is compatible with the approach. Query compilation uses a tuple-at-a-time processing style that departs from the column-at-a-time or vector-at- a-time approaches that in-memory systems typically use. Especially data-parallel processing techniques, e.g. SIMD or coprocessing-techniques, are challenging to use in combination with the approach. This work presents QCLab, a framework for query compilation on modern hardware platforms. The framework contains several new query compilation techniques that allow us to address the mentioned shortcomings and ultimately to extend the benefit of query compilation to new workloads and platforms. The techniques cover three aspects: compilation, communication, and processing. Together they serve as basis for building highly efficient query compilers. The techniques make efficient use of communication channels and of the large processing capacities of modern systems. They were designed for practical use and enable efficient processing, even when workload characteristics are challenging.
- Published
- 2022
25. FP-Hadoop: Efficient processing of skewed MapReduce jobs.
- Author
-
Liroz-Gistau, Miguel, Akbarinia, Reza, Agrawal, Divyakant, and Valduriez, Patrick
- Subjects
- *
ELECTRONIC data processing , *BIG data , *SPARK (Computer program language) , *COMPUTER software execution , *COMPUTER network resources - Abstract
Nowadays, we are witnessing the fast production of very large amount of data, particularly by the users of online systems on the Web. However, processing this big data is very challenging since both space and computational requirements are hard to satisfy. One solution for dealing with such requirements is to take advantage of parallel frameworks, such as MapReduce or Spark, that allow to make powerful computing and storage units on top of ordinary machines. Although these key-based frameworks have been praised for their high scalability and fault tolerance, they show poor performance in the case of data skew. There are important cases where a high percentage of processing in the reduce side ends up being done by only one node. In this paper, we present FP-Hadoop , a Hadoop-based system that renders the reduce side of MapReduce more parallel by efficiently tackling the problem of reduce data skew. FP-Hadoop introduces a new phase, denoted intermediate reduce (IR), where blocks of intermediate values are processed by intermediate reduce workers in parallel. With this approach, even when all intermediate values are associated to the same key, the main part of the reducing work can be performed in parallel taking benefit of the computing power of all available workers. We implemented a prototype of FP-Hadoop, and conducted extensive experiments over synthetic and real datasets. We achieved excellent performance gains compared to native Hadoop, e.g. more than 10 times in reduce time and 5 times in total execution time . [ABSTRACT FROM AUTHOR]
- Published
- 2016
- Full Text
- View/download PDF
26. Platform for Parallel Processing of Intense Experimental Data Flow on Remote Supercomputers.
- Author
-
Shchapov, Vladislav, Masich, Grigoriy, and Masich, Alexey
- Subjects
PARALLEL processing ,DATA flow computing ,SUPERCOMPUTERS ,DATA libraries ,DISTRIBUTED computing - Abstract
Modern experimental facilities generate large amounts of data that should be processed, saved to hard disk and presented to the user as fast as possible. However, in-situ data analysis requires technical resources which are often not available. The existence of accessible high-speed networks allows to forward data processing and storage to a remote supercomputer centers and datacenters. These capabilities can be realized through the development of architectural solutions for effective data transmission trough a long-distance high-speed networks, data input/output and data distribution over computers and data storage systems. In this paper, we describe the results of investigations into the development of a software platform for parallel processing of intense experimental data-streams on ICMM UB RAS (Perm) and IMM UB RAS (Yekaterinburg) supercomputers, interconnected by a high-speed network. The reported studies was partially supported by RFBR, research project No. 14-07-96001-r_ural_a and by Program of UD RAS, project No 15-7-1-25. [ABSTRACT FROM AUTHOR]
- Published
- 2015
- Full Text
- View/download PDF
27. Regular Segmented Single-pass Scan in Futhark
- Author
-
Oancea, Cosmin Eugen, Henriksen, Troels, Clausen, Morten Tychsen, Oancea, Cosmin Eugen, Henriksen, Troels, and Clausen, Morten Tychsen
- Abstract
This thesis presents an implementation of a single-pass scan algorithm described by researchers Merrill & Garland, as an extension to the functional language Futhark. The work consists of a generalization of a prior implementation made by Persson & Nicolaisen, modified to work on regular segments. In addition to generalization of the implementation, two major contributions in the form of an analytical model to estimate optimal workload per thread based on type analysis and safe rewriting of index arithmetic to computationally cheaper calculations are made. The implementation and contributions are tested, and their respective benefits are documented using Futharks built-in benchmarking system.
- Published
- 2021
28. On Software Infrastructure for Scalable Graph Analytics
- Author
-
Bu, Yingyi
- Subjects
Computer science ,Big Data Analytics ,Big Graph Analytics ,Database Systems ,Parallel Data Processing - Abstract
Recently, there is a growing need for distributed graph processing systems that are capable of gracefully scaling to very large datasets. In the mean time, in real-world applications, it is highly desirable to reduce the tedious, inefficient ETL (extract, transform, load) gap between tabular data processing systems and graph processing systems. Unfortunately, those challenges have not been easily met due to the intense memory pressure imposed by process-centric, message passing designs that many graph processing systems follow, as well as the separation of tabular data processing runtimes and graph processing runtimes.In this thesis, we explore the application of programming techniques and algorithms from the database systems world to the problem of scalable graph analysis. We first propose a bloat-aware design paradigm towards the development of efficient and scalable Big Data applications in object-oriented, GC enabled languages and demonstrate that programming under this paradigm does not incur significant programming burden but obtains remarkable performance gains (e.g., 2.5X). Based on the design paradigm, we then build Pregelix, an open source distributed graph processing system which is based on an iterative dataflow design that is better tuned to handle both in-memory and out-of-core workloads. As such, Pregelix offers improved performance characteristics and scaling properties over current open source systems (e.g., we have seen up to 15X speedup compared to Apache Giraph and up to 35X speedup compared to distributed GraphLab). Finally, we integrate Pregelix with the open source Big Data management system AsterixDB to offer users a mix of a vertex-oriented programming model and a declarative query language for richer forms of Big Graph analytics with reduced ETL pains.
- Published
- 2015
29. libxtc: an efficient library for reading XTC-compressed MD trajectory data
- Author
-
Roman G. Efremov and Nikolay A. Krylov
- Subjects
Computer science ,lcsh:Medicine ,Molecular dynamics ,Molecular Dynamics Simulation ,010402 general chemistry ,01 natural sciences ,Storage efficiency ,General Biochemistry, Genetics and Molecular Biology ,Parallel data processing ,Computational science ,Software ,0103 physical sciences ,Code (cryptography) ,lcsh:Science (General) ,Biomolecular simulations ,lcsh:QH301-705.5 ,Disk space ,Gene Library ,010304 chemical physics ,business.industry ,Reading (computer) ,lcsh:R ,General Medicine ,0104 chemical sciences ,Research Note ,lcsh:Biology (General) ,Efficiency of MD trajectories reading ,Reading ,Trajectory ,business ,Algorithms ,lcsh:Q1-390 - Abstract
Objective The purpose of this work is to optimize the processing of molecular dynamics (MD) trajectory data obtained for large biomolecular systems. Two popular software tools were chosen as the reference: the tng and the xdrfile libraries. Current implementation of tng algorithms and library is either fast or storage efficient and xdrfile is storage efficient but slow. Our aim was to combine speed and storage efficiency through the xdrfile’s code modification. Results Here we present libxtc, a ready-to-use library for reading MD trajectory files in xtc format. The effectiveness of libxtc is demonstrated for several biomolecular systems of various sizes (~ 2 × 104 to ~ 2 × 105 atoms). In sequential mode, the performance of libxtc is up to 1.8 times higher and 1.4 times lower than xdrfile and tng, respectively. In parallel mode, libxtc is about 3 and 1.3 times faster than xdrfile and tng. At the same time, MD data stored in the xtc format require about 1.3 times less disk space than those treated with the tng algorithm in the fastest reading mode, which is a noticeable saving especially when the MD trajectory is long and the number of atoms is large—this applies to most biologically relevant systems.
- Published
- 2021
30. Analog Programmable Distance Calculation Circuit for Winner Takes All Neural Network Realized in the CMOS Technology.
- Author
-
Talaska, Tomasz, Kolasa, Marta, Dlugosz, Rafal, and Pedrycz, Witold
- Subjects
- *
ASYNCHRONOUS circuits , *COMPLEMENTARY metal oxide semiconductors , *TRANSISTOR-transistor logic circuits , *MEASUREMENT of distances , *EUCLIDEAN algorithm - Abstract
This paper presents a programmable analog current-mode circuit used to calculate the distance between two vectors of currents, following two distance measures. The Euclidean (L2) distance is commonly used. However, in many situations, it can be replaced with the Manhattan (L1) one, which is computationally less intensive, whose realization comes with less power dissipation and lower hardware complexity. The presented circuit can be easily reprogrammed to operate with one of these distances. The circuit is one of the components of an analog winner takes all neural network (NN) implemented in the complementary metal–oxide–semiconductor 0.18- \mu \textm technology. The learning process of the realized NN has been successfully verified by the laboratory tests of the fabricated chip. The proposed distance calculation circuit (DCC) features a simple structure, which makes it suitable for networks with a relatively large number of neurons realized in hardware and operating in parallel. For example, the network with three inputs occupies a relatively small area of 3900 \mu m^2 . When operating in the L2 mode, the circuit dissipates 85 \mu \textW of power from the 1.5 V voltage supply, at maximum data rate of 10 MHz. In the L1 mode, an average dissipated power is reduced to 55 \mu \textW from 1.2 V voltage supply, while data rate is 12 MHz in this case. The given data rates are provided for the worst case scenario, where input currents differ by 1%–2% only. In this case, the settling time of the comparators used in the DCC is quite long. However, that kind of situation is very rare in the overall learning process. [ABSTRACT FROM PUBLISHER]
- Published
- 2016
- Full Text
- View/download PDF
31. A Workflow Application for Parallel Processing of Big Data from an Internet Portal.
- Author
-
Czarnul, Paweł
- Subjects
WORKFLOW management ,PARALLEL processing ,BIG data ,DATA analysis ,WEB portals ,SIMULTANEOUS multithreading processors - Abstract
Abstract: The paper presents a workflow application for efficient parallel processing of data downloaded from an Internet portal. The workflow partitions input files into subdirectories which are further split for parallel processing by services installed on distinct computer nodes. This way, analysis of the first ready sub-directories can start fast and is handled by services implemented as parallel multithreaded applications using multiple cores of modern CPUs. The goal is to assess achievable speed-ups and determine which factors influence scalability and to what degree. Data processing services were implemented for assessment of context (positive or negative) in which the given keyword appears in a document. The testbed application used these services to determine how a particular brand was recognized by either authors of articles or readers in comments in a specific Internet portal focused on new technologies. Obtained execution times as well as speed-ups are presented for data sets of various sizes along with discussion on how factors such as load imbalance and memory/disk bottlenecks limit performance. [Copyright &y& Elsevier]
- Published
- 2014
- Full Text
- View/download PDF
32. Experimental Realization of a Passive Gigahertz Frequency-Division Demultiplexer for Magnonic Logic Networks
- Author
-
Thomas J. Meyer, Alexander A. Serga, Burkard Hillebrands, Florin Ciubotaru, Thomas Brächer, Giacomo Talmelli, Frank Heussner, Philipp Pirro, Moritz Geilen, Kei Yamamoto, Björn Heinz, and Christoph Adelmann
- Subjects
Technology ,Demultiplexer ,Materials Science ,FOS: Physical sciences ,Materials Science, Multidisciplinary ,frequency-division multiplexing ,Applied Physics (physics.app-ph) ,02 engineering and technology ,01 natural sciences ,Multiplexing ,law.invention ,Physics, Applied ,Frequency divider ,Brillouin light scattering ,law ,Mesoscale and Nanoscale Physics (cond-mat.mes-hall) ,0103 physical sciences ,Electronic engineering ,General Materials Science ,010306 general physics ,wave-based logics ,Electronic circuit ,Physics ,Magnonics ,Science & Technology ,Condensed Matter - Mesoscale and Nanoscale Physics ,Magnon ,Transistor ,Physics - Applied Physics ,021001 nanoscience & nanotechnology ,Condensed Matter Physics ,parallel data processing ,Physics, Condensed Matter ,Physical Sciences ,Condensed Matter::Strongly Correlated Electrons ,0210 nano-technology ,Realization (systems) ,spin-wave caustics - Abstract
The emerging field of magnonics employs spin waves and their quanta, magnons, to implement wave-based computing on the micro- and nanoscale. Multi-frequency magnon networks would allow for parallel data processing within single logic elements whereas this is not the case with conventional transistor-based electronic logic. However, a lack of experimentally proven solutions to efficiently combine and separate magnons of different frequencies has impeded the intensive use of this concept. In this Letter, the experimental realization of a spin-wave demultiplexer enabling frequency-dependent separation of magnonic signals in the GHz range is demonstrated. The device is based on two-dimensional magnon transport in the form of spin-wave beams in unpatterned magnetic films. The intrinsic frequency-dependence of the beam direction is exploited to realize a passive functioning obviating an external control and additional power consumption. This approach paves the way to magnonic multiplexing circuits enabling simultaneous information transport and processing., 16 pages, 3 figures
- Published
- 2020
33. Improved Task Graph-based Parallel Data Processing for Dynamic Resource Allocation in Cloud.
- Author
-
Ajitha, A. and Ramesh, D.
- Abstract
Abstract: In recent years large-set parallel data processing has gained quantum as one of the predominant applications of Infrastructure-as-a-Service (IaaS) clouds. Data processing frameworks like Google''s MapReduce and its open source implementation Hadoop, Microsoft''s Dryad and so on are currently in use for parallel data processing in cloud-based companies. But the problem with them is that they are designed for homogeneous environments like clusters and disregard the dynamic and heterogeneous nature of a cloud. As a result, allocation and de-allocation of compute nodes at runtime is ineffective thereby increasing processing time and cost. In this paper we present our approach towards parallel data processing exploiting dynamic resource allocation in IaaS clouds. Our architecture ensures parallel data processing using Directed Acyclic task graph. To reduce the latency and to improve throughput, load balancing is introduced in the architecture. Incoming jobs are divided into tasks that are assigned to different types of virtual machines that are dynamically instantiated and terminated during job execution. [Copyright &y& Elsevier]
- Published
- 2012
- Full Text
- View/download PDF
34. A programmable triangular neighborhood function for a Kohonen self-organizing map implemented on chip
- Author
-
Kolasa, Marta, Długosz, Rafał, Pedrycz, Witold, and Szulc, Michał
- Subjects
- *
SELF-organizing maps , *MATHEMATICAL functions , *APPLICATION-specific integrated circuits , *GAUSSIAN processes , *APPROXIMATION theory , *ENERGY consumption , *ELECTRIC potential , *COMPUTER simulation - Abstract
Abstract: An efficient transistor level implementation of a flexible, programmable Triangular Function (TF) that can be used as a Triangular Neighborhood Function (TNF) in ultra-low power, self-organizing maps (SOMs) realized as Application-Specific Integrated Circuit (ASIC) is presented. The proposed TNF block is a component of a larger neighborhood mechanism, whose role is to determine the distance between the winning neuron and all neighboring neurons. Detailed simulations carried out for the software model of such network show that the TNF forms a good approximation of the Gaussian Neighborhood Function (GNF), while being implemented in a much easier way in hardware. The overall mechanism is very fast. In the CMOS 0.18 μm technology, distances to all neighboring neurons are determined in parallel, within the time not exceeding 11 ns, for an example neighborhood range, , of 15. The TNF blocks in particular neurons require another 6 ns to calculate the output values directly used in the adaptation process. This is also performed in parallel in all neurons. As a result, after determining the winning neuron, the entire map is ready for the adaptation after the time not exceeding 17 ns, even for large numbers of neurons. This feature allows for the realization of ultra low power SOMs, which are hundred times faster than similar SOMs realized on PC. The signal resolution at the output of the TNF block has a dominant impact on the overall energy consumption as well as the silicon area. Detailed system level simulations of the SOM show that even for low resolutions of 3 to 6 bits, the learning abilities of the SOM are not affected. The circuit performance has been verified by means of transistor level Hspice simulations carried out for different transistor models and different values of supply voltage and the environment temperature — a typical procedure completed in case of commercial chips that makes the obtained results reliable. [Copyright &y& Elsevier]
- Published
- 2012
- Full Text
- View/download PDF
35. Realization of the Conscience Mechanism in CMOS Implementation of Winner-Takes-All Self-Organizing Neural Networks.
- Author
-
Długosz, Rafał, Talaśka, Tomasz, Pedrycz, Witold, and Wojtyna, Ryszard
- Subjects
- *
ANALOG computer circuits , *COMPLEMENTARY metal oxide semiconductors , *ARTIFICIAL neural networks , *EVOLUTIONARY computation , *TRANSISTOR-transistor logic circuits - Abstract
This paper presents a complementary metal-oxide-semiconductor (CMOS) implementation of a conscience mechanism used to improve the effectiveness of learning in the winner-takes-all (WTA) artificial neural networks (ANNs) realized at the transistor level. This mechanism makes it possible to eliminate the effect of the so-called "dead neurons," which do not take part in the learning phase competition. These neurons usually have a detrimental effect on the network performance, increasing the quantization error. The proposed mechanism comes as part of the analog implementation of the WTA neural networks (NNs) designed for applications to ultralow power portable diagnostic devices for online analysis of ECG biomedical signals. The study presents Matlab simulations of the network's model, discusses postlayout circuit level simulations and includes results of measurement completed for the physical realization of the circuit. [ABSTRACT FROM AUTHOR]
- Published
- 2010
- Full Text
- View/download PDF
36. Low power current-mode binary-tree asynchronous Min/Max circuit
- Author
-
DŁugosz, RafaŁ and Talaśka, Tomasz
- Subjects
- *
ELECTRIC currents , *ASYNCHRONOUS circuits , *NONLINEAR theories , *ARTIFICIAL neural networks , *ENERGY dissipation , *COMPLEMENTARY metal oxide semiconductors - Abstract
Abstract: A novel, current-mode, binary-tree, asynchronous Min/Max circuit for application in nonlinear filters as well as in analog artificial neural networks is proposed. The relatively high precision above 99% can be achieved by eliminating the copying of the input signals from one layer to the other in the tree. In the proposed solution, the input signals are always directly copied to particular layers using separate signal paths. This makes the precision almost independent on the number of the layers i.e. the number of the inputs. The circuit is a flexible solution. The power dissipation, as well as data rate can be scaled up and down in a wide range. For an average value of the input currents of 20μA and data rate of 11MHz the circuit dissipates 505μW, while for the signals of 200nA and data rate of 500kHz the power dissipation is reduced to 1μW. The prototype circuit with four inputs, realized in the CMOS 0.18μm technology, occupies the area of 1800μm2. [Copyright &y& Elsevier]
- Published
- 2010
- Full Text
- View/download PDF
37. МАТЕМАТИЧЕСКОЕ ОБЕСПЕЧЕНИЕ ПОДДЕРЖКИ ПРОЦЕССА УПРАВЛЕНИЯ СХЕМОЙ РЕЛЯЦИОННОЙ БАЗЫ ДАННЫХ В ЗАДАЧАХ ГОРИЗОНТАЛЬНОГО МАСШТАБИРОВАНИЯ
- Subjects
параллельная обработка данных ,parallel data processing ,competitive queries ,database management system ,конкурентные запросы ,оптимизация запросов ,система управления базами данных ,query optimization - Abstract
В статье предложен подход к разработке математического обеспечения для поддержки процесса управления схемой реляционной СУБД, позволяющего учитывать статистику конкурентного доступа потока запросов к данным в иерархии памяти, используемой ядром СУБД. Отмечены недостатки существующих подходов, основанных на условной стоимости выполнения плана запроса, рассмотрен вопрос игнорирования издержек кооперативного доступа к данным в разделяемой памяти ЭВМ. Дается теоретикомножественное представление процесса обработки потока запросов, учитывающее недостатки существующих теоретикомножественных моделей. Предлагаемое представление обеспечивает учет фактических временных затрат для потока параллельно выполняемых запросов. Операции на высокоуровневом языке запросов предоставляются посредством множеств типовых операций доступа к данным в памяти. Разработана классификация операций доступа к памяти, позволяющая рассчитывать степень конкуренции при кооперативном выполнение запросов. Вводится формальное представление конкурентных запросов и условие выбора оптимального распределения данных в ходе функционирования базы данных на некотором промежутке времени. Полученные в работе результаты могут найти применение при разработке математического и программного обеспечения автономных систем управления базами данных, автоматизирующих управление физической схемой базы данных., The article proposes an approach to the development of mathematical support system for the relational database (database control system) schematic control process, allowing to take into account the statistics of competitive access to the data query flow in the memory used by the database control system engine by hierarchy. The articles notes shortcomings of existing approaches based on the conditional costs of the query plan and the issue of ignoring the costs related to cooperative access to the data contained in a shared computer memory. There is a settheoretic presentation of the query flow processing, which also takes into consideration the shortcomings of the existing settheoretic models. The proposed approach ensures the actual time costs calculations for the flow of concurrently executed queries. Highlevel query language operations are provided through a variety of typical access operations to the InMemory Data. The article presents a classification of the memory access operations, which allows to calculate the degree of competition during the cooperative query execution. This study also introduces a formal presentation of competitive queries and the conditions for choosing the optimal data distribution method during the database operation in a given period of time. The proposed settheoretic model allows us to calculate the memory segments, access to which leads to a competitive query. The results obtained in this paper can be used in the development of mathematical support systems and other software for the Autonomous database control systems that automate the management of the physical database scheme., №2(25) (2019)
- Published
- 2019
- Full Text
- View/download PDF
38. Platform for Parallel Processing of Intense Experimental Data Flow on Remote Supercomputers
- Author
-
Alexey Masich, Vladislav Shchapov, and Grigoriy Masich
- Subjects
Middleware ,business.industry ,Computer science ,Distributed computing ,Experimental data ,Supercomputer ,Parallel data processing ,Long fat network ,Software ,Parallel processing (DSP implementation) ,Computer data storage ,Distributed system ,General Earth and Planetary Sciences ,business ,General Environmental Science ,Data transmission - Abstract
Modern experimental facilities generate large amounts of data that should be processed, saved to hard disk and presented to the user as fast as possible. However, in-situ data analysis requires technical resources which are often not available. The existence of accessible high-speed networks allows to forward data processing and storage to a remote supercomputer centers and datacenters. These capabilities can be realized through the development of architectural solutions for effective data transmission trough a long-distance high-speed networks, data input/output and data distribution over computers and data storage systems. In this paper, we describe the results of investigations into the development of a software platform for parallel processing of intense experimental data-streams on ICMM UB RAS (Perm) and IMM UB RAS (Yekaterinburg) supercomputers, interconnected by a high-speed network. The reported studies was partially supported by RFBR, research project No. 14-07-96001-r_ural_a and by Program of UD RAS, project No 15-7-1-25.
- Published
- 2015
39. РЕАЛИЗАЦИЯ ПАРАЛЛЕЛЬНОЙ ОБРАБОТКИ ДАННЫХ В ОБЛАЧНЫХ СИСТЕМАХ
- Author
-
Munerman, V.I.
- Subjects
параллельная обработка данных ,методы доступа ,parallel data processing ,massively data processing ,архитектура программно-аппаратных комплексов ,data access methods ,Cloud systems and computing ,Облачные системы и вычисления ,массовая обработка данных ,software- hardware complexes architecture - Abstract
Рассмотрены методы использования технологии облачных вычислений для разработки программного обеспечения, реализующего параллельную обработку распределенных данных. Рассмотрен метод параллельной реализации операции JOIN на основе принципа симметричного горизонтального распределения. Даны описания виртуальных программно-аппаратных комплексов, которые реализуют распараллеливание на уровне обработки запросов и на уровне операции JOIN средствами Microsoft Azure. Приведены экспериментальные данные, подтверждающие эффективность предложенного подхода., The methods of using cloud computing technology for the development of software that implements parallel processing of distributed data are considered. The method of parallel implementation of the JOIN operation based on the symmetric horizontal distribution principle is considered. There are descriptions of virtual software and hardware systems that implement parallelization at the level of processing requests and at the level of the JOIN operation using Microsoft Azure. Experimental data confirming the effectiveness of the proposed approach are presented., Международный научный журнал "Современные информационные технологии и ИТ-образование", № (2017)
- Published
- 2017
- Full Text
- View/download PDF
40. Security assurance assessment for multi-layered and multi-tenant hybrid clouds
- Author
-
Hudic, Aleksandar
- Subjects
FOS: Computer and information sciences ,Parallel Data processing ,Computer security ,Cloud System Architecture ,Distributed Computing - Abstract
This thesis is based on five publications related to the area of security assurance for hybrid clouds which were published at conferences or in journals by IEEE and Elsevier. Cloud computing is an ongoing research field that received an increasing attention in last several years, as new challenges arise in the cloud ecosystem on daily basis especially with the new hybrid cloud models coming to the scene. Meanwhile, the old challenges with regards to security, privacy and especially transparency haven't been comprehended or addressed properly to keep up with the technological momentum caused by the cloud paradigm. Sharing security sensitive information in a cloud environments has become a main obstacle due to the immense lack of transparency. Hence, this thesis addresses this challenges, in particular, transparency of cloud providers that for the given security objectives there are supporting measures in place. The thesis is especially concerned with the security and transparency with regards to security critical services especially with regards to hosting them in hybrid multi-layered and multi-tenant environments. To comprehend the complexity of hybrid cloud environments that can be composed of multiple layers and owned by multiple stakeholder we illustrate a composite multi-layer reference architecture model. The main objective of this model is to observe multidimensional critical infrastructure systems at individual levels from different viewpoints, namely those of multi-provider and multi-tenant, and different stakeholders. Furthermore, we analyze the challenges, objectives and requirements for deploying critical infrastructure services to cloud environments with regards to transparency and security. The challenges that we identified highlight the shortcoming of cloud providers to support transparency especially with regards to the hybrid cloud solutions. To overcome this gap this thesis propose a novel model for holistic security assurance assessment that addresses the interdependencies between both individual components and abstraction levels in hybrid cloud environments. The approach offers the ability to address each individual component of a cloud based infrastructure, regardless if it is a physical server, virtual container, or a high level service, in a structural manner by including all its interdependencies. The flexibility of the approach lies in the composite structural design of the security assurance assessment framework that adheres the Common Criteria and enhances it to achieve higher level of granularity when assessing services. Most importantly, unlike standard approaches for security assessment like certification or auditing our model offers continuous security assessment ability of hybrid cloud environments where we can have competitive cloud provider that deliver one single service. Lastly, our security assurance assessment model prevents the exposure of internal security sensitive information of a cloud provider via its novel security assurance assessment model that operates on abstracted security information sets. Furthermore, we propose a comprehensive life-cycle for designing, developing and deploying secure cloud services in line with standards, regulative compliance, and best practices. In addition, the proposed life-cycle integrates iterative security requirements engineering from high level objectives to security properties used for security validation entities through both development and production phase of cloud services. The life-cycle aligns and integrates the security assurance assessment model, by at the same time supporting it with security requirements, in the final production phase to enhance transparency.
- Published
- 2017
- Full Text
- View/download PDF
41. An Energy-Efficient, Parallel Neighborhood and Adaptation Functions for Hardware Implemented Self-Organizing Maps Applied in Smart Grid.
- Author
-
Kolasa, Marta
- Subjects
- *
SELF-organizing maps , *INTELLIGENT sensors , *CIRCUIT complexity , *ARTIFICIAL neural networks , *WIRELESS sensor networks , *ENERGY conservation , *PHYSIOLOGICAL adaptation - Abstract
Smart Grids (SGs) can be successfully supported by Wireless Sensor Networks (WSNs), especially through these consisting of intelligent sensors, which are able to efficiently process the still growing amount of data. We propose a contribution to the development of such intelligent sensors, which in an advanced version can be equipped with embedded low-power artificial neural networks (ANNs), supporting the analysis and the classification of collected data. This approach allows to reduce the energy consumed by particular sensors during the communication with other nodes of a larger WSN. This in turn, facilitates the maintenance of a net of such sensors, which is a paramount feature in case of their application in SG devices distributed over a large area. In this work, we focus on a novel, energy-efficient neighborhood mechanism (NM) with the neighborhood function (NF). This mechanism belongs to main components of self learning ANNs. We propose a realization of this component as a specialized chip in the CMOS technology and its optimization in terms of the circuit complexity and the consumed energy. The circuit was realized as a prototype chip in the CMOS 130 nm technology, and verified by means of transistor level simulations and measurements. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
42. FP-Hadoop: Efficient Processing of Skewed MapReduce Jobs
- Author
-
Divyakant Agrawal, Miguel Liroz-Gistau, Reza Akbarinia, Patrick Valduriez, Scientific Data Management (ZENITH), Laboratoire d'Informatique de Robotique et de Microélectronique de Montpellier (LIRMM), Centre National de la Recherche Scientifique (CNRS)-Université de Montpellier (UM)-Centre National de la Recherche Scientifique (CNRS)-Université de Montpellier (UM)-Inria Sophia Antipolis - Méditerranée (CRISAM), Institut National de Recherche en Informatique et en Automatique (Inria)-Institut National de Recherche en Informatique et en Automatique (Inria), University of California [Santa Barbara] (UCSB), University of California, Université de Montpellier (UM)-Centre National de la Recherche Scientifique (CNRS)-Université de Montpellier (UM)-Centre National de la Recherche Scientifique (CNRS)-Inria Sophia Antipolis - Méditerranée (CRISAM), University of California [Santa Barbara] (UC Santa Barbara), and University of California (UC)
- Subjects
Data Skew ,Parallel Data Processing ,business.industry ,Computer science ,Distributed computing ,Big data ,Skew ,020206 networking & telecommunications ,Fault tolerance ,02 engineering and technology ,Power (physics) ,Hardware and Architecture ,020204 information systems ,[INFO.INFO-IR]Computer Science [cs]/Information Retrieval [cs.IR] ,Spark (mathematics) ,Scalability ,0202 electrical engineering, electronic engineering, information engineering ,Key (cryptography) ,Data-intensive computing ,MapReduce ,business ,Software ,Information Systems - Abstract
Nowadays, we are witnessing the fast production of very large amount of data, particularly by the users of online systems on the Web. However, processing this big data is very challenging since both space and computational requirements are hard to satisfy. One solution for dealing with such requirements is to take advantage of parallel frameworks, such as MapReduce or Spark, that allow to make powerful computing and storage units on top of ordinary machines. Although these key-based frameworks have been praised for their high scalability and fault tolerance, they show poor performance in the case of data skew. There are important cases where a high percentage of processing in the reduce side ends up being done by only one node.In this paper, we present FP-Hadoop, a Hadoop-based system that renders the reduce side of MapReduce more parallel by efficiently tackling the problem of reduce data skew. FP-Hadoop introduces a new phase, denoted intermediate reduce (IR), where blocks of intermediate values are processed by intermediate reduce workers in parallel. With this approach, even when all intermediate values are associated to the same key, the main part of the reducing work can be performed in parallel taking benefit of the computing power of all available workers.We implemented a prototype of FP-Hadoop, and conducted extensive experiments over synthetic and real datasets. We achieved excellent performance gains compared to native Hadoop, e.g. more than 10 times in reduce time and 5 times in total execution time. HighlightsA novel approach for dealing with data skew in the reduce side of MapReduce.Parallel reducing of each key, using multiple reduce workers.Hierarchical execution of MapReduce jobs.Non-overwhelming reducing of intermediate data.
- Published
- 2016
43. Spezifikation und Optimierung analytischer Datenflüsse
- Author
-
Hüske, Fabian, Markl, Volker, Technische Universität Berlin, Kao, Odej, and Gemulla, Rainer
- Subjects
PACT programming model ,Datenflüsse ,data analysis ,Apache Flink ,parallel data processing ,data flows ,Datenanalyse ,parallele Datenverarbeitung ,big data ,000 Informatik, Informationswissenschaft, allgemeine Werke ,ddc:000 ,MapReduce ,PACT Programmiermodell ,Anfrageoptimierung ,Datenbanksystem ,query optimization ,database systems - Abstract
In the past, the majority of data analysis use cases was addressed by aggregating relational data. Since a few years, a trend is evolving, which is called “Big Data” and which has several implications on the field of data analysis. Compared to previous applications, much larger data sets are analyzed using more elaborate and diverse analysis methods such as information extraction techniques, data mining algorithms, and machine learning methods. At the same time, analysis applications include data sets with less or even no structure at all. This evolution has implications on the requirements on data processing systems. Due to the growing size of data sets and the increasing computational complexity of advanced analysis methods, data must be processed in a massively parallel fashion. The large number and diversity of data analysis techniques as well as the lack of data structure determine the use of user-defined functions and data types. Many traditional database systems are not flexible enough to satisfy these requirements. Hence, there is a need for programming abstractions to define and efficiently execute complex parallel data analysis programs that support custom user-defined operations. The success of the SQL query language has shown the advantages of declarative query specification, such as potential for optimization and ease of use. Today, most relational database management systems feature a query optimizer that compiles declarative queries into physical execution plans. Cost-based optimizers choose from billions of plan candidates the plan with the least estimated cost. However, traditional optimization techniques cannot be readily integrated into systems that aim to support novel data analysis use cases. For example, the use of user-defined functions (UDFs) can significantly limit the optimization potential of data analysis programs. Furthermore, lack of detailed data statistics is common when large amounts of unstructured data is analyzed. This leads to imprecise optimizer cost estimates, which can cause sub-optimal plan choices. In this thesis we address three challenges that arise in the context of specifying and optimizing data analysis programs. First, we propose a parallel programming model with declarative properties to specify data analysis tasks as data flow programs. In this model, data processing operators are composed of a system-provided second-order function and a user-defined first-order function. A cost-based optimizer compiles data flow programs specified in this abstraction into parallel data flows. The optimizer borrows techniques from relational optimizers and ports them to the domain of general-purpose parallel programming models. Second, we propose an approach to enhance the optimization of data flow programs that include UDF operators with unknown semantics. We identify operator properties and conditions to reorder neighboring UDF operators without changing the semantics of the program. We show how to automatically extract these properties from UDF operators by leveraging static code analysis techniques. Our approach is able to emulate relational optimizations such as filter and join reordering and holistic aggregation push-down while not being limited to relational operators. Finally, we analyze the impact of changing execution conditions such as varying predicate selectivities and memory budgets on the performance of relational query plans. We identify plan patterns that cause significantly varying execution performance for changing execution conditions. Plans that include such risky patterns are prone to cause problems in presence of imprecise optimizer estimates. Based on our findings, we introduce an approach to avoid risky plan choices. Moreover, we present a method to assess the risk of a query execution plan using a machine-learned prediction model. Experiments show that the prediction model outperforms risk predictions which are computed from optimizer estimates., In der Vergangenheit wurde die überwiegende Mehrheit der Datenanalyseanwendungen durch die Aggregation von relationalen Daten abgedeckt. Seit einigen Jahren entwickelt sich ein Trend der “Big Data” genannt wird und der große Auswirkungen auf den Bereich der Datenanalyse hat. Im Vergleich zu bisherigen Analyseanwendungen, werden nun wesentlich größere Datenmengen mit deutlich aufwändigeren und vielfältigeren Analysemethoden wie zum Beispiel Techniken der Informationsextraktion, des Data Minings, und Verfahren des maschinellen Lernens ausgewertet. Dabei werden auch Daten in die Analyse einbezogen, die weniger stark oder überhaupt nicht strukturiert sind. Die Veränderungen der Eigenschaften von Datenanalyseanwendungen wirken sich auch auf die Anforderungen an Systeme zur Datenverarbeitung aus. Aufgrund des gestiegenen Datenvolumens und des durch komplexere Analyseverfahren deutlich höheren Berechnungsaufwands müssen Daten massiv parallel verarbeitet werden. Die gestiegene Vielfalt von Analyseverfahren und die geringere Struktur der Daten erfordern häufig den Einsatz von benutzerdefinierten Funktionen und Datenstrukturen. Viele traditionelle Datenbanksysteme sind nicht flexibel genug, um diesen Anforderungen gerecht zu werden. Deshalb gibt es ein großes Interesse an neuen Programmierabstraktionen mit denen komplexe und parallele Datenanalyseanwendungen spezifiziert und effizient ausgeführt werden können. Der Erfolg der Anfragesprache SQL hat die Vorzüge von deklarativer Anfragespezifikation, wie zum Beispiel Optimierungspotenzial und Benutzerfreundlichkeit, deutlich gezeigt. Heute nutzt nahezu jedes relationale Datenbanksystem einen Anfrageoptimierer der deklarative Anfragen in physische Ausführungspläne übersetzt. Kostenbasierte Optimierer sind in der Lage aus Milliarden von möglichen Plänen einen effizienten Plan auszuwählen. Allerdings lassen sich traditionelle Optimierungsmethoden nicht ohne weiteres in Systeme integrieren, die neuartige Anwendungsfälle von Datenanalyse unterstützen wollen. Zum Beispiel kann der Einsatz von benutzerdefinierten Operationen das Optimierungspotenzial sehr stark reduzieren. Darüberhinaus sind selten detaillierte Datenstatistiken verfügbar, wenn große unstrukturierte Datensätze analysiert werden. Fehlende Statistiken haben häufig ungenaue Kostenschätzungen des Optimierers und somit die Auswahl von suboptimalen Ausführungsplänen zur Folge. In dieser Arbeit adressieren wir drei Herausforderungen im Kontext der Spezifikation und Optimierung von parallelen Datenanalyseprogrammen mit benutzerdefinierten Funktionen. Zunächst stellen wir ein paralleles Programmiermodell mit deklarativen Eigenschaften vor um Datenanalyseprogramme als Datenflußprogramme zu spezifizieren. In diesem Modell bestehen Datenverarbeitungsoperatoren aus einer systemeigenen Funktion zweiter Ordnung und einer benutzerdefinierten Funktion erster Ordnung. Ein kostenbasierter Optimierer übersetzt Datenflußprogramme, die in unserem Programmiermodell definiert wurden, in parallele Datenflüße. Unser Optimierer baut auf viele Techniken der relationalen Optimierung auf und überträgt sie in die Domäne von universellen parallelen Programmiermodellen. Zweitens präsentieren wir einen Ansatz zur Verbesserung der Optimierung von Datenflußprogrammen, die benutzerdefinierte Operatoren mit unbekannter Semantik enthalten. Wir identifizieren Eigenschaften von Operatoren und Bedingungen, um die Reihenfolge von benachbarten benutzerdefinierten Operatoren zu verändern ohne die Semantik eines Programms zu ändern. Wir zeigen wie diese Eigenschaften für benutzerdefinierte Operatoren vollautomatisch mit Hilfe von statischer Codeanalyse aus deren Quellcode extrahiert werden können. Mit unserem Ansatz können viele relational Optimierungen wie zum Beispiel die Optimierung der Reihenfolge von Filtern, Joins und Aggregationen emuliert werden ohne jedoch auf relationale Operatoren beschränkt zu sein. Drittens analysieren wir den Einfluß von sich verändernden Ausführungsbedingungen wie zum Beispiel variierenden Prädikatselektivitäten und verfügbaren Hauptspeichermengen auf die Laufzeit von relationalen Ausführungsplänen. Wir identifizieren Planeigenschaften, die deutliche Laufzeitschwankungen auslösen können. Im Fall von ungenauen Optimiererschätzungen können Pläne, die diese Eigenschaften enthalten, ein sehr großes Risiko darstellen. Wir präsentieren einen Ansatz, um die Auswahl von riskanten Plänen zu vermeiden. Darüberhinaus stellen wir eine Methode vor, um das Risiko von Ausführungsplänen mit Hilfe eines maschinell-gelernten Modells vorher zusagen. Unsere Evaluation zeigt, dass mit unserem Vorhersagemodell das Risikopotenzial eines Plans besser abgeschätzt werden kann als mit Hilfe eines kostenbasierten Optimierers.
- Published
- 2016
44. Entwicklung und Umsetzung eines echtzeitfähigen Datenverarbeitungs- und Rekonstruktionsalgorithmus für die ultraschnelle Elektronenstrahl-Röntgen-CT
- Author
-
Frust, T.
- Subjects
parallel data processing ,in-situ ,ROFEX ,cuda - Abstract
This thesis demonstrates the development and the implementation of a real-time capable data processing and reconstruction algorithm for the ulftrafast X-ray scanner ROssendorf Fast Electron beam X-ray tomograph (ROFEX). This measuring system is built for non-invasive imaging of multiphase fluids. Thus, it requires a high scan rate of more than 1 kHz. This is achieved by an arrangement without mechanically rotating parts providing scan rates of up to 8 kHz. Current data processing is not suited to reconstruct a data stream of around 1,3 GB/s. Hence, visual inspection or active process feedback control is not possible yet. This work presents the design and implementation of real-time capable data processing providing a better usability and new fields of application for ROFEX. Therefore, the adjustment of data transfer from the measuring system to the reconstruction workstation as well as the implementation of a new software is essential. The application is implemented using a generic software pipeline consisting of distinct processing units exploiting data parallelism of NVIDIA GPUs with CUDA. Measurements on a NVIDIA GeForce GTX 1080 and NVIDIA Tesla K20c yield a reconstruction rate of more than 1 kHz, which is well-suited for an online application of ROFEX. Furthermore, a test system is introduced simulating an online capable detector. It shows the online visualization of reconstructed images as a first new application.
- Published
- 2016
45. Frequency‐Division Multiplexing in Magnonic Logic Networks Based on Caustic‐Like Spin‐Wave Beams.
- Author
-
Heussner, Frank, Nabinger, Matthias, Fischer, Tobias, Brächer, Thomas, Serga, Alexander A., Hillebrands, Burkard, and Pirro, Philipp
- Subjects
- *
SPIN waves , *FERROMAGNETIC materials , *MAGNONS , *FREQUENCY division multiple access , *COMPLEMENTARY metal oxide semiconductors , *MICROSTRUCTURE - Abstract
Wave‐based data processing by spin waves (SW) and their quanta, magnons, is a promising technique to overcome the challenges which CMOS‐based logic networks are facing nowadays. The advantage of these quasi‐particles lies in their potential for the realization of energy efficient devices on the micro‐ to nanometer scale due to their charge‐less propagation in magnetic materials. In this paper, the frequency dependence of the propagation direction of caustic‐like spin‐wave beams in microstructured ferromagnets is studied by micromagnetic simulations. Based on the observed alteration of the propagation angle, an approach to spatially combine and separate spin‐wave signals of different frequencies is demonstrated. The presented magnetic structure constitutes a prototype design of a passive circuit enabling frequency‐division multiplexing (FDM) in magnonic logic networks. It is verified that spin‐wave signals of different frequencies can be transmitted through the device simultaneously without any interaction or creation of spurious signals. Due to the wave‐based approach of computing in magnonic networks, the technique of FDM can be the basis for parallel data processing in single magnonic devices, enabling the multiplication of the data throughput. The developed prototype design of a multi‐frequency spin‐wave circuit on the micrometer scale can help to drastically enhance the throughput of wave‐based logic networks. Utilizing the frequency dependence of the propagation direction of caustic‐like spin‐wave beams enables the combination and separation of spin waves of different frequencies, which can be used for simultaneous data processing in single devices. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
46. Components of Artificial Neural Networks Realized in CMOS Technology to be Used in Intelligent Sensors in Wireless Sensor Networks.
- Author
-
Talaśka, Tomasz
- Subjects
- *
ARTIFICIAL neural networks , *COMPLEMENTARY metal oxide semiconductors , *WIRELESS sensor networks , *WIRELESS communications , *SIMULATION methods & models - Abstract
The article presents novel hardware solutions for new intelligent sensors that can be used in wireless sensor networks (WSN). A substantial reduction of the amount of data sent by the sensor to the base station in the WSN may extend the possible sensor working time. Miniature integrated artificial neural networks (ANN) applied directly in the sensor can take over the analysis of data collected from the environment, thus reducing amount of data sent over the RF communication block. A prototype specialized chip with components of the ANN was designed in the CMOS 130 nm technology. An adaptation mechanism and a programmable multi-phase clock generator—components of the ANN—are described in more detail. Both simulation and measurement results of selected blocks are presented to demonstrate the correctness of the design. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
47. Paralelní zpracování dat a možnosti datové analytiky v rámci Big Data
- Author
-
Tvrdíková, Milena, Derján, Lukáš, Tvrdíková, Milena, and Derján, Lukáš
- Abstract
V rámci diplomové práce je analyzována práce s velkoobjemovými a nestrukturovanými daty, neboli Big Data. Čtenáři je objasněna architektura Big Data orientovaného řešení a její srovnání s tradiční architekturou Business Intelligence. Právě tradiční Business Intelligence nástroje a řešení stále nejsou technologicky připraveny pro zpracování Big Data, což dalo za vznik jak novým přístupům v paralelním zpracování dat, tak vzniku nových, Big Data orientovaných, technologií. Důležitou roli ve spojení s Big Data hraje datová analytika. Pomocí relevantních analýz mohou organizace získat více informací o svých zákaznících, odhalit v datech skryté souvislosti a zvýšit tak své zisky i věrnost zákazníků. Platformou, která je technologicky připravená pro zpracování a analýzu Big Data, je Apache Hadoop. Tato platforma je více přiblížena nejen v teoretické části, kde je i je definován pojem Big Data a problematika paralelního zpracování dat, ale i v rámci části praktické, kdy platforma slouží pro analytické zpracování vybraného datového souboru. Diplomová práce tak popisuje základní rysy programového frameworku MapReduce i distribuovaného souborového systému HDFS, dohromady tvořící implementaci Hadoop. Z hlediska uplatnitelnosti se tedy jedná o implementaci analytické úlohy dle zákaznických požadavků s reálným výstupem. Stále vyšší počet nasazení analytických platforem nad stávajícími BI řešeními v organizacích a stále narůstající objem veřejně dostupných dat, je pak ze sociálního hlediska potenciálně problematická oblast, která dříve, či později narazí na bariéry osobního soukromí. Praktická část práce vychází ze zadání projektu zákaznické společnosti, v rámci které byla vypracovávána. Projekt je zaměřen na zjištění vhodnosti Big Data platformy Hadoop pro spouštění analytických úloh nad relativně malými soubory. K ověření vhodnosti sloužila analýza n-gramů v rámci vybraného datového souboru, kdy byla využita kromě klasického MapReduce frameworku i in-memory řešení Spark a TEZ. Z, The diploma thesis focuses on analysing the way of working and processing the high-volume unstructured datasets, called Big Data. Reader will find out more about the architecture of Big Data-oriented solutions and its comparison with the traditional architecture of Business Intelligence solutions (BI). Now traditional Business Intelligence tools and solutions are still not technologically ready for processing Big Data. This has led into emergence of new approaches to parallel data processing and the new Big Data-oriented, technologies. Data analytics is playing an important role when talking about the Big Data. If using relevant analysis, organizations can get more information about their customers, uncover hidden relationships in data and increase their profits and customers loyalty. There is a platform that is technologically ready for processing and analysing Big Data. The Apache Hadoop. This platform is more described within the theoretical part, where the terms of Big Data and parallel data processing are explained, as well as in practical part of the diploma thesis, where the platform is used for analytical processing of the pre-selected data file. Thus basic features of a programming framework MapReduce and a distributed file system HDFS (together forming the Hadoop implementation) are explained. In terms of applicability the implementation of analytical tasks according to customer requirements is the real outcome. An increasing number of analytical platforms deployment on top of existing BI solutions in organizations and the ever-increasing volume of publicly available data, is then in social terms, a potentially problematic area that sooner or later hit the barriers personal privacy. The practical part of the thesis is based on the project requirements from the client company. The project is focused on finding the suitability of Big Data Hadoop platform for running analytical tasks over the relatively small datasets. To verify the suitability the n-gram ana, Ve zpracování, Import 22/07/2015
- Published
- 2015
48. Efficient face detection on Epiphany multicore processor
- Author
-
Sukhinov, A. A. and Ostrobrod, G. B.
- Subjects
УДК 004.272.23 ,УДК 004.272.45 ,distributed memory ,УДК 004.93’1 ,распределенная память ,параллельная обработка данных ,parallel data processing ,face detection ,детекция лиц ,specialized processors ,ГРНТИ 50.33 ,специализированные микропроцессоры ,local binary patterns ,УДК 004.258 ,локальные бинарные шаблоны - Abstract
В статье рассматривается возможность использования энергоэффективного микропроцессора Epiphany для решения актуальной прикладной задачи - детекции лиц на изображении. Этот микропроцессор представляет собой многоядерную вычислительную систему с распределенной памятью, выполненную на одном кристалле. Из-за малой площади кристалла микропроцессор обладает существенными аппаратными ограничениями (в частности, он имеет всего 32 килобайта памяти на ядро), которые ограничивают выбор алгоритма и затрудняют его программную реализацию. Для детекции лиц адаптирован известный алгоритм, основанный на каскадном классификаторе, использующем LBP-признаки (Local Binary Patterns). Показано, что микропроцессор Epiphany, имеющий 16 ядер, может на этой задаче в 2,5 раза обогнать одноядерный процессор персонального компьютера той же тактовой частоты, при этом потребляя лишь 0,5 ватта электрической мощности. I t is studied the possibility of usage of energy-efficient Epiphany microprocessor for solving actual applied problem of face detection at still image. The microprocessor is a multicore system with distributed memory, implemented in a single chip. Due to small die area the microprocessor has significant hardware limitations (in particular it has only 32 kilobytes of memory per core) which limit the range of usable algorithms and complicate their software implementation. Common face-detection algorithm based on local binary patterns (LBP) and cascading classifier was adapted for parallel implementation. It is shown that Epiphany microprocessor having 16 cores can outperform single-core CPU of personal computer having the same clock rate by a factor of 2.5, while consuming only 0.5 watts of electric power. Сухинов Антон Александрович, к.ф. м.н., научный сотрудник, Сколковский институт науки и технологий (Сколково, Российская Федерация), soukhinov@gmail.com. Остроброд Георгий Борисович, программист ООО «СиВижинЛаб» (Таганрог, Российская Федерация), wdf.gost@gmail.com. A.A. Sukhinov, Skolkovo Institute of Science and Technology (Skolkovo, Russia), G.B. Ostrobrod, CVisionLab (Taganrog, Russia)
- Published
- 2014
49. Vyhodnocování relačních dotazů v proudově orientovaném prostředí
- Author
-
Kikta, Marcel, Bednárek, David, and Černý, Tomáš
- Subjects
pipeline ,parallel data processing ,relational algebra ,data streaming ,relační algebra ,paralelní zpracování dat - Abstract
This thesis deals with the design and implementation of an optimizer and a transformer of relational queries. Firstly, the thesis describes the theory of the relational query compilers. Secondly, we present the data structures and algorithms used in the implemented tool. Finally, the important implementation details of the developed tool are discussed. Part of the thesis is the selection of used relational algebra operators and design of an appropriate input. Input of the implemented software is a query written in a XML file in the form of relational algebra. Query is optimized and transformed into physical plan which will be executed in the parallelization framework Bobox. Developed compiler outputs physical plan written in the Bobolang language, which serves as an input for the Bobox. Powered by TCPDF (www.tcpdf.org)
- Published
- 2014
50. Increasing the performance of the software framework for implementing the algorithms of the group method of data handling
- Subjects
параллельная обработка данных ,parallel data processing ,object-oriented analysis and design ,объектно-ориентированное проектирование ,объектно-ориентированный анализ ,метод группового учёта аргументов ,эффективность ,распараллеливание ,software framework ,efficiency of paralleling ,программные платформы ,group method of data handling - Abstract
В предыдущих работах автором была предложена универсальная программная платформа, позволяющая реализовать известные алгоритмы метода группового учета аргументов, базисы, методы обучения и критерии селекции моделей. В статье приводится решение актуальной задачи увеличения производительности этой платформы. На основе проведенного обзора существующих архитектур вычислительных систем для параллельной обработки данных и программных систем индуктивного моделирования с поддержкой параллельных вычислений выработаны требования к подсистемам параллельных вычислений и управления памятью программной платформы. С использованием методологии объектно-ориентированного анализа и проектирования разработана объектно-ориентированная структура этих подсистем, приведены особенности их работы для каждой из указанных архитектур вычислительных систем. Производительность параллельной реализации комбинаторного алгоритма метода группового учета аргументов на базе программной платформы оценена экспериментально для многоядерных процессоров. In previous works the author has proposed a universal software framework that allows implementing the known algorithms of group method of data handling, model bases, training methods and model selection criteria. This paper introduces the solution of a topical problem of increasing the performance of the framework. Based on the review of existing computing architectures for parallel data processing and software systems for inductive modeling supporting parallel computations the author has worked out the requirements for the subsystems of parallel computing and memory management of the software framework. Using the methodology of object-oriented analysis and design the author developed the object-oriented structure of these subsystems and introduced the specifics of their operation on each of the mentioned computing architectures. The performance of the parallel implementation of the combinatorial group method of data handling algorithm on basis of the software framework was evaluated experimentally for multi-core processors.
- Published
- 2013
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.