978 results on '"Tassiulas, Leandros"'
Search Results
2. A Brief Introduction to Quantum Network Control
- Author
-
Valls, Víctor, Promponas, Panagiotis, and Tassiulas, Leandros
- Subjects
Mathematics - Optimization and Control ,Quantum Physics - Abstract
Quantum networking is an emerging area with the potential to transform information processing and communications. In this paper, we present a brief introduction to quantum network control, an area in quantum networking dedicated to designing algorithms for distributing entanglement (i.e., entangled qubits). We start by explaining what qubits and entanglement are and how they furnish quantum network control operations such as entanglement swapping and teleportation. With those operations, we present a model for distributing entanglement in a multi-hop quantum network to enable applications such as quantum key distribution and distributed quantum computing. We conclude the paper by presenting open research problems in the field, including the characterization of the quantum network capacity region and the design of throughput-optimal policies.
- Published
- 2024
3. A Genetic Approach to Minimising Gate and Qubit Teleportations for Multi-Processor Quantum Circuit Distribution
- Author
-
Crampton, Oliver, Promponas, Panagiotis, Chen, Richard, Polakos, Paul, Tassiulas, Leandros, and Samuel, Louis
- Subjects
Quantum Physics - Abstract
Distributed Quantum Computing (DQC) provides a means for scaling available quantum computation by interconnecting multiple quantum processor units (QPUs). A key challenge in this domain is efficiently allocating logical qubits from quantum circuits to the physical qubits within QPUs, a task known to be NP-hard. Traditional approaches, primarily focused on graph partitioning strategies, have sought to reduce the number of required Bell pairs for executing non-local CNOT operations, a form of gate teleportation. However, these methods have limitations in terms of efficiency and scalability. Addressing this, our work jointly considers gate and qubit teleportations introducing a novel meta-heuristic algorithm to minimise the network cost of executing a quantum circuit. By allowing dynamic reallocation of qubits along with gate teleportations during circuit execution, our method significantly enhances the overall efficacy and potential scalability of DQC frameworks. In our numerical analysis, we demonstrate that integrating qubit teleportations into our genetic algorithm for optimising circuit blocking reduces the required resources, specifically the number of EPR pairs, compared to traditional graph partitioning methods. Our results, derived from both benchmark and randomly generated circuits, show that as circuit complexity increases - demanding more qubit teleportations - our approach effectively optimises these teleportations throughout the execution, thereby enhancing performance through strategic circuit partitioning. This is a step forward in the pursuit of a global quantum compiler which will ultimately enable the efficient use of a 'quantum data center' in the future., Comment: 9 pages, 9 figures
- Published
- 2024
4. Compiler for Distributed Quantum Computing: a Reinforcement Learning Approach
- Author
-
Promponas, Panagiotis, Mudvari, Akrit, Della Chiesa, Luca, Polakos, Paul, Samuel, Louis, and Tassiulas, Leandros
- Subjects
Quantum Physics ,Computer Science - Networking and Internet Architecture - Abstract
The practical realization of quantum programs that require large-scale qubit systems is hindered by current technological limitations. Distributed Quantum Computing (DQC) presents a viable path to scalability by interconnecting multiple Quantum Processing Units (QPUs) through quantum links, facilitating the distributed execution of quantum circuits. In DQC, EPR pairs are generated and shared between distant QPUs, which enables quantum teleportation and facilitates the seamless execution of circuits. A primary obstacle in DQC is the efficient mapping and routing of logical qubits to physical qubits across different QPUs, necessitating sophisticated strategies to overcome hardware constraints and optimize communication. We introduce a novel compiler that, unlike existing approaches, prioritizes reducing the expected execution time by jointly managing the generation and routing of EPR pairs, scheduling remote operations, and injecting SWAP gates to facilitate the execution of local gates. We present a real-time, adaptive approach to compiler design, accounting for the stochastic nature of entanglement generation and the operational demands of quantum circuits. Our contributions are twofold: (i) we model the optimal compiler for DQC using a Markov Decision Process (MDP) formulation, establishing the existence of an optimal algorithm, and (ii) we introduce a constrained Reinforcement Learning (RL) method to approximate this optimal compiler, tailored to the complexities of DQC environments. Our simulations demonstrate that Double Deep Q-Networks (DDQNs) are effective in learning policies that minimize the depth of the compiled circuit, leading to a lower expected execution time and likelihood of successful operation before qubits decohere.
- Published
- 2024
5. Adaptive Heterogeneous Client Sampling for Federated Learning over Wireless Networks
- Author
-
Luo, Bing, Xiao, Wenli, Wang, Shiqiang, Huang, Jianwei, and Tassiulas, Leandros
- Subjects
Computer Science - Distributed, Parallel, and Cluster Computing ,Computer Science - Machine Learning ,Computer Science - Networking and Internet Architecture ,Electrical Engineering and Systems Science - Systems and Control - Abstract
Federated learning (FL) algorithms usually sample a fraction of clients in each round (partial participation) when the number of participants is large and the server's communication bandwidth is limited. Recent works on the convergence analysis of FL have focused on unbiased client sampling, e.g., sampling uniformly at random, which suffers from slow wall-clock time for convergence due to high degrees of system heterogeneity and statistical heterogeneity. This paper aims to design an adaptive client sampling algorithm for FL over wireless networks that tackles both system and statistical heterogeneity to minimize the wall-clock convergence time. We obtain a new tractable convergence bound for FL algorithms with arbitrary client sampling probability. Based on the bound, we analytically establish the relationship between the total learning time and sampling probability with an adaptive bandwidth allocation scheme, which results in a non-convex optimization problem. We design an efficient algorithm for learning the unknown parameters in the convergence bound and develop a low-complexity algorithm to approximately solve the non-convex problem. Our solution reveals the impact of system and statistical heterogeneity parameters on the optimal client sampling design. Moreover, our solution shows that as the number of sampled clients increases, the total convergence time first decreases and then increases because a larger sampling number reduces the number of rounds for convergence but results in a longer expected time per-round due to limited wireless bandwidth. Experimental results from both hardware prototype and simulation demonstrate that our proposed sampling scheme significantly reduces the convergence time compared to several baseline sampling schemes., Comment: Published in IEEE Transactions on Mobile Computing (TMC). arXiv admin note: substantial text overlap with arXiv:2112.11256
- Published
- 2024
6. Predictive Handover Strategy in 6G and Beyond: A Deep and Transfer Learning Approach
- Author
-
Panitsas, Ioannis, Mudvari, Akrit, Maatouk, Ali, and Tassiulas, Leandros
- Subjects
Computer Science - Networking and Internet Architecture ,Computer Science - Artificial Intelligence - Abstract
Next-generation cellular networks will evolve into more complex and virtualized systems, employing machine learning for enhanced optimization and leveraging higher frequency bands and denser deployments to meet varied service demands. This evolution, while bringing numerous advantages, will also pose challenges, especially in mobility management, as it will increase the overall number of handovers due to smaller coverage areas and the higher signal attenuation. To address these challenges, we propose a deep learning based algorithm for predicting the future serving cell utilizing sequential user equipment measurements to minimize the handover failures and interruption time. Our algorithm enables network operators to dynamically adjust handover triggering events or incorporate UAV base stations for enhanced coverage and capacity, optimizing network objectives like load balancing and energy efficiency through transfer learning techniques. Our framework complies with the O-RAN specifications and can be deployed in a Near-Real-Time RAN Intelligent Controller as an xApp leveraging the E2SM-KPM service model. The evaluation results demonstrate that our algorithm achieves a 92% accuracy in predicting future serving cells with high probability. Finally, by utilizing transfer learning, our algorithm significantly reduces the retraining time by 91% and 77% when new handover trigger decisions or UAV base stations are introduced to the network dynamically.
- Published
- 2024
7. From Similarity to Superiority: Channel Clustering for Time Series Forecasting
- Author
-
Chen, Jialin, Lenssen, Jan Eric, Feng, Aosong, Hu, Weihua, Fey, Matthias, Tassiulas, Leandros, Leskovec, Jure, and Ying, Rex
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence - Abstract
Time series forecasting has attracted significant attention in recent decades. Previous studies have demonstrated that the Channel-Independent (CI) strategy improves forecasting performance by treating different channels individually, while it leads to poor generalization on unseen instances and ignores potentially necessary interactions between channels. Conversely, the Channel-Dependent (CD) strategy mixes all channels with even irrelevant and indiscriminate information, which, however, results in oversmoothing issues and limits forecasting accuracy. There is a lack of channel strategy that effectively balances individual channel treatment for improved forecasting performance without overlooking essential interactions between channels. Motivated by our observation of a correlation between the time series model's performance boost against channel mixing and the intrinsic similarity on a pair of channels, we developed a novel and adaptable Channel Clustering Module (CCM). CCM dynamically groups channels characterized by intrinsic similarities and leverages cluster identity instead of channel identity, combining the best of CD and CI worlds. Extensive experiments on real-world datasets demonstrate that CCM can (1) boost the performance of CI and CD models by an average margin of 2.4% and 7.2% on long-term and short-term forecasting, respectively; (2) enable zero-shot forecasting with mainstream time series forecasting models; (3) uncover intrinsic time series patterns among channels and improve interpretability of complex time series models., Comment: 20 pages, 6 figures
- Published
- 2024
8. Machine Learning on Blockchain Data: A Systematic Mapping Study
- Author
-
Palaiokrassas, Georgios, Bouraga, Sarah, and Tassiulas, Leandros
- Subjects
Computer Science - Cryptography and Security ,Computer Science - Machine Learning - Abstract
Context: Blockchain technology has drawn growing attention in the literature and in practice. Blockchain technology generates considerable amounts of data and has thus been a topic of interest for Machine Learning (ML). Objective: The objective of this paper is to provide a comprehensive review of the state of the art on machine learning applied to blockchain data. This work aims to systematically identify, analyze, and classify the literature on ML applied to blockchain data. This will allow us to discover the fields where more effort should be placed in future research. Method: A systematic mapping study has been conducted to identify the relevant literature. Ultimately, 159 articles were selected and classified according to various dimensions, specifically, the domain use case, the blockchain, the data, and the machine learning models. Results: The majority of the papers (49.7%) fall within the Anomaly use case. Bitcoin (47.2%) was the blockchain that drew the most attention. A dataset consisting of more than 1.000.000 data points was used by 31.4% of the papers. And Classification (46.5%) was the ML task most applied to blockchain data. Conclusion: The results confirm that ML applied to blockchain data is a relevant and a growing topic of interest both in the literature and in practice. Nevertheless, some open challenges and gaps remain, which can lead to future research directions. Specifically, we identify novel machine learning algorithms, the lack of a standardization framework, blockchain scalability issues and cross-chain interactions as areas worth exploring in the future.
- Published
- 2024
9. Efficient High-Resolution Time Series Classification via Attention Kronecker Decomposition
- Author
-
Feng, Aosong, Chen, Jialin, Garza, Juan, Berry, Brooklyn, Salazar, Francisco, Gao, Yifeng, Ying, Rex, and Tassiulas, Leandros
- Subjects
Computer Science - Machine Learning - Abstract
The high-resolution time series classification problem is essential due to the increasing availability of detailed temporal data in various domains. To tackle this challenge effectively, it is imperative that the state-of-the-art attention model is scalable to accommodate the growing sequence lengths typically encountered in high-resolution time series data, while also demonstrating robustness in handling the inherent noise prevalent in such datasets. To address this, we propose to hierarchically encode the long time series into multiple levels based on the interaction ranges. By capturing relationships at different levels, we can build more robust, expressive, and efficient models that are capable of capturing both short-term fluctuations and long-term trends in the data. We then propose a new time series transformer backbone (KronTime) by introducing Kronecker-decomposed attention to process such multi-level time series, which sequentially calculates attention from the lower level to the upper level. Experiments on four long time series datasets demonstrate superior classification results with improved efficiency compared to baseline methods.
- Published
- 2024
10. An Item is Worth a Prompt: Versatile Image Editing with Disentangled Control
- Author
-
Feng, Aosong, Qiu, Weikang, Bai, Jinbin, Zhang, Xiao, Dong, Zhen, Zhou, Kaicheng, Ying, Rex, and Tassiulas, Leandros
- Subjects
Computer Science - Computer Vision and Pattern Recognition - Abstract
Building on the success of text-to-image diffusion models (DPMs), image editing is an important application to enable human interaction with AI-generated content. Among various editing methods, editing within the prompt space gains more attention due to its capacity and simplicity of controlling semantics. However, since diffusion models are commonly pretrained on descriptive text captions, direct editing of words in text prompts usually leads to completely different generated images, violating the requirements for image editing. On the other hand, existing editing methods usually consider introducing spatial masks to preserve the identity of unedited regions, which are usually ignored by DPMs and therefore lead to inharmonic editing results. Targeting these two challenges, in this work, we propose to disentangle the comprehensive image-prompt interaction into several item-prompt interactions, with each item linked to a special learned prompt. The resulting framework, named D-Edit, is based on pretrained diffusion models with cross-attention layers disentangled and adopts a two-step optimization to build item-prompt associations. Versatile image editing can then be applied to specific items by manipulating the corresponding prompts. We demonstrate state-of-the-art results in four types of editing operations including image-based, text-based, mask-based editing, and item removal, covering most types of editing applications, all within a single unified framework. Notably, D-Edit is the first framework that can (1) achieve item editing through mask editing and (2) combine image and text-based editing. We demonstrate the quality and versatility of the editing results for a diverse collection of images through both qualitative and quantitative evaluations.
- Published
- 2024
11. Cyber-Twin: Digital Twin-boosted Autonomous Attack Detection for Vehicular Ad-Hoc Networks
- Author
-
Yigit, Yagmur, Panitsas, Ioannis, Maglaras, Leandros, Tassiulas, Leandros, and Canberk, Berk
- Subjects
Computer Science - Cryptography and Security - Abstract
The rapid evolution of Vehicular Ad-hoc NETworks (VANETs) has ushered in a transformative era for intelligent transportation systems (ITS), significantly enhancing road safety and vehicular communication. However, the intricate and dynamic nature of VANETs presents formidable challenges, particularly in vehicle-to-infrastructure (V2I) communications. Roadside Units (RSUs), integral components of VANETs, are increasingly susceptible to cyberattacks, such as jamming and distributed denial of service (DDoS) attacks. These vulnerabilities pose grave risks to road safety, potentially leading to traffic congestion and vehicle malfunctions. Existing methods face difficulties in detecting dynamic attacks and integrating digital twin technology and artificial intelligence (AI) models to enhance VANET cybersecurity. Our study proposes a novel framework that combines digital twin technology with AI to enhance the security of RSUs in VANETs and address this gap. This framework enables real-time monitoring and efficient threat detection while also improving computational efficiency and reducing data transmission delay for increased energy efficiency and hardware durability. Our framework outperforms existing solutions in resource management and attack detection. It reduces RSU load and data transmission delay while achieving an optimal balance between resource consumption and high attack detection effectiveness. This highlights our commitment to secure and sustainable vehicular communication systems for smart cities., Comment: 6 pages, 5 figures, IEEE International Conference on Communications (ICC) 2024
- Published
- 2024
12. Constrained Reinforcement Learning for Adaptive Controller Synchronization in Distributed SDN
- Author
-
Panitsas, Ioannis, Mudvari, Akrit, and Tassiulas, Leandros
- Subjects
Computer Science - Networking and Internet Architecture ,Computer Science - Artificial Intelligence - Abstract
In software-defined networking (SDN), the implementation of distributed SDN controllers, with each controller responsible for managing a specific sub-network or domain, plays a critical role in achieving a balance between centralized control, scalability, reliability, and network efficiency. These controllers must be synchronized to maintain a logically centralized view of the entire network. While there are various approaches for synchronizing distributed SDN controllers, most tend to prioritize goals such as optimization of communication latency or load balancing, often neglecting to address both the aspects simultaneously. This limitation becomes particularly significant when considering applications like Augmented and Virtual Reality (AR/VR), which demand constrained network latencies and substantial computational resources. Additionally, many existing studies in this field predominantly rely on value-based reinforcement learning (RL) methods, overlooking the potential advantages offered by state-of-the-art policy-based RL algorithms. To bridge this gap, our work focuses on examining deep reinforcement learning (DRL) techniques, encompassing both value-based and policy-based methods, to guarantee an upper latency threshold for AR/VR task offloading within SDN environments, while selecting the most cost-effective servers for AR/VR task offloading. Our evaluation results indicate that while value-based methods excel in optimizing individual network metrics such as latency or load balancing, policy-based approaches exhibit greater robustness in adapting to sudden network changes or reconfiguration.
- Published
- 2024
13. Joint SDN Synchronization and Controller Placement in Wireless Networks using Deep Reinforcement Learning
- Author
-
Mudvari, Akrit and Tassiulas, Leandros
- Subjects
Computer Science - Networking and Internet Architecture - Abstract
Software Defined Networking has afforded numerous benefits to the network users but there are certain persisting issues with this technology, two of which are scalability and privacy. The natural solution to overcoming these limitations is a distributed SDN controller architecture where multiple controllers are deployed over the network, with each controller orchestrating a certain segment of the network. However, since the centralized control is the key attribute of SDN that allows it to be so beneficial, a centralized logical view of the network will have to be maintained by each of these controllers; this can be done through synchronization of the distributed controllers, where each controller communicates with the others to ensure that they remain informed about the entire network. There is however a network cost associated with constantly having to update each others about different aspects of the network, which will become a greater issue in dynamic wireless networks. To minimize this network cost, there is a need to consider not only when to get the update information from the neighboring controllers, but also where to dynamically place the controllers such that the network costs may be minimized. The placement should take into consideration both communication for synchronization among the distributed controllers and communication of the controllers with the network devices that they manage. In this work, we show that our multi-objective deep reinforcement learning-based method performs the best at achieving different application goals by developing policy for controller synchronization as well as placement, outperforming different other possible approaches, under a wide variety of network conditions., Comment: Submitted to IEEE NOMS'24
- Published
- 2023
14. Adaptive Compression-Aware Split Learning and Inference for Enhanced Network Efficiency
- Author
-
Mudvari, Akrit, Vainio, Antero, Ofeidis, Iason, Tarkoma, Sasu, and Tassiulas, Leandros
- Subjects
Computer Science - Networking and Internet Architecture ,Computer Science - Machine Learning - Abstract
The growing number of AI-driven applications in mobile devices has led to solutions that integrate deep learning models with the available edge-cloud resources. Due to multiple benefits such as reduction in on-device energy consumption, improved latency, improved network usage, and certain privacy improvements, split learning, where deep learning models are split away from the mobile device and computed in a distributed manner, has become an extensively explored topic. Incorporating compression-aware methods (where learning adapts to compression level of the communicated data) has made split learning even more advantageous. This method could even offer a viable alternative to traditional methods, such as federated learning techniques. In this work, we develop an adaptive compression-aware split learning method ('deprune') to improve and train deep learning models so that they are much more network-efficient, which would make them ideal to deploy in weaker devices with the help of edge-cloud resources. This method is also extended ('prune') to very quickly train deep learning models through a transfer learning approach, which trades off little accuracy for much more network-efficient inference abilities. We show that the 'deprune' method can reduce network usage by 4x when compared with a split-learning approach (that does not use our method) without loss of accuracy, while also improving accuracy over compression-aware split-learning by 4 percent. Lastly, we show that the 'prune' method can reduce the training time for certain models by up to 6x without affecting the accuracy when compared against a compression-aware split-learning approach.
- Published
- 2023
15. Mobility as a Resource (MaaR) for resilient human-centric automation: a vision paper
- Author
-
Waller, S. Travis, Polydoropoulou, Amalia, Tassiulas, Leandros, Ziliaskopoulos, Athanasios, Jian, Sisi, Wagenknecht, Susann, Hirte, Georg, Ukkusuri, Satish, Ramadurai, Gitakrishnan, and Bednarz, Tomasz
- Subjects
Electrical Engineering and Systems Science - Systems and Control - Abstract
With technological advances, mobility has been moving from a product (i.e., traditional modes and vehicles), to a service (i.e., Mobility as a Service, MaaS). However, as observed in other fields (e.g. cloud computing resource management) we argue that mobility will evolve from a service to a resource (i.e., Mobility as a Resource, MaaR). Further, due to increasing scarcity of shared mobility spaces across traditional and emerging modes, the transition must be viewed within the critical need for ethical and equitable solutions for the traveling public (i.e., research is needed to avoid hyper-market driven outcomes for society). The evolution of mobility into a resource requires novel conceptual frameworks, technologies, processes and perspectives of analysis. A key component of the future MaaR system is the technological capacity to observe, allocate and manage (in real-time) the smallest envisionable units of mobility (i.e., atomic units of mobility capacity) while providing prioritized attention to human movement and ethical metrics related to access, consumption and impact. To facilitate research into the envisioned future system, this paper proposes initial frameworks which synthesize and advance methodologies relating to highly dynamic capacity reservation systems. Future research requires synthesis across transport network management, demand behavior, mixed-mode usage, and equitable mobility.
- Published
- 2023
16. Age Optimum Sampling in Non-Stationary Environment
- Author
-
Zhang, Jinheng, Tang, Haoyue, Wang, Jintao, Kompella, Sastry, and Tassiulas, Leandros
- Subjects
Computer Science - Networking and Internet Architecture ,Electrical Engineering and Systems Science - Signal Processing - Abstract
In this work, we consider a status update system with a sensor and a receiver. The status update information is sampled by the sensor and then forwarded to the receiver through a channel with non-stationary delay distribution. The data freshness at the receiver is quantified by the Age-of-Information (AoI). The goal is to design an online sampling strategy that can minimize the average AoI when the non-stationary delay distribution is unknown. Assuming that channel delay distribution may change over time, to minimize the average AoI, we propose a joint stochastic approximation and non-parametric change point detection algorithm that can: (1) learn the optimum update threshold when the delay distribution remains static; (2) detect the change in transmission delay distribution quickly and then restart the learning process. Simulation results show that the proposed algorithm can quickly detect the delay changes, and the average AoI obtained by the proposed policy converges to the minimum AoI.
- Published
- 2023
17. Optimization methods for the capacitated refueling station location problem with routing
- Author
-
Nordlund, Nicholas, Tassiulas, Leandros, and Lange, Jan-Hendrik
- Subjects
Mathematics - Optimization and Control - Abstract
The energy transition in transportation benefits from demand-based models to determine the optimal placement of refueling stations for alternative fuel vehicles such as battery electric trucks. A formulation known as the refueling station location problem with routing (RSLP-R) is concerned with minimizing the number of stations necessary to cover a set of origin-destination trips such that the transit time does not exceed a given threshold. In this paper we extend the RSLP-R by station capacities to limit the number of vehicles that can be refueled at individual stations. The solution to the capacitated RSLP-R (CRSLP-R) avoids congestion of refueling stations by satisfying capacity constraints. We devise two optimization methods to deal with the increased difficulty to solve the CRSLP-R. The first method extends a prior branch-and-cut approach and the second method is a branch-cut-and-price algorithm based on variables associated with feasible routes. We evaluate both our methods on instances from the literature as well as a newly constructed network and find that the relative performance of the algorithms depends on the strictness of the capacity constraints. Furthermore, we show some runtime improvements over prior work on uncapacitated instances.
- Published
- 2023
18. Sampling for Remote Estimation of an Ornstein-Uhlenbeck Process through Channel with Unknown Delay Statistics
- Author
-
Chen, Yuchao, Tang, Haoyue, Wang, Jintao, Yang, Pengkun, and Tassiulas, Leandros
- Subjects
Computer Science - Information Theory ,Computer Science - Networking and Internet Architecture - Abstract
In this paper, we consider sampling an Ornstein-Uhlenbeck (OU) process through a channel for remote estimation. The goal is to minimize the mean square error (MSE) at the estimator under a sampling frequency constraint when the channel delay statistics is unknown. Sampling for MSE minimization is reformulated into an optimal stopping problem. By revisiting the threshold structure of the optimal stopping policy when the delay statistics is known, we propose an online sampling algorithm to learn the optimum threshold using stochastic approximation algorithm and the virtual queue method. We prove that with probability 1, the MSE of the proposed online algorithm converges to the minimum MSE that is achieved when the channel delay statistics is known. The cumulative MSE gap of our proposed algorithm compared with the minimum MSE up to the $(k+1)$-th sample grows with rate at most $\mathcal{O}(\ln k)$. Our proposed online algorithm can satisfy the sampling frequency constraint theoretically. Finally, simulation results are provided to demonstrate the performance of the proposed algorithm., Comment: Accepted and to appear, JCN special issues
- Published
- 2023
19. Optimizing Sectorized Wireless Networks: Model, Analysis, and Algorithm
- Author
-
Promponas, Panagiotis, Chen, Tingjun, and Tassiulas, Leandros
- Subjects
Computer Science - Networking and Internet Architecture - Abstract
Future wireless networks need to support the increasing demands for high data rates and improved coverage. One promising solution is sectorization, where an infrastructure node (e.g., a base station) is equipped with multiple sectors employing directional communication. Although the concept of sectorization is not new, it is critical to fully understand the potential of sectorized networks, such as the rate gain achieved when multiple sectors can be simultaneously activated. In this paper, we focus on sectorized wireless networks, where sectorized infrastructure nodes with beam-steering capabilities form a multi-hop mesh network for data forwarding and routing. We present a sectorized node model and characterize the capacity region of these sectorized networks. We define the flow extension ratio and the corresponding sectorization gain, which quantitatively measure the performance gain introduced by node sectorization as a function of the network flow. Our objective is to find the optimal sectorization of each node that achieves the maximum flow extension ratio, and thus the sectorization gain. Towards this goal, we formulate the corresponding optimization problem and develop an efficient distributed algorithm that obtains the node sectorization under a given network flow with an approximation ratio of 2/3. Through extensive simulations, we evaluate the sectorization gain and the performance of the proposed algorithm in various network scenarios with varying network flows. The simulation results show that the approximate sectorization gain increases sublinearly as a function of the number of sectors per node.
- Published
- 2023
20. Leveraging Machine Learning for Multichain DeFi Fraud Detection
- Author
-
Palaiokrassas, Georgios, Scherrers, Sandro, Ofeidis, Iason, and Tassiulas, Leandros
- Subjects
Quantitative Finance - General Finance ,Computer Science - Cryptography and Security ,Computer Science - Machine Learning - Abstract
Since the inception of permissionless blockchains with Bitcoin in 2008, it became apparent that their most well-suited use case is related to making the financial system and its advantages available to everyone seamlessly without depending on any trusted intermediaries. Smart contracts across chains provide an ecosystem of decentralized finance (DeFi), where users can interact with lending pools, Automated Market Maker (AMM) exchanges, stablecoins, derivatives, etc. with a cumulative locked value which had exceeded 160B USD. While DeFi comes with high rewards, it also carries plenty of risks. Many financial crimes have occurred over the years making the early detection of malicious activity an issue of high priority. The proposed framework introduces an effective method for extracting a set of features from different chains, including the largest one, Ethereum and it is evaluated over an extensive dataset we gathered with the transactions of the most widely used DeFi protocols (23 in total, including Aave, Compound, Curve, Lido, and Yearn) based on a novel dataset in collaboration with Covalent. Different Machine Learning methods were employed, such as XGBoost and a Neural Network for identifying fraud accounts detection interacting with DeFi and we demonstrate that the introduction of novel DeFi-related features, significantly improves the evaluation results, where Accuracy, Precision, Recall, F1-score and F2-score where utilized.
- Published
- 2023
21. Full Exploitation of Limited Memory in Quantum Entanglement Switching
- Author
-
Promponas, Panagiotis, Valls, Víctor, and Tassiulas, Leandros
- Subjects
Computer Science - Networking and Internet Architecture - Abstract
We study the problem of operating a quantum switch with memory constraints. In particular, the switch has to allocate quantum memories to clients to generate link-level entanglements (LLEs), and then use these to serve end-to-end entanglements requests. The paper's main contributions are (i) to characterize the switch's capacity region, and (ii) to propose a memory allocation policy (MEW) that is throughput optimal. The worst-case time complexity of MEW is exponential on the system parameters. However, when the requests are bipartite and the LLE attempts are always successful, we propose a variant of MEW (MEW2) that has polynomial time complexity. We evaluate the proposed policies numerically and illustrate their performance depending on the requests arrivals characteristics and the time available to obtain a memory allocation.
- Published
- 2023
22. Incentive Mechanism Design for Unbiased Federated Learning with Randomized Client Participation
- Author
-
Luo, Bing, Feng, Yutong, Wang, Shiqiang, Huang, Jianwei, and Tassiulas, Leandros
- Subjects
Computer Science - Computer Science and Game Theory ,Computer Science - Artificial Intelligence ,Computer Science - Machine Learning - Abstract
Incentive mechanism is crucial for federated learning (FL) when rational clients do not have the same interests in the global model as the server. However, due to system heterogeneity and limited budget, it is generally impractical for the server to incentivize all clients to participate in all training rounds (known as full participation). The existing FL incentive mechanisms are typically designed by stimulating a fixed subset of clients based on their data quantity or system resources. Hence, FL is performed only using this subset of clients throughout the entire training process, leading to a biased model because of data heterogeneity. This paper proposes a game theoretic incentive mechanism for FL with randomized client participation, where the server adopts a customized pricing strategy that motivates different clients to join with different participation levels (probabilities) for obtaining an unbiased and high performance model. Each client responds to the server's monetary incentive by choosing its best participation level, to maximize its profit based on not only the incurred local cost but also its intrinsic value for the global model. To effectively evaluate clients' contribution to the model performance, we derive a new convergence bound which analytically predicts how clients' arbitrary participation levels and their heterogeneous data affect the model performance. By solving a non-convex optimization problem, our analysis reveals that the intrinsic value leads to the interesting possibility of bidirectional payment between the server and clients. Experimental results using real datasets on a hardware prototype demonstrate the superiority of our mechanism in achieving higher model performance for the server as well as higher profits for the clients., Comment: Accepted in ICDCS 2023
- Published
- 2023
23. Monetary Policy, Digital Assets, and DeFi Activity
- Author
-
Kyriazis, Antzelos, Ofeidis, Iason, Palaiokrassas, Georgios, and Tassiulas, Leandros
- Subjects
Quantitative Finance - Statistical Finance - Abstract
This paper studies the effects of unexpected changes in US monetary policy on digital asset returns. We use event study regressions and find that monetary policy surprises negatively affect BTC and ETH, the two largest digital assets, but do not significantly affect the rest of the market. Second, we use high-frequency price data to examine the effect of the FOMC statements release and Minutes release on the prices of the assets with the higher collateral usage on the Ethereum Blockchain Decentralized Finance (DeFi) ecosystem. The FOMC statement release strongly affects the volatility of digital asset returns, while the effect of the Minutes release is weaker. The volatility effect strengthened after December 2021, when the Federal Reserve changed its policy to fight inflation. We also show that some borrowing interest rates in the Ethereum DeFi ecosystem are affected positively by unexpected changes in monetary policy. In contrast, the debt outstanding and the total value locked are negatively affected. Finally, we utilize a local Ethereum Blockchain node to record the activity history of primary DeFi functions, such as depositing, borrowing, and liquidating, and study how these are influenced by the FOMC announcements over time., Comment: 33 pages, 11 figures, 9 tables
- Published
- 2023
24. Network Slicing: Market Mechanism and Competitive Equilibria
- Author
-
Promponas, Panagiotis and Tassiulas, Leandros
- Subjects
Computer Science - Networking and Internet Architecture - Abstract
Towards addressing spectral scarcity and enhancing resource utilization in 5G networks, network slicing is a promising technology to establish end-to-end virtual networks without requiring additional infrastructure investments. By leveraging Software Defined Networks (SDN) and Network Function Virtualization (NFV), we can realize slices completely isolated and dedicated to satisfy the users' diverse Quality of Service (QoS) prerequisites and Service Level Agreements (SLAs). This paper focuses on the technical and economic challenges that emerge from the application of the network slicing architecture to real-world scenarios. We consider a market where multiple Network Providers (NPs) own the physical infrastructure and offer their resources to multiple Service Providers (SPs). Then, the SPs offer those resources as slices to their associated users. We propose a holistic iterative model for the network slicing market along with a clock auction that converges to a robust $\epsilon$-competitive equilibrium. At the end of each cycle of the market, the slices are reconfigured and the SPs aim to learn the private parameters of their users. Numerical results are provided that validate and evaluate the convergence of the clock auction and the capability of the proposed market architecture to express the incentives of the different entities of the system.
- Published
- 2023
- Full Text
- View/download PDF
25. Robust and Resource-efficient Machine Learning Aided Viewport Prediction in Virtual Reality
- Author
-
Jiang, Yuang, Poularakis, Konstantinos, Kiedanski, Diego, Kompella, Sastry, and Tassiulas, Leandros
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Artificial Intelligence - Abstract
360-degree panoramic videos have gained considerable attention in recent years due to the rapid development of head-mounted displays (HMDs) and panoramic cameras. One major problem in streaming panoramic videos is that panoramic videos are much larger in size compared to traditional ones. Moreover, the user devices are often in a wireless environment, with limited battery, computation power, and bandwidth. To reduce resource consumption, researchers have proposed ways to predict the users' viewports so that only part of the entire video needs to be transmitted from the server. However, the robustness of such prediction approaches has been overlooked in the literature: it is usually assumed that only a few models, pre-trained on past users' experiences, are applied for prediction to all users. We observe that those pre-trained models can perform poorly for some users because they might have drastically different behaviors from the majority, and the pre-trained models cannot capture the features in unseen videos. In this work, we propose a novel meta learning based viewport prediction paradigm to alleviate the worst prediction performance and ensure the robustness of viewport prediction. This paradigm uses two machine learning models, where the first model predicts the viewing direction, and the second model predicts the minimum video prefetch size that can include the actual viewport. We first train two meta models so that they are sensitive to new training data, and then quickly adapt them to users while they are watching the videos. Evaluation results reveal that the meta models can adapt quickly to each user, and can significantly increase the prediction accuracy, especially for the worst-performing predictions., Comment: Accepted for publication in 2022 IEEE International Conference on Big Data (IEEE BigData 2022)
- Published
- 2022
26. On the Capacity Region of a Quantum Switch with Entanglement Purification
- Author
-
Panigrahy, Nitish K., Vasantam, Thirupathaiah, Towsley, Don, and Tassiulas, Leandros
- Subjects
Quantum Physics ,Computer Science - Networking and Internet Architecture ,Computer Science - Performance - Abstract
Quantum switches are envisioned to be an integral component of future entanglement distribution networks. They can provide high quality entanglement distribution service to end-users by performing quantum operations such as entanglement swapping and entanglement purification. In this work, we characterize the capacity region of such a quantum switch under noisy channel transmissions and imperfect quantum operations. We express the capacity region as a function of the channel and network parameters (link and entanglement swap success probability), entanglement purification yield and application level parameters (target fidelity threshold). In particular, we provide necessary conditions to verify if a set of request rates belong to the capacity region of the switch. We use these conditions to find the maximum achievable end-to-end user entanglement generation throughput by solving a set of linear optimization problems. We develop a max-weight scheduling policy and prove that the policy stabilizes the switch for all feasible request arrival rates. As we develop scheduling policies, we also generate new results for computing the conditional yield distribution of different classes of purification protocols. From numerical experiments, we discover that performing link-level entanglement purification followed by entanglement swaps provides a larger capacity region than doing entanglement swaps followed by end-to-end entanglement purification. The conclusions obtained in this work can yield useful guidelines for subsequent quantum switch designs., Comment: 10 pages, 4 figures, accepted for a talk at the IEEE International Conference on Computer Communications (INFOCOM), 2023
- Published
- 2022
27. Deep Reinforcement Learning-based Rebalancing Policies for Profit Maximization of Relay Nodes in Payment Channel Networks
- Author
-
Papadis, Nikolaos and Tassiulas, Leandros
- Subjects
Computer Science - Distributed, Parallel, and Cluster Computing ,Computer Science - Cryptography and Security ,Computer Science - Machine Learning ,Computer Science - Networking and Internet Architecture ,Electrical Engineering and Systems Science - Systems and Control - Abstract
Payment channel networks (PCNs) are a layer-2 blockchain scalability solution, with its main entity, the payment channel, enabling transactions between pairs of nodes "off-chain," thus reducing the burden on the layer-1 network. Nodes with multiple channels can serve as relays for multihop payments by providing their liquidity and withholding part of the payment amount as a fee. Relay nodes might after a while end up with one or more unbalanced channels, and thus need to trigger a rebalancing operation. In this paper, we study how a relay node can maximize its profits from fees by using the rebalancing method of submarine swaps. We introduce a stochastic model to capture the dynamics of a relay node observing random transaction arrivals and performing occasional rebalancing operations, and express the system evolution as a Markov Decision Process. We formulate the problem of the maximization of the node's fortune over time over all rebalancing policies, and approximate the optimal solution by designing a Deep Reinforcement Learning (DRL)-based rebalancing policy. We build a discrete event simulator of the system and use it to demonstrate the DRL policy's superior performance under most conditions by conducting a comparative study of different policies and parameterizations. Our work is the first to introduce DRL for liquidity management in the complex world of PCNs., Comment: Best Paper Award at the 4th International Conference on Mathematical Research for the Blockchain Economy (MARBLE 2023). 28 pages; minor language edits and fixes; acknowledgments added; results unchanged
- Published
- 2022
28. A Quantitative Theory of Bottleneck Structures for Data Networks
- Author
-
Ros-Giralt, Jordi, Amsel, Noah, Yellamraju, Sruthi, Ezick, James, Lethin, Richard, Jiang, Yuang, Feng, Aosong, and Tassiulas, Leandros
- Subjects
Computer Science - Networking and Internet Architecture - Abstract
The conventional view of the congestion control problem in data networks is based on the principle that a flow's performance is uniquely determined by the state of its bottleneck link, regardless of the topological properties of the network. However, recent work has shown that the behavior of congestion-controlled networks is better explained by models that account for the interactions between bottleneck links. These interactions are captured by a latent \textit{bottleneck structure}, a model describing the complex ripple effects that changes in one part of the network exert on the other parts. In this paper, we present a \textit{quantitative} theory of bottleneck structures (QTBS), a mathematical and engineering framework comprising a family of polynomial-time algorithms that can be used to reason about a wide variety of network optimization problems, including routing, capacity planning and flow control. QTBS can contribute to traffic engineering by making clear predictions about the relative performance of alternative flow routes, and by providing numerical recommendations for the optimal rate settings of traffic shapers. A particularly novel result in the domain of capacity planning indicates that previously established rules for the design of folded-Clos networks are suboptimal when flows are congestion controlled. We show that QTBS can be used to derive the optimal rules for this important class of topologies, and empirically demonstrate the correctness and efficacy of these results using the BBR and Cubic congestion-control algorithms.
- Published
- 2022
29. An Overview of the Data-Loader Landscape: Comparative Performance Analysis
- Author
-
Ofeidis, Iason, Kiedanski, Diego, and Tassiulas, Leandros
- Subjects
Computer Science - Distributed, Parallel, and Cluster Computing ,Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Machine Learning ,Computer Science - Performance - Abstract
Dataloaders, in charge of moving data from storage into GPUs while training machine learning models, might hold the key to drastically improving the performance of training jobs. Recent advances have shown promise not only by considerably decreasing training time but also by offering new features such as loading data from remote storage like S3. In this paper, we are the first to distinguish the dataloader as a separate component in the Deep Learning (DL) workflow and to outline its structure and features. Finally, we offer a comprehensive comparison of the different dataloading libraries available, their trade-offs in terms of functionality, usability, and performance and the insights derived from them., Comment: 17 pages, 28 figures
- Published
- 2022
30. Fundamentals of Clustered Molecular Nanonetworks
- Author
-
Azimi-Abarghouyi, Seyed Mohammad, Dhillon, Harpreet S., and Tassiulas, Leandros
- Subjects
Computer Science - Information Theory ,Mathematics - Probability ,Mathematics - Statistics Theory - Abstract
We present a comprehensive approach to the modeling, performance analysis, and design of clustered molecular nanonetworks in which nano-machines of different clusters release an appropriate number of molecules to transmit their sensed information to their respective fusion centers. The fusion centers decode this information by counting the number of molecules received in the given time slot. Owing to the propagation properties of the biological media, this setup suffers from both inter- and intra-cluster interference that needs to be carefully modeled. To facilitate rigorous analysis, we first develop a novel spatial model for this setup by modeling nano-machines as a Poisson cluster process with the fusion centers forming its parent point process. For this setup, we first derive a new set of distance distributions in the three-dimensional space, resulting in a remarkably simple result for the special case of the Thomas cluster process. Using this, total interference from previous symbols and different clusters is characterized and its expected value and Laplace transform are obtained. The error probability of a simple detector suitable for biological applications is analyzed, and approximate and upper-bound results are provided. The impact of different parameters on the performance is also investigated., Comment: Accepted for publication
- Published
- 2022
31. Accelerated Convex Optimization with Stochastic Gradients: Generalizing the Strong-Growth Condition
- Author
-
Valls, Víctor, Wang, Shiqiang, Jiang, Yuang, and Tassiulas, Leandros
- Subjects
Mathematics - Optimization and Control - Abstract
This paper presents a sufficient condition for stochastic gradients not to slow down the convergence of Nesterov's accelerated gradient method. The new condition has the strong-growth condition by Schmidt \& Roux as a special case, and it also allows us to (i) model problems with constraints and (ii) design new types of oracles (e.g., oracles for finite-sum problems such as SAGA). Our results are obtained by revisiting Nesterov's accelerated algorithm and are useful for designing stochastic oracles without changing the underlying first-order method.
- Published
- 2022
32. Sampling of the Wiener Process for Remote Estimation over a Channel with Unknown Delay Statistics
- Author
-
Tang, Haoyue, Sun, Yin, and Tassiulas, Leandros
- Subjects
Computer Science - Information Theory - Abstract
In this paper, we study an online sampling problem of the Wiener process. The goal is to minimize the mean squared error (MSE) of the remote estimator under a sampling frequency constraint when the transmission delay distribution is unknown. The sampling problem is reformulated into an optional stopping problem, and we propose an online sampling algorithm that can adaptively learn the optimal stopping threshold through stochastic approximation. We prove that the cumulative MSE regret grows with rate $\mathcal{O}(\ln k)$, where $k$ is the number of samples. Through Le Cam's two point method, we show that the worst-case cumulative MSE regret of any online sampling algorithm is lower bounded by $\Omega(\ln k)$. Hence, the proposed online sampling algorithm is minimax order-optimal. Finally, we validate the performance of the proposed algorithm via numerical simulations., Comment: Conference Version: Mobihoc 2022, submitted to IEEE/ACM Transactions on Networking
- Published
- 2022
33. Adaptive Graph Spatial-Temporal Transformer Network for Traffic Flow Forecasting
- Author
-
Feng, Aosong and Tassiulas, Leandros
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence - Abstract
Traffic flow forecasting on graphs has real-world applications in many fields, such as transportation system and computer networks. Traffic forecasting can be highly challenging due to complex spatial-temporal correlations and non-linear traffic patterns. Existing works mostly model such spatial-temporal dependencies by considering spatial correlations and temporal correlations separately and fail to model the direct spatial-temporal correlations. Inspired by the recent success of transformers in the graph domain, in this paper, we propose to directly model the cross-spatial-temporal correlations on the spatial-temporal graph using local multi-head self-attentions. To reduce the time complexity, we set the attention receptive field to the spatially neighboring nodes, and we also introduce an adaptive graph to capture the hidden spatial-temporal dependencies. Based on these attention mechanisms, we propose a novel Adaptive Graph Spatial-Temporal Transformer Network (ASTTN), which stacks multiple spatial-temporal attention layers to apply self-attention on the input graph, followed by linear layers for predictions. Experimental results on public traffic network datasets, METR-LA PEMS-BAY, PeMSD4, and PeMSD7, demonstrate the superior performance of our model.
- Published
- 2022
34. Optimal Entanglement Distribution using Satellite Based Quantum Networks
- Author
-
Panigrahy, Nitish K., Dhara, Prajit, Towsley, Don, Guha, Saikat, and Tassiulas, Leandros
- Subjects
Quantum Physics ,Computer Science - Networking and Internet Architecture - Abstract
Recent technological advancements in satellite based quantum communication has made it a promising technology for realizing global scale quantum networks. Due to better loss distance scaling compared to ground based fiber communication, satellite quantum communication can distribute high quality quantum entanglements among ground stations that are geographically separated at very long distances. This work focuses on optimal distribution of bipartite entanglements to a set of pair of ground stations using a constellation of orbiting satellites. In particular, we characterize the optimal satellite-to-ground station transmission scheduling policy with respect to the aggregate entanglement distribution rate subject to various resource constraints at the satellites and ground stations. We cast the optimal transmission scheduling problem as an integer linear programming problem and solve it efficiently for some specific scenarios. Our framework can also be used as a benchmark tool to measure the performance of other potential transmission scheduling policies.
- Published
- 2022
35. Debt-Financed Collateral and Stability Risks in the DeFi Ecosystem
- Author
-
Darlin, Michael, Palaiokrassas, Georgios, and Tassiulas, Leandros
- Subjects
Quantitative Finance - Trading and Market Microstructure - Abstract
The rise of Decentralized Finance ("DeFi") on the Ethereum blockchain has enabled the creation of lending platforms, which serve as marketplaces to lend and borrow digital currencies. We first categorize the activity of lending platforms within a standard regulatory framework. We then employ a novel grouping and classification algorithm to calculate the percentage of fund flows into DeFi lending platforms that can be attributed to debt created elsewhere in the system ("debt-financed collateral"). Based on our results, we conclude that the wide-spread use of stablecoins as debt-financed collateral increases financial stability risks in the DeFi ecosystem.
- Published
- 2022
36. Age Optimal Sampling Under Unknown Delay Statistics
- Author
-
Tang, Haoyue, Chen, Yuchao, Wang, Jintao, Yang, Pengkun, and Tassiulas, Leandros
- Subjects
Computer Science - Information Theory - Abstract
This paper revisits the problem of sampling and transmitting status updates through a channel with random delay under a sampling frequency constraint \cite{sun_17_tit}. We use the Age of Information (AoI) to characterize the status information freshness at the receiver. The goal is to design a sampling policy that can minimize the average AoI when the statistics of delay is unknown. We reformulate the problem as the optimization of a renewal-reward process, and propose an online sampling strategy based on the Robbins-Monro algorithm. We prove that the proposed algorithm satisfies the sampling frequency constraint. Moreover, when the transmission delay is bounded and its distribution is absolutely continuous, the average AoI obtained by the proposed algorithm converges to the minimum AoI when the number of samples $K$ goes to infinity with probability 1. We show that the optimality gap decays with rate $\mathcal{O}\left(\ln K/K\right)$, and the proposed algorithm is minimax rate optimal. Simulation results validate the performance of our proposed algorithm., Comment: Accepted and to appear, IEEE Transactions on Information Theory
- Published
- 2022
- Full Text
- View/download PDF
37. KerGNNs: Interpretable Graph Neural Networks with Graph Kernels
- Author
-
Feng, Aosong, You, Chenyu, Wang, Shiqiang, and Tassiulas, Leandros
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence - Abstract
Graph kernels are historically the most widely-used technique for graph classification tasks. However, these methods suffer from limited performance because of the hand-crafted combinatorial features of graphs. In recent years, graph neural networks (GNNs) have become the state-of-the-art method in downstream graph-related tasks due to their superior performance. Most GNNs are based on Message Passing Neural Network (MPNN) frameworks. However, recent studies show that MPNNs can not exceed the power of the Weisfeiler-Lehman (WL) algorithm in graph isomorphism test. To address the limitations of existing graph kernel and GNN methods, in this paper, we propose a novel GNN framework, termed \textit{Kernel Graph Neural Networks} (KerGNNs), which integrates graph kernels into the message passing process of GNNs. Inspired by convolution filters in convolutional neural networks (CNNs), KerGNNs adopt trainable hidden graphs as graph filters which are combined with subgraphs to update node embeddings using graph kernels. In addition, we show that MPNNs can be viewed as special cases of KerGNNs. We apply KerGNNs to multiple graph-related tasks and use cross-validation to make fair comparisons with benchmarks. We show that our method achieves competitive performance compared with existing state-of-the-art methods, demonstrating the potential to increase the representation ability of GNNs. We also show that the trained graph filters in KerGNNs can reveal the local graph structures of the dataset, which significantly improves the model interpretability compared with conventional GNN models.
- Published
- 2022
38. Tackling System and Statistical Heterogeneity for Federated Learning with Adaptive Client Sampling
- Author
-
Luo, Bing, Xiao, Wenli, Wang, Shiqiang, Huang, Jianwei, and Tassiulas, Leandros
- Subjects
Computer Science - Machine Learning ,Computer Science - Artificial Intelligence ,Computer Science - Distributed, Parallel, and Cluster Computing ,Computer Science - Networking and Internet Architecture ,Mathematics - Optimization and Control - Abstract
Federated learning (FL) algorithms usually sample a fraction of clients in each round (partial participation) when the number of participants is large and the server's communication bandwidth is limited. Recent works on the convergence analysis of FL have focused on unbiased client sampling, e.g., sampling uniformly at random, which suffers from slow wall-clock time for convergence due to high degrees of system heterogeneity and statistical heterogeneity. This paper aims to design an adaptive client sampling algorithm that tackles both system and statistical heterogeneity to minimize the wall-clock convergence time. We obtain a new tractable convergence bound for FL algorithms with arbitrary client sampling probabilities. Based on the bound, we analytically establish the relationship between the total learning time and sampling probabilities, which results in a non-convex optimization problem for training time minimization. We design an efficient algorithm for learning the unknown parameters in the convergence bound and develop a low-complexity algorithm to approximately solve the non-convex problem. Experimental results from both hardware prototype and simulation demonstrate that our proposed sampling scheme significantly reduces the convergence time compared to several baseline sampling schemes. Notably, our scheme in hardware prototype spends 73% less time than the uniform sampling baseline for reaching the same target loss., Comment: Accepted in IEEE INFOCOM 2022
- Published
- 2021
39. Cost-Effective Federated Learning in Mobile Edge Networks
- Author
-
Luo, Bing, Li, Xiang, Wang, Shiqiang, Huang, Jianwei, and Tassiulas, Leandros
- Subjects
Computer Science - Machine Learning ,Computer Science - Distributed, Parallel, and Cluster Computing ,Computer Science - Networking and Internet Architecture ,Electrical Engineering and Systems Science - Systems and Control ,Mathematics - Optimization and Control - Abstract
Federated learning (FL) is a distributed learning paradigm that enables a large number of mobile devices to collaboratively learn a model under the coordination of a central server without sharing their raw data. Despite its practical efficiency and effectiveness, the iterative on-device learning process (e.g., local computations and global communications with the server) incurs a considerable cost in terms of learning time and energy consumption, which depends crucially on the number of selected clients and the number of local iterations in each training round. In this paper, we analyze how to design adaptive FL in mobile edge networks that optimally chooses these essential control variables to minimize the total cost while ensuring convergence. We establish the analytical relationship between the total cost and the control variables with the convergence upper bound. To efficiently solve the cost minimization problem, we develop a low-cost sampling-based algorithm to learn the convergence related unknown parameters. We derive important solution properties that effectively identify the design principles for different optimization metrics. Practically, we evaluate our theoretical results both in a simulated environment and on a hardware prototype. Experimental evidence verifies our derived properties and demonstrates that our proposed solution achieves near-optimal performance for different optimization metrics for various datasets and heterogeneous system and statistical settings., Comment: Accepted in IEEE JSAC Special Issue on Distributed Learning over Wireless Edge Networks. arXiv admin note: substantial text overlap with arXiv:2012.08336
- Published
- 2021
40. Federated Learning with Spiking Neural Networks
- Author
-
Venkatesha, Yeshwanth, Kim, Youngeun, Tassiulas, Leandros, and Panda, Priyadarshini
- Subjects
Computer Science - Machine Learning ,Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Distributed, Parallel, and Cluster Computing ,Computer Science - Neural and Evolutionary Computing - Abstract
As neural networks get widespread adoption in resource-constrained embedded devices, there is a growing need for low-power neural systems. Spiking Neural Networks (SNNs)are emerging to be an energy-efficient alternative to the traditional Artificial Neural Networks (ANNs) which are known to be computationally intensive. From an application perspective, as federated learning involves multiple energy-constrained devices, there is a huge scope to leverage energy efficiency provided by SNNs. Despite its importance, there has been little attention on training SNNs on a large-scale distributed system like federated learning. In this paper, we bring SNNs to a more realistic federated learning scenario. Specifically, we propose a federated learning framework for decentralized and privacy-preserving training of SNNs. To validate the proposed federated learning framework, we experimentally evaluate the advantages of SNNs on various aspects of federated learning with CIFAR10 and CIFAR100 benchmarks. We observe that SNNs outperform ANNs in terms of overall accuracy by over 15% when the data is distributed across a large number of clients in the federation while providing up to5.3x energy efficiency. In addition to efficiency, we also analyze the sensitivity of the proposed federated SNN framework to data distribution among the clients, stragglers, and gradient noise and perform a comprehensive comparison with ANNs.
- Published
- 2021
- Full Text
- View/download PDF
41. Deep Reinforcement Learning-Based Rebalancing Policies for Profit Maximization of Relay Nodes in Payment Channel Networks
- Author
-
Papadis, Nikolaos, Tassiulas, Leandros, Barbosa-Povoa, Ana Paula, Editorial Board Member, de Almeida, Adiel Teixeira, Editorial Board Member, Gans, Noah, Editorial Board Member, Gupta, Jatinder N. D., Editorial Board Member, Heim, Gregory R., Editorial Board Member, Hua, Guowei, Editorial Board Member, Kimms, Alf, Editorial Board Member, Li, Xiang, Editorial Board Member, Masri, Hatem, Editorial Board Member, Nickel, Stefan, Editorial Board Member, Qiu, Robin, Editorial Board Member, Shankar, Ravi, Editorial Board Member, Slowiński, Roman, Editorial Board Member, Tang, Christopher S., Editorial Board Member, Wu, Yuzhe, Editorial Board Member, Zhu, Joe, Editorial Board Member, Zopounidis, Constantin, Editorial Board Member, Pardalos, Panos, editor, Kotsireas, Ilias, editor, Knottenbelt, William J., editor, and Leonardos, Stefanos, editor
- Published
- 2023
- Full Text
- View/download PDF
42. State-Dependent Processing in Payment Channel Networks for Throughput Optimization
- Author
-
Papadis, Nikolaos and Tassiulas, Leandros
- Subjects
Electrical Engineering and Systems Science - Systems and Control ,Computer Science - Networking and Internet Architecture ,Computer Science - Social and Information Networks - Abstract
Payment channel networks (PCNs) have emerged as a scalability solution for blockchains built on the concept of a payment channel: a setting that allows two nodes to safely transact between themselves in high frequencies based on pre-committed peer-to-peer balances. Transaction requests in these networks may be declined because of unavailability of funds due to temporary uneven distribution of the channel balances. In this paper, we investigate how to alleviate unnecessary payment blockage via proper prioritization of the transaction execution order. Specifically, we consider the scheduling problem in PCNs: as transactions continuously arrive on both sides of a channel, nodes need to decide which ones to process and when in order to maximize their objective, which in our case is the channel throughput. We introduce a stochastic model to capture the dynamics of a payment channel under random arrivals, and propose that channels can hold incoming transactions in buffers up to some deadline in order to enable more elaborate processing decisions. We describe a policy that maximizes the channel success rate/throughput for uniform transaction requests of fixed amounts, both in the presence and absence of buffering capabilities, and formally prove its optimality. We also develop a discrete event simulator of a payment channel, and evaluate different heuristic scheduling policies in the more general heterogeneous amounts case, with the results showing superiority of the heuristic extension of our policy in this case as well. Our work opens the way for more formal research on improving PCN performance via joint consideration of routing and scheduling decisions., Comment: 28 pages
- Published
- 2021
43. Cost-Effective Federated Learning Design
- Author
-
Luo, Bing, Li, Xiang, Wang, Shiqiang, Huang, Jianwei, and Tassiulas, Leandros
- Subjects
Computer Science - Machine Learning ,Computer Science - Distributed, Parallel, and Cluster Computing ,Computer Science - Networking and Internet Architecture ,Mathematics - Optimization and Control - Abstract
Federated learning (FL) is a distributed learning paradigm that enables a large number of devices to collaboratively learn a model without sharing their raw data. Despite its practical efficiency and effectiveness, the iterative on-device learning process incurs a considerable cost in terms of learning time and energy consumption, which depends crucially on the number of selected clients and the number of local iterations in each training round. In this paper, we analyze how to design adaptive FL that optimally chooses these essential control variables to minimize the total cost while ensuring convergence. Theoretically, we analytically establish the relationship between the total cost and the control variables with the convergence upper bound. To efficiently solve the cost minimization problem, we develop a low-cost sampling-based algorithm to learn the convergence related unknown parameters. We derive important solution properties that effectively identify the design principles for different metric preferences. Practically, we evaluate our theoretical results both in a simulated environment and on a hardware prototype. Experimental evidence verifies our derived properties and demonstrates that our proposed solution achieves near-optimal performance for various datasets, different machine learning models, and heterogeneous system settings., Comment: Accepted in IEEE INFOCOM 2021
- Published
- 2020
44. Birkhoff's Decomposition Revisited: Sparse Scheduling for High-Speed Circuit Switches
- Author
-
Valls, Víctor, Iosifidis, George, and Tassiulas, Leandros
- Subjects
Mathematics - Optimization and Control ,Computer Science - Networking and Internet Architecture - Abstract
Data centers are increasingly using high-speed circuit switches to cope with the growing demand and reduce operational costs. One of the fundamental tasks of circuit switches is to compute a sparse collection of switching configurations to support a traffic demand matrix. Such a problem has been addressed in the literature with variations of the approach proposed by Birkhoff in 1946 to decompose a doubly stochastic matrix exactly. However, the existing methods are heuristic and do not have theoretical guarantees on how well a collection of switching configurations (i.e., permutations) can approximate a traffic matrix (i.e., a scaled doubly stochastic matrix). In this paper, we revisit Birkhoff's approach and make three contributions. First, we establish the first theoretical bound on the sparsity of Birkhoff's algorithm (i.e., the number of switching configurations necessary to approximate a traffic matrix). In particular, we show that by using a subset of the admissible permutation matrices, Birkhoff's algorithm obtains an $\epsilon$-approximate decomposition with at most $O( \log(1 / \epsilon))$ permutations. Second, we propose a new algorithm, Birkhoff+, which combines the wealth of Frank-Wolfe with Birkhoff's approach to obtain sparse decompositions in a fast manner. And third, we evaluate the performance of the proposed algorithm numerically and study how this affects the performance of a circuit switch. Our results show that Birkhoff+ is superior to previous algorithms in terms of throughput, running time, and number of switching configurations.
- Published
- 2020
45. Optimal Bidding Strategy for Maker Auctions
- Author
-
Darlin, Michael, Papadis, Nikolaos, and Tassiulas, Leandros
- Subjects
Quantitative Finance - Trading and Market Microstructure ,Quantitative Finance - Portfolio Management - Abstract
The Maker Protocol is a decentralized finance application that enables collateralized lending. The application uses open-bid, second-price auctions to complete its loan liquidation process. In this paper, we develop a bidding function for these auctions, focusing on the costs incurred to participate in the auctions. We then optimize these costs using parameters from historical auction data, and compare our optimal bidding prices to the historical auction prices. We find that the majority of auctions end at higher prices than our recommended optimal prices, and we propose several theories for these results.
- Published
- 2020
46. A Blockchain-based Decentralized Data Sharing Infrastructure for Off-grid Networking
- Author
-
Niavis, Harris, Papadis, Nikolaos, and Tassiulas, Leandros
- Subjects
Computer Science - Distributed, Parallel, and Cluster Computing ,Computer Science - Social and Information Networks - Abstract
Off-grid networks are recently emerging as a solution to connect the unconnected or provide alternative services to networks of possibly untrusted participants. The systems currently used, however, exhibit limitations due to their centralized nature and thus prove inadequate to secure trust. Blockchain technology can be the tool that will enable trust and transparency in such networks. In this paper, we introduce a platform for secure and privacy-respecting decentralized data sharing among untrusted participants in off-grid networks. The proposed architecture realizes this goal via the integration of existing blockchain frameworks (Hyperledger Fabric, Indy, Aries) with an off-grid network device and a distributed file system. We evaluate the proposed platform through experiments and show results for its throughput and latency, which indicate its adequate performance for supporting off-grid decentralized applications., Comment: An abridged version of this work appeared in ICBC 2020, fixed minor typos and layout issues
- Published
- 2020
47. Model Pruning Enables Efficient Federated Learning on Edge Devices
- Author
-
Jiang, Yuang, Wang, Shiqiang, Valls, Victor, Ko, Bong Jun, Lee, Wei-Han, Leung, Kin K., and Tassiulas, Leandros
- Subjects
Computer Science - Machine Learning ,Computer Science - Distributed, Parallel, and Cluster Computing ,Statistics - Machine Learning - Abstract
Federated learning (FL) allows model training from local data collected by edge/mobile devices while preserving data privacy, which has wide applicability to image and vision applications. A challenge is that client devices in FL usually have much more limited computation and communication resources compared to servers in a datacenter. To overcome this challenge, we propose PruneFL -- a novel FL approach with adaptive and distributed parameter pruning, which adapts the model size during FL to reduce both communication and computation overhead and minimize the overall training time, while maintaining a similar accuracy as the original model. PruneFL includes initial pruning at a selected client and further pruning as part of the FL process. The model size is adapted during this process, which includes maximizing the approximate empirical risk reduction divided by the time of one FL round. Our experiments with various datasets on edge devices (e.g., Raspberry Pi) show that: (i) we significantly reduce the training time compared to conventional FL and various other pruning-based methods; (ii) the pruned model with automatically determined size converges to an accuracy that is very similar to the original model, and it is also a lottery ticket of the original model., Comment: Accepted for publication in IEEE Transactions on Neural Networks and Learning Systems (TNNLS)
- Published
- 2019
48. Online Convex Optimization with Perturbed Constraints
- Author
-
Valls, Víctor, Iosifidis, George, Leith, Douglas J., and Tassiulas, Leandros
- Subjects
Mathematics - Optimization and Control - Abstract
This paper addresses Online Convex Optimization (OCO) problems where the constraints have additive perturbations that (i) vary over time and (ii) are not known at the time to make a decision. Perturbations may not be i.i.d. generated and can be used to model a time-varying budget or commodity in resource allocation problems. The problem is to design a policy that obtains sublinear regret while ensuring that the constraints are satisfied on average. To solve this problem, we present a primal-dual proximal gradient algorithm that has $O(T^\epsilon \vee T^{1-\epsilon})$ regret and $O(T^\epsilon)$ constraint violation, where $\epsilon \in [0,1)$ is a parameter in the learning rate. Our results match the bounds of previous work on OCO with time-varying constraints when $\epsilon = 1/2$; however, we (i) define the regret using a time-varying set of best fixed decisions; (ii) can balance between regret and constraint violation; and (iii) use an adaptive learning rate that allows us to run the algorithm for any time horizon.
- Published
- 2019
49. Joint Service Placement and Request Routing in Multi-cell Mobile Edge Computing Networks
- Author
-
Poularakis, Konstantinos, Llorca, Jaime, Tulino, Antonia M., Taylor, Ian, and Tassiulas, Leandros
- Subjects
Computer Science - Networking and Internet Architecture - Abstract
The proliferation of innovative mobile services such as augmented reality, networked gaming, and autonomous driving has spurred a growing need for low-latency access to computing resources that cannot be met solely by existing centralized cloud systems. Mobile Edge Computing (MEC) is expected to be an effective solution to meet the demand for low-latency services by enabling the execution of computing tasks at the network-periphery, in proximity to end-users. While a number of recent studies have addressed the problem of determining the execution of service tasks and the routing of user requests to corresponding edge servers, the focus has primarily been on the efficient utilization of computing resources, neglecting the fact that non-trivial amounts of data need to be stored to enable service execution, and that many emerging services exhibit asymmetric bandwidth requirements. To fill this gap, we study the joint optimization of service placement and request routing in MEC-enabled multi-cell networks with multidimensional (storage-computation-communication) constraints. We show that this problem generalizes several problems in literature and propose an algorithm that achieves close-to-optimal performance using randomized rounding. Evaluation results demonstrate that our approach can effectively utilize the available resources to maximize the number of requests served by low-latency edge cloud servers., Comment: IEEE Infocom 2019
- Published
- 2019
50. Learning the Optimal Synchronization Rates in Distributed SDN Control Architectures
- Author
-
Poularakis, Konstantinos, Qin, Qiaofeng, Ma, Liang, Kompella, Sastry, Leung, Kin K., and Tassiulas, Leandros
- Subjects
Computer Science - Networking and Internet Architecture - Abstract
Since the early development of Software-Defined Network (SDN) technology, researchers have been concerned with the idea of physical distribution of the control plane to address scalability and reliability challenges of centralized designs. However, having multiple controllers managing the network while maintaining a "logically-centralized" network view brings additional challenges. One such challenge is how to coordinate the management decisions made by the controllers which is usually achieved by disseminating synchronization messages in a peer-to-peer manner. While there exist many architectures and protocols to ensure synchronized network views and drive coordination among controllers, there is no systematic methodology for deciding the optimal frequency (or rate) of message dissemination. In this paper, we fill this gap by introducing the SDN synchronization problem: how often to synchronize the network views for each controller pair. We consider two different objectives; first, the maximization of the number of controller pairs that are synchronized, and second, the maximization of the performance of applications of interest which may be affected by the synchronization rate. Using techniques from knapsack optimization and learning theory, we derive algorithms with provable performance guarantees for each objective. Evaluation results demonstrate significant benefits over baseline schemes that synchronize all controller pairs at equal rate., Comment: IEEE Infocom 2019
- Published
- 2019
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.