With the widespread adoption of blockchain applications, the imperative for seamless data migration among decentralized applications has intensified. This necessity arises from various factors, including the depletion of blockchain disk space, transitions between blockchain systems, and specific requirements such as temporal data analysis. To meet these challenges and ensure the sustained functionality of applications, it is imperative to conduct time-aware cross-blockchain data migration. This process is designed to facilitate the smooth iteration of decentralized applications and the construction of a temporal index for historical data, all while preserving the integrity of the original data. In various application scenarios, this migration task may encompass the transfer of data between multiple blockchains, involving movements from one chain to another, from one chain to several chains, or from multiple chains to a single chain. However, the success of data migration hinges on the careful consideration of factors such as the reliability of the data source, data consistency, and migration efficiency. This paper introduces a time-aware cross-blockchain data migration approach tailored to accommodate diverse application scenarios, including migration between multiple chains. The proposed solution integrates a collective mechanism for controlling, executing, and storing procedures to address the complexities of data migration, incorporating elements such as transaction classification and matching. Extensive experiments have been conducted to validate the efficacy of the proposed approach.
- Article type
- Year
- Co-author
Nonnegative Matrix Factorization (NMF) is one of the most popular feature learning technologies in the field of machine learning and pattern recognition. It has been widely used and studied in the multi-view clustering tasks because of its effectiveness. This study proposes a general semi-supervised multi-view nonnegative matrix factorization algorithm. This algorithm incorporates discriminative and geometric information on data to learn a better-fused representation, and adopts a feature normalizing strategy to align the different views. Two specific implementations of this algorithm are developed to validate the effectiveness of the proposed framework: Graph regularization based Discriminatively Constrained Multi-View Nonnegative Matrix Factorization (GDCMVNMF) and Extended Multi-View Constrained Nonnegative Matrix Factorization (ExMVCNMF). The intrinsic connection between these two specific implementations is discussed, and the optimization based on multiply update rules is presented. Experiments on six datasets show that the effectiveness of GDCMVNMF and ExMVCNMF outperforms several representative unsupervised and semi-supervised multi-view NMF approaches.
Recently, analyzing big data on the move is booming. It requires that the hardware resource should be low volume, low power, light in weight, high-performance, and highly scalable whereas the management software should be flexible and consume little hardware resource. To meet these requirements, we present a system named SOCA-DOM that encompasses a mobile system-on-chip array architecture and a two-tier “software-defined” resource manager named Chameleon. First, we design an Ethernet communication board to support an array of mobile system-on-chips. Second, we propose a two-tier software architecture for Chameleon to make it flexible. Third, we devise data, configuration, and control planes for Chameleon to make it “software-defined” and in turn consume hardware resources on demand. Fourth, we design an accurate synthetic metric that represents the computational power of a computing node. We employ 12 Apache Spark benchmarks to evaluate SOCA-DOM. Surprisingly, SOCA-DOM consumes up to 9:4x less CPU resources and 13.5x less memory than Mesos which is an existing resource manager. In addition, we show that a 16-node SOCA-DOM consumes up to 4x less energy than two standard Xeon servers. Based on the results, we conclude that an array architecture with fine-grained hardware resources and a software-defined resource manager works well for analyzing big data on the move.