International Journal of High Performance Computing and Networking (136 papers in press)
Revenue maximisation for scheduling deadline-constrained moldable jobs on high performance computing as a service platforms
by Kuo-Chan Huang, Chun-Hao Hung, Wei Hsieh
Abstract: Traditionally, High-Performance Computing (HPC) systems usually deal with the so-called best-effort jobs, which do not have deadlines and are scheduled in an as-quick-as-possible manner. Recently, the concept of HPC as a Service (HPCaaS) was proposed, aiming to transform HPC facilities and applications into a more convenient and accessible service model. To achieve that goal, there will be new issues to explore, such as scheduling jobs with deadlines and maximising the revenue of service providers. This paper presents a reservation-based dynamic scheduling approach for scheduling deadline-constrained moldable jobs with the aim of maximising a service providers revenue. The proposed approach has been evaluated with a series of simulation experiments. The experimental results indicate that our scheduling approach can achieve significantly higher revenue than previous methods. In the experiments, we also explored several research issues, including waiting queue sequencing, processor allocation decisions on time and space, admission control, and partial rescheduling.
Keywords: moldable jobs; scheduling; deadline-constrained jobs; high performance computing; revenue maximisation; processor allocation.
Secure and verifiable outsourcing protocol for non-negative matrix factorisation
by Zhenhua Liu, Bin Li, Qi Han
Abstract: With the rapid development of cloud computing services, resource-constrained clients can outsource their expensive computation tasks, such as scientific computations, to untrusted cloud servers. Furthermore, it is essential for these clients to protect their sensitive data and verify the validity of the returned computation results. In this paper, we focus on outsourcing protocol of non-negative matrix factorisation, which is an expensive computation task and has been widely applied to image processing, face recognition, text analysis, and so on. The permutation technique is employed to transform the original problem into a new one in our proposed protocol so as to protect the privacy, and the matrix 1-norm technique is used to verify the result returned from the cloud server in order to reduce the verification cost. Based on these two techniques, we construct a secure and verifiable outsourcing protocol for non-negative matrix factorisation. Moreover, the theoretical analysis and the experimental results show that our proposed protocol brings great computation savings for resource-constrained clients and fulfills the goals of correctness, security, verifiability and high efficiency.
Keywords: cloud computing; permutation; outsourcing; cloud security; verifiability; matrix factorisation.
Comparable encryption scheme supporting multiple users in cloud computing
by Jun Ye, Meixia Miao, Peng Chen, Xiaofeng Chen
Abstract: Cloud computing provides a way for secure data sharing, where data should be converted into ciphertext forms before uploading to the cloud servers. At the same time, we need to compare the data in many occasions. Although the privacy is enhanced when data is encrypted, the comparison of encrypted data is inconvenient, especially the data encrypted by different keys. The techniques for secure ciphertexts comparison is becoming a hot research topic both in the academic research and the industry practice. However, to the best of our knowledge, only a single user's data encrypted by a single key can be compared in the existing schemes.
In this paper, we designed the first comparable encryption scheme among multiple users, and a novel token transformation technique was provided. In our scheme, different ciphertexts encrypted by different keys can be compared with each other by using our new algorithm and token transformation method. In addition, the security and verifiability characteristics of our scheme guarantee that users can easily verify the comparison result, and the value of the data will not be revealed.
Finally, our scheme is flexible, which can be widely used in cloud computing.
Keywords: comparable encryption; multiple users; token transformation; security.
Estimating data stream tendencies to adapt clustering parameters
by Marcelo Keese Albertini, Rodrigo Mello
Abstract: A wide range of applications based on processing of data streams have emerged in the last decade. They require specialised techniques to obtain representative models and extract information. Traditional data clustering algorithms have been adapted to include continuously arriving data by updating the current model. Most data stream clustering algorithms aggregate new data into models according to parameters usually set by users. Problems arise when choosing the values of given parameters. When the phenomenon under study is stable, an analysis of a sample of the data stream or a priori knowledge can be used. However, when the behaviour changes over collection, parameters become obsolete and, consequently, the performance is degraded. In this paper we study the problem of automatically adapt control parameters of data stream clustering algorithms. In this sense, we introduce a novel approach to estimate and use data tendencies in order to automatically modify control parameters. We present a proof of the convergence of our approach towards an ideal and unknown value of the control parameter. Experimental results confirm the estimation of data tendency improves learning control parametrisation.
Keywords: big data; data clustering ; data stream ; data sequence ; adaptive clustering ; data analysis.
Fully secure hierarchical inner product encryption for privacy preserving keyword searching in cloud
by Leyou Zhang, Zhuanning Wang, Yi Mu, Yupu Hu
Abstract: Cloud computing provides dynamically scalable resources provisioned as a service over networks. But an untrustworthy Cloud Service Provider (CSP) offers a big obstacle for the adoption of the cloud service since a CSP can access data in the cloud without the data owner's permission. Hierarchical Inner Product Encryption (HIPE) covers all applications of anonymous encryption, fully private communication and search on encrypted data, and provides trusted data access control policy to CSP. However, the existing works only achieve either selective attribute-hiding or adaptive attribute-hiding under some strong assumptions in the public key setting. These schemes only support limited class of functions and cannot protect the privacy of the query. To overcome these shortcomings, a novel HIPE in private key setting is proposed. The scheme achieves both fully secure and security reduction under the Decisional Linear (DLIN) assumption in the standard model. It has longer O(l)-size secret keys and O(n)-size ciphertexts, and brings high computational complexity. Therefore, a variant of the basic scheme is presented with the same security but shorter ciphertexts.
Keywords: cloud security; searching encryption; HIPE; DLIN assumption.
Improved pre-copy algorithm using statistical prediction and compression model for efficient live memory migration
by Minal Patel, Sanjay Chaudhary, Sanjay Garg
Abstract: Cloud computing provides on-demand network access based on computing resources such as networks, servers, storage, etc. The core part of the cloud is the virtualisation that runs multiple instances of different operating systems simultaneously. Virtual machines can be moved between physical hosts with the help of the migration feature of virtualisation. The migration of a virtual machine is useful for load balancing, energy efficiency, fault tolerance, etc. Xen hypervisor is used to execute and migrate the guests on different architectures using a pre-copy algorithm. There are three major categories to improve pre-copy using live migration algorithms: i) reducing dirty pages; ii) predicating dirty pages; iii) compressing memory pages. The methods based on reducing dirty pages can lead to performance degradation so a new approach called combined approach (including prediction and compression) is proposed in this paper. The prediction of dirty pages during a migration is performed using the ARIMA (Auto-Regressive Integrated Moving Average) model to avoid repeatedly sending them. This method is used to predict a time series based low or high dirty pages as linear combinations of its past and error values. An LRU (Least Recently Used) stack distance based delta compression algorithm is proposed for the compression model to achieve efficient virtual machine migration. The results show that the ARIMA based model is able to predict 93% in the case of high dirty pages environment. The combined approach is able to reduce downtime by 19.16% and total migration time by 10.76% on average compared with Xen's pre-copy algorithm.
Keywords: virtualisation; pre-copy; ARIMA; LRU; delta compression
A revocable certificateless encryption scheme with high performance
by Sun Yinxia, Zhang Zhuoran, Shen Limin
Abstract: Featuring no key escrow and no certificate, certificateless public key cryptography (CLPKC) has received much attention. A necessary problem in CLPKC is how to revoke a compromised/misbehaving user, which still remains to be efficiently solved: all existing methods are not practical for use owing to either enormous computation (secret channels) or a costly mediator. In this paper, we present a new approach for revocation in CLPKC and give a revocable certificateless encryption (RCLE) scheme. In the new scheme, a user's private key contains three parts: an initial partial private key, a time key and a secret value. The time key is updated at every time period and is transmitted over a public channel. Our construction offers higher performance than previous solutions. Based on the bilinear Diffie-Hellman problem, our RCLE scheme is provably secure in the random oracle model.
Keywords: revocation; certificateless encryption; Bilinear Diffie-Hellman problem; random oracle model.
Towards distributed acceleration of image processing applications using reconfigurable active SSD clusters: a case study of seismic data analysis
by Mageda Sharafeddin, Hmayag Partamian, Mariette Awad, Mazen A. R. Saghir, Haitham Akkary, Hassan Artail, Hazem Hajj, Mohammed Baydoun
Abstract: We propose a high performance distributed system that consists of several middleware servers (MWS) each connected to a number of FPGAs with extended solid state storage that we call reconfigurable active solid state device (RASSD) nodes. A MWS manages a group of RASSD nodes and bridges the connection between a client and the RASSD nodes within a collaborative distributed environment. We provide in this system full data communication solution between middleware and RASSD nodes. Because of the high levels of parallelism possible with the configurable fabric and the extended storage available by adding SSD storage to FPGAs, RASSD nodes in our system cooperate to deliver unprecedented speedup in image data analysis. In this work we use seismic data analysis as a case study to quantify how and by how much our RASSD nodes can accelerate computational throughput. We present speedup of seismic data prediction time in our system when we accelerate Gray Level Co-occurrence Matrix (GLCM) calculations and discuss the impact of data communication on speedup. When both GLCM and Haralick features are accelerated, the data traffic is reduced and our system achieves 102
Keywords: active solid state devices, distributed systems, FPGA, GLCM, Haralick attributes, intelligent systems, machine learning, parallel architectures, reconfigurable computing, seismic data analysis.
Distributed predictive performance anomaly detection for virtualised platforms
by Ali imran Jehangiri, Ramin Yahyapour, Edwin Yaqub, Philipp Wieder
Abstract: Predicting subsequent values of Quality of Service (QoS) properties is a key component of autonomic solutions. Predictions help in the management of cloud-based applications by preventing QoS breaches from happening. The huge amount of monitoring data generated by cloud platforms motivated the applicability of scalable data mining and machine learning techniques for predicting performance anomalies. Building prediction models individually for thousands of Virtual Machines (VMs) requires a robust generic methodology with minimal human intervention. In this work, we focus on these issues and present three main contributions. First, we compare several time series modelling approaches to evidence the predictive capabilities of these approaches. Second, we propose estimation-classification models that augment the predictive capabilities of machine learning classification methods (random forest, decision tree, support vector machine) by combining them with time series analysis methods (AR, ARIMA and ETS). Third, we show how the data mining techniques in conjunction with the Hadoop framework can be a useful, practical, and inexpensive method for predicting QoS attributes.
Keywords: cloud; performance prediction; monitoring; QoS; analytics; distributed time-series database;
A GPU parallel optimised blockwise NLM algorithm in a distributed computing system
by Salvatore Cuomo, Ardelio Galletti, Livia Marcellino
Abstract: Recently, advanced computing systems are widely adopted in order to intensively elaborate a huge amount of biomedical data in the e-health field. An interesting challenge is to perform real-time diagnosis by means of complex computational environments. In this paper, we suggest how to deal with the most computationally expensive processing steps of a distributed cloud e-health system by using Graphics Processing Units (GPUs). In the case study of magnetic resonance imaging (MRI), for improving the quality of denoising and helping the real time diagnosis, we have implemented a GPU parallel algorithm based on the Optimized Blockwise Non-Local Means (OB-NLM) method. Experimental results have shown a significant improvement of healthcare processing practice in terms of execution time.
Keywords: cloud systems; e-health; MRI denoising; non-local means; GPU computing.
Active vehicle: a new approach integrating AmI technology for trust management in VANET
by Amel Ltifi
Abstract: Ambient intelligence (AmI) is a new concept that interests many researchers in different domains. It requires highly innovative services such as Vehicular Ad-hoc networks (VANETs). In fact, concerning vehicle trust management, the majority of the proposed solutions are dependent on roadside equipment and an investigation into a scalable infrastructure for all roads is unreasonable. Therefore, road security should be managed by vehicles themselves. Hence, we have introduced the concept of 'active vehicle', which combines the power of ambient intelligence and the V2V technologies. In this paper, we propose a new secure communication model between active vehicles for alert spreading. Through this new model, traffic accident alerts are verified depending on the trust level of the sender. A trust management system has the responsibility to cut the spread of false warnings through the network.
Keywords: active vehicle; ambient communication; AmI; cooperation; cluster; security; trust management; VANET; V2V
Mining frequent itemsets over uncertain data streams
by Huiting Liu, Kaishen Zhou, Peng Zhao, Sheng Yao
Abstract: In recent years, owing to the wide applications of sensor network monitoring, RFID, moving object search and LBS, mining frequent itemsets over uncertain data streams has attracted much attention. However, existing hyper-structure-based algorithms cannot achieve high mining accuracy. In this paper, we present two sliding-window-based false-positive-oriented algorithms, called UFIM (Uncertain data stream Frequent Itemsets Mining) and UFIMTopK, to find threshold-based and rank-based frequent itemsets from uncertain data streams efficiently. UFIM uses a global GT-tree to maintain frequent itemsets in the sliding window and outputs them when needed. In addition, an efficient deleting strategy is designed to reduce time overhead. UFIMTopK is designed to find top-k frequent itemsets, and it is modified from UFIM. Experimental results show that our proposed algorithm UFIM can obtain higher mining accuracy than previous algorithms on synthetic and real-life datasets.
Keywords: frequent itemsets; uncertain data streams; sliding window.
A semantic similarity calculation model for event ontology mapping
by Xu Wang, Wei Liu, Yue Tan
Abstract: Ontology mapping is quite a useful solution for matching semantics between ontologies or schemas that were designed independently of each other. The traditional ontology is mainly to describe concepts and hierarchy relations between concepts, which easily cause Tennis Problem and impact the accuracy of concept mapping, especially while integrating data in event-centred domains. In this paper, we give the definition of event ontology mapping, and then propose a semantic similarity calculation model for event ontology mapping, which enables the mapping between event-based information with more abundant semantics, especially enabling the semantic mapping between event elements in different event classes. Experiments show that the proposed model can effectively identify the semantic relations of event classes in two event ontologies.
Keywords: ontology mapping; event ontology; event semantic similarity.
Lattice-based ring signature scheme under the random oracle model
by Shangping Wang, Ru Zhao, Yaling Zhang
Abstract: On the basis of the signatures scheme without trapdoors from lattice, which was proposed by Lyubashevsky in 2012, we present a new ring signature scheme from lattice. The proposed scheme is an extension of the signature scheme without trapdoors. We prove that our scheme is strongly unforgeable against adaptive chosen message in the random oracle model, and prove that the security of our scheme can be reduced to the hardness of the small integer solution (SIS) problem by rejection samplings. Compared with the existing lattice-based ring signature schemes, our new scheme is more efficient and with shorter signature length.
Keywords: lattice; ring signature; random oracle model; unforgeable;.
Robust Student's-t mixture modelling via Markov random field and its application in image segmentation
by Taisong Xiong, Yuanyuan Huang, Xin Luo
Abstract: The finite mixture model has been widely applied to image segmentation. However, the technique does not consider the spatial information in images, which leads to unsatisfactory results for image segmentation. To address this problem, in this paper, a Student's-t mixture model is proposed for image segmentation based on Markov random field. There are three advantages in the proposed model. Firstly, a representation of spatial relationships among pixels is given. Secondly, Student's-t distribution is chosen to be the component function of the proposed model instead of the Gaussian distribution because of its heavy tail. Thirdly, to infer the parameters of the proposed model, the gradient descent method is applied during the inference process. Comprehensive experiments are carried out on gray scale noisy images and real-world colour images. The experimental results have shown the effectiveness and robustness of the proposed model.
Keywords: Student's-t mixture model; Markov random field; image segmentation; gradient descent.
Towards workload-aware cloud resource provisioning using a multi-controller fuzzy switching approach
by Amjad Ullah, Jingpeng Li, Amir Hussain
Abstract: Elasticity enables cloud customers to enrich their applications to dynamically
adjust the underlying cloud resources as per their needs, in order to minimise the cost of infrastructure as well as to satisfy their performance goals. Over the past few years,a plethora of techniques have been introduced in order to implement elasticity. Control theory is one such technique that offers a systematic method to design feedback controllers to implement elasticity. A number of proposals based on feedback controller concepts have been introduced in the recent past in order to guarantee the QoS needs of a system deployed over a cloud. Many of these are based on the use of a single controller approach of various types, such as adaptive and fixed. However, for systems that operate in time-varying and unpredictable operating conditions, it becomes difficult for such approaches to perform effectively at all times, in order to comply with the system stated performance goals. The systems deployed over cloud are subject to unpredictable workload conditions that vary from time to time, e.g. an e-commerce website may face higher workloads than normal during festival or promotional schemes. This paper exploits the novel use of a recently developed multi-controller based approach, where each controller is specifically designed for one operating region. Moreover, the use of fuzzy logic is exploited to enable qualitative specification for the selection of the most suitable controller in run-time, based on system current behaviour. Initial experimental evaluation in comparison with the conventional single-controller approach demonstrates that our proposed method enhances the capability of an elastic application to comply with system performance goals.
Keywords: cloud elasticity; control theory; multi-controller; fuzzy logic; auto-scaling;rndynamic resource provisioning
Exploiting node mobility for fault management in RPL-based wireless sensor networks
by Djamila Bendouda, Lynda Mokdad, Hafid Haffaf
Abstract: This paper aims to propose an effective new local repair technique for RPL protocol (IPv6 Routing Protocol for Low Power and Lossy Networks) in Wireless Sensor Networks (WSNs), using the mobile node to replace failure node in three cases of failure position. In addition, RPL was originally designed for static networks, with no support for mobility. However, handling the local repair mechanism for the RPL with mobile nodes is a real challenge. Simultaneous with replace nodes failures by the mobility of their predecessor without rebuild of the RPL tree, is a new proposed solution to avoided triggering to rebuilding the graph Destination Oriented Directed Acyclic Graph (DODAG), estimate to ensure continuity of service, collect and transmit data in the monitored environment. Our work is the first effort implemented for fault management using the Local Repair technique by Mobile Node with RPL protocol (MN-LR_RPL). The MN-LR_RPL technique is presented by an algorithm which is simulated using COOJA simulator in the Contiki operating system. The performance of this algorithm is evaluated for different network settings, in terms of the control traffic, delay, energy consumption and PDR (%). Performance evaluation results show that MN-LR_RPL method allows a great improvement compared with the standard specification in the case of local repair, mainly in terms of packet loss ratio and average network latency. Based on our new method, we offer a few suggestions for MN-LR_RPL implemented in Vehicular Ad hoc Networks applications (VANETs) and another point to using Wireless Software Defined Network (WSDN) to improve RPL protocol in the internet of things.
Keywords: WSN; VANETs; fault management; local repair; mobile node; RPL protocol; Contiki; Cooja.
VKSE-DO: verifiable keyword search over encrypted data for dynamic data-owner
by Yinbin Miao, Jianfeng Ma, Fushan Wei, Kai Zhang, Zhiquan Liu, Xu An Wang
Abstract: With the popularity of cloud storage, the search over encrypted data technique enables data-owners to securely store and efficiently retrieve ciphertext with a privacy-preserving way. However, in practice, the cloud service provider (CSP) is a semi-honest-but-curious entity that may honestly execute a fraction of search operations and return partial or incorrect search results to save its computation resources. Therefore, the searchable encryption (SE) scheme should provide the result verification mechanism to guarantee the correctness of search results. To the best of our knowledge, most existing SE schemes rarely consider the dynamic data-owner scenario. Moreover, these schemes incur extra computational overhead when updating the whole ciphertext. To this end, we devise a secure and efficient cryptographic primitive called verifiable keyword search over encrypted data for the dynamic data-owner scheme. Formal security analysis proves that our scheme is secure against chosen keyword attack without random oracle. As a further contribution, experimental results over real-world dataset show its efficiency in practice.
Keywords: cloud storage; result verification; dynamic data-owner; chosen keyword attack.
Acceleration of sequential Monte Carlo for computationally intensive target distribution by parallel lookup table and GPU computing
by Di Zhao
Abstract: Sequential Monte Carlo (SMC) is the key solver for applications such as object tracking, signal processing and statistical distribution approximation. However, if the target distribution is complicated, the solution speed of the conventional SMC is too slow to satisfy the real-time requirement of applications. In this paper, by the novel idea of GPU based lookup table (GPU-LTU), the acceleration method for the conventional SMC (LTU-GPU accelerated SMC) is developed, and the efficiency of LTU-GPU accelerated SMC by a statistical approximation problem is illustrated. Computational results show that, LTU-GPU accelerated SMC is significantly faster than the conventional SMC from hours to seconds.
Keywords: sequential Monte Carlo, parallel lookup table, GPU computing, statistical distribution approximation
HtComp: bringing reconfigurable hardware to future high-performance applications
by Alessandro Cilardo
Abstract: Current trends in computer architecture are increasingly moving towards heterogeneous platforms, now including FPGAs as first-class components enabling unprecedented levels of performance and power-efficiency. Programming such next-generation machines is however extremely difficult as it requires architecture-specific code and low-level hardware design. This paper describes the main outcomes of the HtComp project, a two-year research programme aimed at exploring methodologies and tools allowing the automated generation of FPGA-based accelerators from high-level applications written in traditional software languages. In particular, the paper describes the main contributions brought by the project, covering the generation of hardware systems from high-level parallel code, the performance-oriented optimisation of memory architectures tailored on the application access patterns, as well as the automated definition of application-driven special-purpose on-chip interconnects. Overall, the above innovations contributed to creating a viable path allowing generic software developers to access tomorrow's hardware-accelerated high-performance platforms with minimum development overheads.
Keywords: HPC; FPGA; design automation; programming models
Enhanced e-learning system performance with a cloud and crowd oriented approach
by Chun- Hsiung Tseng, Ching-Lien Huang, Yung-Hui Chen, Chu-Chun Chuang, Han-Ci Syu, Yan-Ru Jiang, Fang-Chi Tsai, Pin-Yu Su, Jun-Yan Chen
Abstract: Flipped classroom is undoubtedly a very popular emerging teaching methodology and is highly emphasised in the education system today. There are several e-learning systems focusing on the flipped classroom concept. For example, with the "moodle" e-learning system, tutors prepare their materials in advance and request students to read these materials. The basic concept of flipped classroom is to have students learn by themselves before attending a real class at school. Once the background learning stage is performed outside class time, tutors have free time to lead students to participate in higher-order thinking. However, as shown in the report of Katie Ash (9), the performance of the flipped classroom method is in fact still arguable. Our survey shows that the contents offered by most modern e-learning systems are relatively static. Consider how fast new information appears on the web! Of course, teachers, or material providers, can upload new contents to e-learning systems. However, creating contents requires efforts. Today, teachers are ordered to do a lot of administrative tasks, such as recruiting students. Additionally, they have to devote their efforts to academic-industrial cooperative projects and research. The workload of our teachers is already heavy, so expecting teachers to update contents very frequently is not practical. Hence, the researchers believe that one of the major challenges faced by e-learning systems today is the richness of contents. Besides, crowd intelligence is usually considered helpful in enhancing learning performance. In the research, a system to use both cloud materials and crowd intelligence is proposed, and an experiment to validate the system is included.
Keywords: e-learning; crowd intelligence
Broker-based resource management in dynamic multi-cloud environment
by Naidila Sadashiv, S. M. Dilip Kumar
Abstract: Cloud computing is being largely embraced by small, medium and large business organisations to host interactive web-based applications, as they provide more unlimited services than the classical computing approach. However, providing uninterrupted service at an economic price with efficient use of resources is the challenge faced by cloud service providers, especially in serving users spread across the globe. Services from many different clouds can be reaped to address resource availability issues and impart the desired QoS. This paper presents a resource management approach for deploying three-tier applications over a broker-based multi-cloud environment. Strategies for quick cloud site selection, dynamic resource adaptation, and two-level load balancing with high availability are considered as part of this approach. Experiments are carried out on an extended Cloudsim simulator using realistic session workloads that are synthesised based on different statistical distributions. Performance evaluation of the approach reveals that these strategies lead to improved resource use, throughput and compliance with SLA, even under varying workload scenarios.
Keywords: cloud computing; resource management; multi-cloud environment; cloud site selection; dynamic adaptation; fair share load balancer; broker.
An efficient traceable data-sharing scheme in cloud computing for mobile devices
by Zhiying Wang, Jun Ye, Jianfeng Wang
Abstract: Owing to the convenience of mobile devices and the advantages of cloud computing, it is a fashion to efficiently share data by mobile devices in cloud computing. However, it might be insecure to upload the plaintext directly to the remote cloud server. In order to solve the problem, the data should be encrypted by a secure encryption scheme. However, most existing data-sharing schemes in cloud computing are either insecure against collusion attack or suffer from a large computational overhead. In this paper, an efficient data-sharing scheme in cloud computing for mobile devices is proposed based on convergent encryption and polynomial reconstruction techniques. Moreover, pre-decryption is used to further reduce the computation overhead of mobile devices. Finally, we use traitor tracing technology to track the users who have leaked private keys for their own benefit. Experimental results indicate that the proposed scheme is extremely suitable for mobile devices.
Keywords: data sharing; mobile cloud computing; convergent encryption; polynomial reconstruction; traitor tracing.
Malicious URL detection with feature extraction based on machine learning
by Baojiang Cui, Shanshan He, Xi Yao, Peilin Shi
Abstract: Many web applications suffer from various web attacks due to the lack of awareness concerning security. To create a web attack, executable code is usually embedded into a URL. Therefore, it is necessary to improve the reliability and security of web applications by accurately detecting malicious URLs. In this paper, statistical analyses based on gradient learning and feature extraction using sigmoidal threshold level are combined to propose a new detection approach based on machine learning. Twenty-one features from five different dimensions are extracted by analysing the characteristics to distinguish benign from malicious behaviour. Moreover, the na
Keywords: malicious URLs; feature selection; machine learning; multiple algorithms.
An efficient symmetric searchable encryption scheme for dynamic dataset in cloud computing paradigms
by Minghao Zhao, Han Jiang, Zhen Li, Qiuliang Xu, Hao Wang, Shaojing Li
Abstract: Searchable encryption is a significant cryptographic primitive to ensure storage security and data privacy in cloud computing environment. It allows a client to store a collection of encrypted documents on a server, and latterly, according to his specific search criteria, perform keyword based searches and retrieve the documents, meanwhile ensuring that it reveals minimal information to the server. Early research on searchable encryption mainly focused on the efficiency, the security and the query expressiveness, and nowadays, the studies have been paying attention to searchable encryption that supports dataset updates dynamically. In this paper, we propose a new dynamic symmetric searchable encryption scheme. In the aspect of efficiency, the complexity of search algorithm is O(1), while file addition and deletion are O(m'n) and O(N), respectively (here m' means the number of keywords in a document; N means the number of document-keyword pairs and n means the size of keywords dictionary), which indicates the overall efficiency is superior to the existing scheme; in terms of security, this scheme can resist selective keyword attack, and compared with former ones, it achieves less information leakage.
Keywords: cloud computing security; dynamic; searchable encryption； security and privacy-preserving.
Modelling and coordinating multisource distributed power system models in service oriented architecture
by Yan Liu, Jared M. Chase, Ian Gorton, Mark Rice, Adam Wynne
Abstract: Distributed power grid applications aim to combine multisource sensor device data with conventional power grid monitoring and control data to improve data redundancy and model accuracy. To support such an application, a software platform is essential that coordinates multiple data sources and integrates with the data communication and computation infrastructure. In this paper, we present a service oriented coordination platform that exposes parallel programs as services and data sources as ports. We model services and ports by extending classes of standard Common Information Model (CIM) for energy management and distribution management. These services are then coordinated based on the typed ports in the context of domain specific power applications. These services are integrated with two platform components: (1) a middleware for data communication and aggregation; and (2) a job launching mechanism to run parallel programs on HPC clusters. Our contribution is building a service oriented platform that is compliant with CIM management standards and provides power engineers an integrated environment for composing distributed power applications, disseminating data and launching HPC jobs. We demonstrate this platform with a use case of deploying distributed state estimation on the IEEE 118 bus model. Our experience leads to insights on future research direction on building a general service oriented for exploring distributed power applications.
Keywords: service oriented architecture; workflow; power grid.
Energy efficiency of a heterogeneous multicore system based on the enhanced Amdahl's law
by Songwen Pei, Junge Zhang, Naixue Xiong, Myoung-Seo Kim, Jean-luc Gaudiot
Abstract: Energy efficiency is one of the main challenges of designing future heterogeneous multicore systems, beyond performance. In this paper, we propose an energy efficiency analytical model for heterogeneous multicore system based on the enhanced Amdahl's law. The model extends the traditional computing-centric model by considering the overhead of data preparation, which potentially includes the overhead of communication, data transfer, synchronisation, etc. The analysis clearly shows that decreasing the overhead of data preparation is a promising approach to reach higher performance of computation and greater energy-efficiency. Therefore, more informed tradeoffs should be taken when we design a modern heterogeneous processor system within limited budget of energy.
Keywords: energy efficiency; overhead of data preparation; dataflow computing model; performance evaluation; heterogeneous multicore system.
Pixel classified colorisation method based on neighborhood similarity priori
by Jie Chen, Zongliang Gan, Xiuchang Zhu, Jin Wang
Abstract: Colorisation is a kind of computer-aided technology that automatically adds colours to grayscale images. This paper presents a scribble-based colorisation method that treats the flat and edge pixels differently. First, we classify the pixels to flat or edge pixel categories using the neighborhood similarity pixels searching algorithm. Then, we compute the weighted coefficients of the edge pixels by solving a constraint quadratic programming problem and compute the weighted coefficients of the flat pixels based on their luminance distances. Finally, we transmit the weighted coefficients to the chrominance images according to the joint correlation property between luminance and chrominance channels, and combine with the colours scribbled on by the user to compute all the unknown colours. The experimental results show that our method is effective, especially on reducing colour bleeding in the boundary parts, and can give better results when only a few colours are scribbled on.
Keywords: colorisation; joint correlation; linear weighted combination; quadratic programming; active set; neighborhood similarity pixels
A novel distributed node searching method in delay-tolerant networks
by Yiqin Deng, Ming Zhao, Zhengjiang Wu, Longhua Xiao
Abstract: Among conventional target tracking methods in delay-tolerant networks (DTNs), there exists a type of distributed approach which is efficient with regard to target tracking. These methods leave the position and time information of each node in their original sub-areas, and select particular nodes in each sub-region as anchor nodes and the messenger nodes are used to collect and update the spread information, so that the detector is able to track the target node. However, some nodes in these methods consume energy too quickly, resulting in energy holes and a short network lifetime. Based on the above, this research developed a novel target track strategy (MAtracking), on the basis of a mobile anchor (MA) and any correlation between nodes. It improves the effect of the target tracking through two new mechanisms: first, a new data collection method (MA) is adopted, which moves to collect real-time mobile location and time of nodes so as to balance the energy consumption of the whole network. Then, the approximate path is proposed by inferring the path of the MA into a non-deterministic polynomial (NP) problem for path optimisation. Secondly, the correlation between nodes is used to track the target in the absence of valid mobile information about it. By doing so, the node that has the shortest time to meet the target node is tracked, thus guaranteeing the success rate. A large number of experimental results from real data simulations show that the proposed MAtracking method can improve the network lifetime and tracking efficiency compared with existing approaches.
Keywords: delay-tolerant networks; target tracking; mobile anchor; correlation of nodes.
ElasticQ: an active queue management algorithm with flow trust
by Su Chenglong, Jin Guang, Jiang Xianliang, Niu Jun
Abstract: Active Queue Management (AQM) can improve the network transmission performance and reduce the delay of packets. However, most of previous algorithms cannot achieve the efficiency and fairness simultaneously when high-bandwidth flows exist. In this paper, a novel scheme, named the ElasticQ (Elastically Fair AQM), is proposed to suppress the high-bandwidth non-responsive flows and enhance the fairness of different flows. Different from previous works, the concept of flow trust is introduced into the design of AQM algorithms to measure the flows reliability and security effectively. The trust degrees of different flows are estimated to decide whether the packets are discarded or not in the proposed scheme with the sample-match mechanism and the packet dropping interval. Simulation experiments results show that ElasticQ could ensure the fairness of various flows, maintain the stability of the queues, and decrease the completion time of responsive flows, especially when responsive flows and non-responsive flows coexist.
Keywords: flow trust, active queue management, elastic fairness, non-responsive flow, responsive flow
A hybrid mutation artificial bee colony algorithm for spectrum sharing
by Ling Huang
Abstract: In order to alleviate the spectrum scarcity problem in wireless networks, achieve efficient allocation of spectrum resources, and balance the users permission to spectrum, a hybrid mutation artificial bee colony algorithm based on the artificial bee colony algorithm is presented. The presented algorithm aims to enhance the efficiency of global searching by improving the way of leaders searching the nectar source with differential evolution algorithm. In addition, the onlookers searching method is improved by bat algorithm to guarantee the convergence efficiency of the algorithm and the precision of the result, and it is assumed that the onlookers are equipped with bats echolocation to get close to the nectar source by adjusting the rate of pulse emission and loudness when they are searching nectar source. The simulation results show that the proposed algorithm has faster convergence, higher efficiency and more optimal solutions compared with other algorithms.
Keywords: spectrum sharing, artificial bee colony algorithm, bat algorithm, differential evolution
Outsourcing privacy-preserving ID3 decision tree over horizontal partitioned data for multiple parties
by Ye Li, Zoe L. Jiang, Xuan Wang, S.M. Yiu
Abstract: Today, many small and medium-sized companies want to share data for data mining; however, privacy and security concerns restrict such data sharing. Privacy-preserving data mining has emerged as a solution to this problem. Nevertheless, the traditional cryptographic solutions are too inefficient and infeasible to allow the large-scale analytics needed for big data. In this paper, we focus on the outsourcing of privacy-preserving ID3 decision trees over horizontally partitioned data for multiple parties. We outsource most of the protocol computation to the cloud and propose the OPPWAP to protect users data privacy. By this method, each party can have the correct results calculated with data from other parties and the cloud, and each partys data are kept private from other parties and the cloud. Our findings indicate that an increase in the number of participating parties results in a slight computing cost increase on the users side.
Keywords: cloud computing; privacy preserving data mining; decision tree; horizontally partitioned data; multiple parties; PPWAP.
When is the immune inspired B-Cell algorithm superior to the (1+1) evolutionary algorithm?
by Xiaoyun Xia, Langping Tang, Xue Peng
Abstract: There exist many experimental investigations of artificial immune systems (AIS), and it has been shown that AIS are useful and efficient for many real-world optimisation problems. However, we know little about that whether the AIS can outperform the traditional evolutionary algorithms on some optimisation problems in theory. This work rigorously proved that a simple AIS called the B-Cell Algorithm (BCA) with somatic contiguous hypermutations can efficiently optimise two instances of the multiprocessor scheduling problem in expected polynomial runtime, whereas the local search algorithms and the (1+1) evolutionary algorithm ((1+1) EA) using only one individual in the search space and with standard bit mutation are highly inefficient. This work is helpful for gaining insight into whether there exists any algorithm that is efficient for all specific problems.
Keywords: artificial immune system; somatic contiguous hypermutations; multiprocessor scheduling problem; runtime analysis.
Scalable bootstrap attribute reduction for massive data
by Suqin Ji, Hongbo Shi, Yali Lv, Min Guo
Abstract: Attribute reduction is one of the fundamental technique for knowledge acquisition in rough set theory. Traditional attribute reduction algorithms have to load the whole dataset into the memory at a time, however it is unfeasible for attribute reduction of the massive decision table due to hard limitation. To solve this problem, we propose the Bag of Little Bootstraps Attribute Reduction algorithm (BLBAR), which combines the bag of little bootstraps with attribute discernibility. Specifically, the algorithm first samples from the original decision table to generate a number of sub decision tables, and then finds the reducts of bootstrap samples of each sub-table through attribute discernibility; finally, all of the reducts are integrated as the reduct of the original massive decision table. Experimental results demonstrate that BLBAR leads to the improved feasibility, scalability and efficiency for attribute reduction on massive decision table.
Keywords: bag of little bootstraps; attribute reduction; massive data; discernibility of attribute; reduct
Assessing nodes importance in complex networks using structural holes
by Hui Xu, Jianpei Zhang, Jing Yang, Lijun Lun
Abstract: Accurate measurement of important nodes in complex networks has great practical and theoretical significance. Mining important nodes should not only consider the core nodes, but also take into account the locations of the nodes in the network. Despite much research on assessing important nodes, the importance of nodes in the structural holes is still easily ignored. Therefore, a local measuring method is proposed, which evaluates the nodes' importance by the total constraints caused by the lack of primary structural holes and secondary structural holes around the nodes. This method simultaneously considers both the centrality and the bridging property of the nodes' first-order and second-order neighbours. Further to prove the accuracy of the method TCM, we carry out deliberate attack simulations through selective deletion in a certain proportion of network nodes. Then we calculate the decreased ratio of network efficiency before and after the attacks. Experiment results show that the average effect of the TCM in four real networks is improved by 50.64% and 14.92% compared to the clustering coefficient index and the k-shell decomposition method, respectively. Obviously, the TCM is more accurate to mine important nodes than the other two methods, and it is suitable for quantitative analysis in large-scale networks.
Keywords: complex networks; structural holes; secondary holes; nodes importance; constraints
Load-balanced overlay recovery scheme with delay minimisation
by Shengwen Tian, Hongyong Yang
Abstract: Recovery from a link or node failure in the internet is often subjected to seconds or minutes of routing convergence, during which certain end-to-end connections may experience seconds or minutes of outage. According to this problem, existing approaches reroute the data traffic to a pre-defined backup path to detour the failed components. However, the maintenance of backup path increases the significant bandwidth expenditure. On the other hand, the diverted traffic may cause congestion on the backup path if it is not carefully split over multiple paths according to their available capacity. In this paper, we propose an efficient recovery scheme by using one-hop overlay multipath source routing, which is a post-failure recovery method. Once a failure happens, multiple one-hop overlay paths are constructed by selecting strategically multiple relay nodes, and the affected traffic is diverted to these paths in a well-balanced manner. We formulate the traffic allocation problem as a tractable linear programming (LP) optimisation problem, whose goal is to minimise the worst-case network congestion ratio. Simulations based on a real ISP network and a synthetic internet topology show that our scheme can effectively balance link usage dramatically and improve the reliability of the network.
Keywords: failure recovery; load balance; overlay routing; linear programming.
Negation scope detection with recurrent neural network models in review texts
by Lydia Lazib, Yanyan Zhao, Bing Qin, Ting Liu
Abstract: Identifying negation scopes in a text is an important subtask of information
extraction, which can benefit other natural language processing tasks, such as relation extraction, question answering and sentiment analysis. It also serves the task of social media text understanding. The task of negation scope detection can be regarded as a token-level sequence labelling problem. In this paper, we propose different models based on recurrent neural networks (RNNs) and word embedding that can be successfully applied to such tasks without any task-specific feature engineering effort. Our experimental results show that RNNs, without using any hand-crafted features, outperform feature-rich CRF-based model.
Keywords: negation scope detection; natural language processing; recurrent neural networks.
The optimisation of speech recognition based on convolutional neural network
by Weipeng Jing, Tao Jiang, Xingge Zhang, Liangkuan Zhu
Abstract: The convolutional neural network (CNN) as acoustic model is introduced into speech recognition system based on mobile computing. To improve speech accuracy, two optimised methods are proposed in speech recognition based on CNN. Firstly, aiming at the problem for existing pooling algorithms ignoring locally relevant characteristics of the speech data, a dynamic adaptive pooling (DA-pooling) algorithm is proposed in the pooling layer of the CNN model. DA-pooling algorithm calculates the Spearman correlation coefficient of the extracted data to determine data correlation, then selects an appropriate pool strategy for different correlativity of data according to weight. Secondly, in order to solve traditional dropout hiding neurons node randomly, a dropout strategy based on sparseness is proposed in the full-connected layer in the CNN model. By adding a unit sparseness determination mechanism in the output stage of the network unit, we can reduce the ratio of the influence of smaller units in the model results, thereby improving the generalisation ability of the model. Experimental results show that these strategies can improve the performance of the acoustic models based on CNN.
Keywords: CNN; speech recognition; DA-pooling; overfitting; sparseness
Hybrid feature selection technique for intrusion detection system
by MUHAMMAD HILMI KAMARUDIN, CARSTEN MAPLE, T.I.M. WATSON
Abstract: The enormous volume of network traffic has created a major challenge to the Intrusion Detection System (IDS). High dimensionalitys problems have rendered feature selection one of the most important criteria in determining the efficiency of the IDS. The objective of this study is to seek the best-fit for selecting features that can provide a high level of detection level with low false positive rate. In this study we have selected a hybrid feature selection model that potentially combines the strengths of both the filter and the wrapper selection procedure. The potential hybrid solution is expected to effectively select the optimal set of features in detecting intrusion. The use of several performance metrics to study the feature selection optimisation impact includes performance accuracy, true positive, true negative, detection rate, false positive, false negative and learning time. The proposed hybrid model was carried out using Correlation Feature Selection (CFS) together with three different search techniques known as best-first, greedy stepwise and genetic algorithm. The wrapper-based subset evaluation uses a Random Forest (RF) classifier to evaluate each of the features that were first selected by the filter method. The reduced feature selection on both KDD99 dataset (Network-based) and DARPA 1999 (Host-based) dataset was tested using RF algorithm classification with 10-fold cross-validation in a supervised environment. The experimental result shows that the hybrid feature selections had produced a satisfactory outcome in terms of least number of features with lower processing time. It also had achieved almost 100% detection rate of known attacks on both datasets. This encouraging result has proven the usage of hybrid feature selection model in IDS environment would be a method of choice.
Keywords: machine learning; filter-subset evaluation; wrapper-subset evaluation; genetic algorithm; random forest.
A dynamic and QoS-effective resource management system
by Amril Nazir
Abstract: This paper presents the design and implementation of HASEX that supports a
dynamic resource provisioning during application run-time and provides effective quality-of-service (QoS) resource management system for QoS- and deadline-driven applications. The most important feature of HASEX is the ability to serve high performance and distributed applications with minimal infrastructure and operating costs. The resource provisioning management is controlled by a 'rental' mechanism which is supported by a pool of computing resources that the system may rent from external resource owners/providers in times of high demand. HASEX differentiates the roles of application management, job scheduling, and resource provisioning tasks. This approach significantly reduces the overhead of managing distributed resources. We demonstrate the effectiveness of HASEX to rent groups of resource nodes across geographically disparate sites. We then perform performance comparison of HASEX with OpenNebula Cloud system to demonstrate its performance, scalability, and QoS effectiveness.
Keywords: deadline-driven jobs; SLA management; SLA/QoS middleware; resource provisioning; autonomic and self management.
Performance identification in large-scale class data from advanced facets of computational intelligence and soft computing techniques
by You-Shyang Chen
Abstract: Enormous hemodialysis (HD) treatments have caused concern regarding negative information as the world's highest prevalence of end-stage renal disease in Taiwan. This topic is the motivation to identify an adequate HD remedy. Although previous researchers have devised various models to address HD adequacy, the following five deficiencies form obstacles: (1) lack of consideration for imbalanced class problems (ICPs) with medical data; (2) lack of methods to season mathematical distributions for the given data; (3) lack of explanatory ability on the given data; (4) lack of effective methods to identify the determinants of HD adequacy; and (5) lack of appropriate classifiers to define HD adequacy. This study proposes hybrid models to integrate expert knowledge, imbalanced resampling methods, decision tree and random forests-based feature-selection methods, LEM2 algorithm, rough set theory, and rule-filtering techniques to process medical practice with ICPs. These models have better performance than the listed methods from the empirical results.
Keywords: rough set theory; decision tree; random forests; imbalanced class problem data.
Assessing complex evolving cyber-physical systems: a case study on smart medical devices
by Jan Sliwa
Abstract: Our environment is more and more permeated by intelligent devices and systems that directly interact with physical objects, including our bodies. In this way, complex cyber-physical systems are created where the cyber part is intertwined with the physical part so that novel dynamical dependencies appear. These intelligent devices produce immense amounts of data that can be stored and analysed. Often, high hopes are raised that processing those data will easily increase our knowledge and permit to take good decisions based on hard facts. If not based on a solid understanding, such data processing can lead to the well-known problem of GIGO (garbage in - garbage out). If presented in a visually compelling way, useless results will look like truth and will be misleading and damaging. In order obtain valid analysis results than can be used as "actionable knowledge", it is necessary to understand the working of the physical systems and also to be aware of possible statistical fallacies, such as biased selection. Even if big data collected by intelligent devices are not perfect, we want nevertheless use them to evaluate cyber-physical systems, their safety, efficiency and quality. One of the major challenges is the changing nature of the technical systems, of the environment in which they operate, and of the humans who use them. This raises the problem of partial invalidation of collected statistics if the conditions change. We will discuss the general problems related to assessing cyber-physical systems and present an important and interesting case study: smart medical devices. We have to stress that the statistical questions raised here are open, and one of the goals of this paper is to raise interest and instigate a cooperation to solve them.
Keywords: cyber-physical systems; quality assessment; smart medical devices; statistical models.
Distributed data-dependent locality sensitive hashing
by Yanping Ma
Abstract: Locality sensitive hashing (LSH) is a popular algorithm for approximate nearest neighbour (ANN) search. As LSH partitions vector space uniformly and the distribution of vectors is usually non-uniform, it poorly fits real datasets and has limited efficiency. In this paper, we propose a novel data-dependent LSH (DP-LSH) algorithm, which has a two-level structure. In the first level, we train a number of cluster centres, and use the centres to divide the dataset. So the vectors in each cluster have near uniform distribution. In the second level, we construct LSH tables for each cluster. Given a query, we first determine a few clusters that it belongs to, and perform an ANN search in the corresponding LSH tables. Furthermore, we present an optimised distributed scheme and a distributed DP-LSH algorithm. Experimental results on the reference datasets show that the search speed of DP-LSH can be increased by 48 times compared to E2LSH, while keeping high search precision; also, the distributed DP-LSH can further improve search efficiency.
Keywords: locality sensitive hashing; approximate nearest neighbour; data-dependent; distributed high dimensional search.
Real-time human action recognition using depth motion maps and convolutional neural networks
by Jiang Li, Xiaojuan Ban, Guang Yang, Yitong Li
Abstract: This paper presents an effective approach for recognising human actions from depth video sequences by employing Depth Motion Maps (DMMs) and Convolutional Neural Networks (CNNs). Depth maps are projected onto three orthogonal planes, frame differences under each view (front/side/top) are then accumulated through an entire depth video sequence thus generating a DMM. We build a model architecture of Multi-View Convolutional Neural Network (MV-CNN) containing multiple networks to deal with three DMMs (DMMf, DMMs, DMMt). The output of full-connected layer under each view is integrated as feature representation, which is then learned in the last softmax regression layer to predict human actions. Experimental results on MSR-Action3D dataset and UTD-MHAD dataset indicate that the proposed approach achieves state-of-the-art recognition performance and is appropriate for real-time recognition.
Keywords: real-time human action recognition; depth motion maps; multi-view convolutional neural networks.
A compact construction for non-monotonic key-policy attributebased encryption
by Junqi Zhang, Haiying Gao
Abstract: The Attribute-Based Encryption (ABE) scheme with monotonic access structure cannot deal with an access structure that is associated with the negation of attributes, which is not convenient for real world applications. In this paper, we attempt to propose a more expressive non-monotonic ABE scheme through a new method. To achieve this goal, we firstly propose a linear Two-mode Identity-Based Broadcast Encryption (TIBBE) scheme based on an Identity-Based Broadcast Encryption (IBBE) scheme. We introduce the concept of Identity-Based Revocation (IBR) for this scheme without increasing the size of parameters in IBBE. The scheme is selective identity secure under the m-DBDHE assumption. Then we convert the TIBBE scheme into a non-monotonic Key-Policy ABE (KP-ABE) scheme with compact parameters. Our KP-ABE scheme could achieve constant-size ciphertexts, and the scale of the private keys grows linearly with the scale of the attribute set. Besides, the computational cost is the lowest compared with other existing non-monotonic KP-ABE schemes.
Keywords: identity-based broadcast encryption; revocation scheme; linear secret-sharing schemes; non-monotonic access structure; selective security.
Research on link blocks recognition of web pages
by Gu Qiong, Wang Xianming
Abstract: The link block is a typical type of block structure of web pages; it is also an important and basic research object in the fields of web data processing and web data mining. Nevertheless, the existing research on links only focuses on granularities such as websites, pages and single links. Research results based on block-level links are extremely rare. In view of the significance and deficiency of this issue, firstly, block and block tree are proposed as the basic concepts of subsequent explorations, and then an approach of building block trees is put forward. Secondly, four rules for link block discrimination and two indicators for recognition results evaluation are put forward based on the concept of block, the two evaluation indicators are named as Link Coverage Rate (LCR) and Code Coverage Rate (CCR) respectively. Finally, a strategy named Forward Algorithm for Discovery of Link Block (FAD) is proposed and a corresponding experiment with different parameters is performed to verify the strategy. The results show that the FAD can be flexible to achieve recognition of link blocks under different granularity conditions. Concepts and approaches presented in this paper have a good prospect in the fields of web data processing, web data mining such as advertising block recognition, web page purification, page importance evaluation and web content extraction.
Keywords: web; block trees; link blocks; discrimination; recognition.
AdaBoost based conformal prediction with high efficiency
by Yingjie Zhang, Jianxing Xu, Hengda Cheng
Abstract: Conformal prediction presents a novel idea whose error rate is provably controlled by given significant levels. So the remaining goal of conformal prediction is its efficiency. High efficiency means that the predictions are as certain as possible. As we know, ensemble methods are able to obtain a better predictive performance than that obtained from any of the constituent models. An ensemble method such as random forest has been used as an underlying method to build a conformal predictor. But we dont know the differences of conformal predictors with and without ensemble methods， or how the corresponding performances are improved. In this paper, an ensemble method, AdaBoost, is used to build a conformal predictor, and we introduce another evaluation metric-correct efficiency, which measures the efficiency of correct classification correctly. The good performance of AdaBoost based conformal predictor (CP-AB) has been validated on seven datasets. The experimental results show that the proposed method has a much higher efficiency.
Keywords: machine learning; conformal prediction; AdaBoost; efficiency; ensemble; support vector machine; decision tree; weak classifiers; p-value; prediction label.
Comparative analysis of hierarchical cluster protocols for wireless sensor networks
by Chirihane Gherbi, Zibouda Aliouat, Mohamed Benmohammed
Abstract: Wireless Sensor Networks (WSNs) basically consist of low cost sensor nodes deployed in an interesting area for collecting data from the environment and relaying them to a sink, where they will be processed and then sent to an end user. Since wireless nodes are severely power-constrained, the major concern is how to conserve nodes' energy so that network lifetime can be last longer enough till the expected normal end of the network mission. Since WSNs may be formed by a large number of nodes, it is rather complex, or even unfeasible, to analytically model a WSN and it usually leads to oversimplified analysis with limited confidence. Besides, deploying test-beds requires a huge effort. Therefore, simulation is essential to study WSN behaviour. However, it requires a suitable model based on solid assumptions and an appropriate framework to ease implementation. In addition, simulation results rely on the particular scenario under study (environment), hardware and physical layer assumptions, which are not usually accurate enough to capture the real behaviour of a WSN, thus jeopardising the credibility of results. However, detailed models yield to scalability and performance issues, owing to the large number of nodes that depend on application to be simulated. Therefore, the tradeoff between scalability and accuracy becomes a major issue when simulating WSN. In particular, we systematically analyse a few prominent WSN clustering routing protocols and compare these different approaches according to our taxonomy and several significant metrics. Finally, we summarise and conclude the paper with some pertinent future directions.
Keywords: energy saving; distributed algorithm; load balancing; cluster-based routing; wireless sensor network.
Virtual cluster optimisation for MapReduce-like applications
by Cairong Yan, Guangwei Xu
Abstract: Infrastructure-as-a-service clouds are becoming ubiquitous for provisioning virtual machines on demand. Cloud service providers expect to use the least resources to deliver the best services. As users frequently request virtual machines to build virtual clusters and run MapReduce-like jobs for big data processing, cloud service providers intend to optimise the virtual cluster to minimise network latency and subsequently reduce data movement cost. In this paper, we focus on the virtual machine placement issue for provisioning virtual clusters with minimum network latency in clouds. We define the distance as the latency between virtual machines and use it to measure the affinity of a virtual cluster. Such metric of distance indicates the considerations of virtual machine placement and the topology of physical nodes in clouds. Then we formulate our problem as the classical shortest distance problem and solve it by building an integer programming model. A greedy virtual machine placement algorithm is designed to get a compact virtual cluster. Furthermore, an improved heuristic algorithm is also presented for achieving a global resource optimisation. The simulation results verify our algorithms and the experiment results validate the improvement achieved by our approaches.
Keywords: virtual cluster; provisioning; resource optimisation; MapReduce programming model; shortest distance.
Harnessing betweenness centrality for virtual network embedding in tree topologies
by Mydhili Palagummi, Ricardo Lent
Abstract: We examine the virtual network embedding problem with QoS constraints and formulate an approach that exploits the betweenness centrality of VNE requests to improve performance. A pay-per-use revenue model is introduced to evaluate the algorithm. An evaluation study using datacentre-like substrates and a wide area topology compares the approach with four embedding methods from the literature and reports on the average revenue rate, embedding success probability, average number of VNE deployments, cost, and impact of substrate failures on the operation of the VNEs, confirming the efficacy of the proposed approach.
Keywords: virtual network embedding; revenue metric; cloud computing; network overlay; datacentre; simulation.
Detecting fake reviews via dynamic multimode network
by Jun Zhao, Hong Wang
Abstract: Online product reviews can greatly affect the consumers shopping decision. Thus, a large number of unscrupulous merchants post fake or unfair reviews to mislead consumers for their profit and fame. The common approaches to find these spam reviews are analysing the text similarity or rating pattern. With these common approaches we can easily identify ordinary spammers, but we cannot find the unusual ones who manipulate their behaviour to act just like genuine reviewers. In this paper, we propose a novel method to recognise these unusual ones by using relations among reviewers, reviews, commodities and stores. Firstly, we present four fundamental concepts, which are the quality of the merchandise, the honesty of the review, the trustworthiness of the reviewer and the reliability of the store, thus enabling us to identify the spam reviewers more efficiently. Secondly, we propose our multimode network model for identifying suspicious reviews and then give three corresponding algorithms. Eventually, we find that the multiview spam detection based on the multimode network can detect more subtle false reviews according to our experiments.
Keywords: fake review detection; honesty degree; shopping behaviour; multiview spam detection; dynamic multimode network.
DBSCAN-PSM: An improvement method of DBSCAN algorithm on Spark
by Guangsheng Chen, Yiqun Cheng, Weipeng Jing
Abstract: DBSCAN is a density-based data clustering algorithm, widely used in image processing, data mining, machine learning and other fields. With the increasing size of clusters, the parallel DBSCAN algorithm is widely used. However, we consider the current partitioning method of DBSCAN is too simple, and steps of GETNEIGHBORS query repeatedly access the dataset on Spark. So we proposed DBSCAN-PSM, which applies a new data partitioning and merging method. In the first stage of our method we import the KD-Tree, combine the partitioning and GETNEIGHBORS query, reduce the number of accesses to the dataset, and decrease the influence of I/O in the algorithm. In the second stage of our method we use the feature of points in merging so as to avoid the time costing of the global label. Experimental results showed that our new method can improve the parallel efficiency and the clustering algorithm performance.
Keywords: big data; DBSCAN; data partitioning; data merging.
Multimedia auto-annotation via label correlation mining
by Feng Tian
Abstract: How to automatically determine the label for multimedia object is crucial for multimedia retrieval. Over the past decade, significant efforts have been devoted to the task of multimedia annotation. The problem is difficult because an arbitrary multimedia object can capture a variety of concepts, each of which would require separate detection. The neighbour voting mechanism is known to be effective for multimedia object annotation. However, it estimates the relevance of a label with respect to multimedia content by labels' frequency derived from its nearest neighbours, which does not take into account the assigned label set as a whole. We propose LSLabel, a novel algorithm that achieves comparable results with label correlation mining. By incorporating the label correlation and label relevance with respect to multimedia content, the problem of assigning labels to multimedia object is formulated into a joint framework. The problem can be efficiently optimised in a heuristic manner, which allows us to incorporate a large number of feature descriptors efficiently. On two standard real-world benchmark datasets, we demonstrate that LSLabel matches the current state-of-the-art in annotation quality, and has lower complexity.
Keywords: label correlation; multimedia annotation; auto-annotation; correlation mining.
Dynamic trust evaluation model based on bidding and multi-attributes for social networks
by Gang Wang, Jaehong Park, Ravi Sandhu
Abstract: Mutual trust is the most important basis in social networks. However, many malicious nodes often deceive, collaboratively cheat, and maliciously recommend other nodes for getting more benefits. Meanwhile, because of the lack of an effective incentive strategy, many nodes can neither evaluate nor recommend. Thus, malicious actions have been aggravated in social networks. To solve these issues, we design a bidding strategy to incentivise nodes to do their best to recommend or evaluate service node. At the same time, we also use the TOPSIS method of selecting a correct service node for the system from networks. To guarantee reliability of the service node selected, we bring recommendation time influential function, service content similarity function and recommendation acquaintance function into the model to compute the general trust of the node. Finally, we give an update method for trust degree of node and experiments analysis.
Keywords: dynamic trust; trust evaluation model; bid; multi-attributes; TOPSIS; information entropy; recommendation trust; direct trust.
A personal local area information interaction system based on NFC and Bluetooth technology
by Tian Wang, Wenhua Wang, Ming Huang, Yongxuan Lai
Abstract: Taking attendance is a regular activity in society. Required class attendance is
common in Chinese colleges and universities. In most traditional classrooms, taking attendance may consume much time and some students may cheat by pretending to be their classmates, which makes the results unbelievable. Moreover, the interaction mode between the teacher and students is single and cannot support fast data interaction. To solve these problems, this paper proposes an information interaction system, which can not only speed up the process of taking attendance but also extend the information exchange mode. Firstly, we propose the NFC (Near Field Communication)based method to take attendance, which uses the rapid information exchange characteristic of NFC in the mobile phone. Furthermore, an ad-hoc scheme is introduced, based on which some students may be selected as the relay, which can greatly accelerate the attendance-taking process. Moreover, a lazy unbinding mechanism is proposed to prevent the students from taking attendance for others. Finally, based on Bluetooth technology, the system realises file transfer, which
extends the information exchange mode. Real experimental results demonstrate the feasibility of the proposed system.
Keywords: taking attendance; information interaction; NFC; lazy unbinding scheme; relay scheme.
A risk adaptive access control model based on Markov for big data in the cloud
by Hongfa Ding, Changgen Peng, Youliang Tian, Shuwen Xiang
Abstract: One of most important problems faced by cloud computing and cloud storage is identity and access management. The main problems of the application of access control in the cloud are the necessary flexibility and scalability to support a large number of users and resources in a dynamic and heterogeneous environment, with collaboration and information sharing needs. This paper proposes a risk self-adaptive dynamic access control model for big data stored in cloud and processed by cloud computing. The suggested model employs the Markov method and Shannon information theory. First, the simple formal adversary model for our risk adaptive access control model is presented. Second, a modification of eXtensible Access Control Markup Language (XACML) framework is given, and some new and enhanced components (including a novel risk evaluation component) are added in the modification. Then, we present Markov based methods to calculate the risk values of access requests, identify the user and supervise the access behaviour according to the job obligations of users and classification of data. At last, an incentive mechanism similar to a credit system is designed to supervise all the access behaviours of subjects, and the risky requests and risky users are restrained effectively by this mechanism. Our method is easy to deploy as the model is extended from the standard XACML. The administrator just needs to label the object data and record the request and access behaviour by comparing with the other work. This method is more effective and suitable to control the access in large-scale information system (e.g. cloud-based system), and protect the sensitive and privacy data for the data owners.
Keywords: risk-based access control; privacy protection; risk management; cloud computing.
Watermarking based authentication and image restoration in multimedia sensor networks
by Hai Huang, Dongqing Xie, Qi Ke
Abstract: In this paper, we proposed a novel watermarking based secure image acquisition scheme for Multimedia Sensor Networks (MSNs), which provides support for image authentication and restoration. In the proposed scheme, the sensor node groups image frames. Two successive frames compose the non-overlapping authentication and restoration group. The watermarking bits, which are the version of the down-sampling image, are computed from the first image and reversibly embedded into the second image. The sink does the verification and restoration by reversible watermark. Compared with previous work, our approach can not only implement authentication but also improve image quality. Experimental results show that our scheme can achieve gains in terms of authentication and packet loss tolerance to improve the image quality.
Keywords: multimedia sensor networks; secure image restoration; authentication; watermarking.
Coverless information hiding method based on the keyword
by Hongyong Ji, Zhangjie Fu
Abstract: Information hiding is an important
Keywords: coverless information hiding; big data; Chinese mathematical expression; word segmentation.
Measurement method of carbon dioxide using spatial decomposed parallel computing
by Nan Liu, Weipeng Jing
Abstract: According to the carbon dioxide's characteristics of weak absorption in the ultraviolet and visible (UV-VIS) band, a measurement method based on spatial decomposed parallel computing of traditional differential optical absorption spectroscopy is proposed to measure CO2 vertical column concentration in ambient atmosphere. First, the American Standard Profile is used to define the solar absorption spectrum, and the spectrum acquisition of the incident light converged by the telescope is described as observed parameters. On these bases, a spectrometer line model is established. Then, atmospheric radiation transmission is simulated using parallel computing, which reduces the computational complexity while balancing the interference that participates in the fitting. Simulation analyses show that the proposed method can reduce the computational complexity, and the run time is reduced by 1.18 s compared with IMLM and IMAP-DOAS in the same configuration. The proposed method can also increase accuracy, with its inversion error reduced by 5.3% and residual reduced by 0.8% compared with differential optical absorption spectroscopy. The spatial decomposed parallel computing method has advantages in processing CO2, and it can be further used in the research into carbon sinks.
Keywords: differential optical absorption spectroscopy; ultraviolet and visible band; spatial decomposed parallel computing method; vertical column concentration; spectrometer; fitting.
Design and implementation of an Openflow SDN controller in the NS-3 discrete-event network simulator
by Ovidiu Mihai Poncea, Andrei Sorin Pistirica, Florica Moldoveanu, Victor Asavei
Abstract: The NS-3 simulator comes with the Openflow protocol and a rudimentary controller for classic layer-2 bridge simulations. This controller lacks the basic functionality provided by real SDN controllers, making it inadequate for experimentation in the scientific community. In this paper, we propose a new controller with an architecture and functionality similar to that of full-fledged controllers yet simple, extensible and easy to modify - all characteristics specific to simulators.
Keywords: networking; software defined networking; SDN controller; NS3; NS-3; simulators.
Exploring traffic conditions based on massive taxi trajectories
by Dongjin Yu, Jiaojiao Wang, Ruiting Wang
Abstract: As increasing volumes of urban traffic data become available, more and more opportunities arise for the data-driven analysis that can lead to the improvements of traffic conditions. In this paper, we focus on a particularly important type of urban traffic dataset: taxi trajectories. With GPS devices installed, moving taxis become valuable sensors for the traffic conditions. However, analysing these GPS data presents many challenges owing to their complex nature. We propose a new approach that transforms the trajectories of each moving taxi as a document consisting of the traversed street names, which enables semantic analysis of massive taxi datasets as document corpora. More specifically, we identify traffic topics through textual topic modelling techniques, and then cluster trajectories under these topics to explore the traffic conditions. The effectiveness of our approach is illustrated by case study using a large taxi trajectory dataset acquired from 3743 taxis in a city.
Keywords: vehicle trajectory; map matching; traffic regions; latent Dirichlet allocation; trajectory clustering; visualisation.
Keyword guessing on multi-user searchable encryption
by Zhen Li, Minghao Zhao, Han Jiang, Qiuliang Xu
Abstract: Searchable encryption provides a practical method that enables a client to store an encrypted database on an untrusted server, while supporting keyword search in a secure manner. It has gained extensive research interests with increasing concerns on security in cloud computing. Multi-user searchable encryption is more compatible for multi-tenancy and massive scalability property of cloud service. Most of these schemes are constructed using public key encryption. However, public key encryption with keyword search is vulnerable to keyword guessing attack. This is mainly because the keyword space is overwhelming smaller than the polynomial level of secure parameter and users usually query commonly-used keywords with low entropy. Consequently, a secure channel is necessarily involved for secret information transformation, which leads to an extra severe burden for the cloud system. This vulnerability is recognised in traditional searchable encryption, but it is still undecided whether it also exists in multi-user setting. In this paper, we firstly point out that keyword guessing attack is also a problem in the multi-user setting without the supposed secure channel. By an in-depth investigation of some multi-user searchable encryption schemes proposed recently and simulating the keyword guessing attack on them, we present that none of these schemes can resist this attack. We make a comprehensive security definition and propose some open problems related to multi-user searchable encryption.
Keywords: cloud computing; keyword guessing; searchable encryption; multi-user.
Semi-supervised dimensionality reduction based on local estimation error
by Xianfa Cai
Abstract: Graph construction is one of the key steps in graph-based semi-supervised learning. However, the neighbourhood graph of most semi-supervised methods is unstable by virtue of sensitivity to the selection of neighbourhood parameter and inaccuracy of the edge weights of the neighbourhood graph, which easily leads to dramatic degradation of performance. Since local models are trained only with the points related to the particular one，local learning methods often outperform global ones. The good performance of local learning methods indicates that the label of a point can be well estimated by its neighbours. Inspired by the good performance of the local learning method, this paper proposes a feasible strategy called semi-supervised dimensionality reduction based on local estimation error (LEESSDR) by using local learning projections (LLP) for semi-supervised dimensionality reduction. The algorithm sets the edge weights of neighbourhood graph through minimising the local estimation error, and can effectively preserve the global geometric structure of the sampled data set as well as preserving its local one. Since LLP does not require its input space to be locally linear, even if it is nonlinear, LLP maps it to the feature space by using kernel functions and then obtains its local estimation error in the feature space. The feasibility and effectiveness of the proposed method are verified on two popular face databases (YaleB and CMU PIE) with promising classification accuracy and favourable robustness.
Keywords: local learning projections; side-information; semi-supervised learning; graph construction.
Securing SDN controller and switches from attacks
by Uday Tupakula, Vijay Varadharajan
Abstract: Software Defined Networking (SDN) enables programmable networks and offers several advantages to simplify network management tasks for administrators. Hence it is increasingly used for the management of complex networks such as the cloud. Although there are several benefits with SDN, it also leads to new types of attack. In this paper we describe how the attacks in traditional networks can lead to attacks in SDN. Then we propose techniques for securing the SDN controller and the switches from malicious end-host attacks. Our model makes use of trusted computing and introspection-based intrusion detection to deal with attacks in SDN. We have developed a security application for the SDN controller to validate the state of the switches in the data plane and enforce the security policies to monitor the virtual machines. The attack detection component takes into account the specific features of the virtual machine in detecting the attacks and isolating the malicious virtual machines. It uses the introspection feature at the hypervisor layer to collect the system call traces of programs running in a monitored VM. We have developed a feature extraction method, named vector of n-grams, which represents the traces in an efficient way without losing the ordering of system calls. The flows from the malicious hosts are dropped before they are processed by the switches or forwarded to the SDN controller. Hence our model protects the switches and the SDN controller from the attacks.
Keywords: SDN security; trusted computing; virtual machine introspection; security attacks.
Multi-objective fuzzy optimisation of knowledge transfer organisations in the big data environment
by Chuanrong Wu, Feng Li
Abstract: With the advent of the big data era, the information from big data has become a type of important knowledge that enterprises need for innovation. The knowledge transfer mode and the influence factors of big data knowledge providers are different from those of traditional knowledge providers. Based on an analysis of organisational composition and characteristics of knowledge transfer in the big data environment, the influence factors and evaluation index systems of big data knowledge providers are established. A multi-objective fuzzy optimisation model is constructed to derive satisfactory knowledge providers by finding optimal sequences. Meanwhile, this model can provide a cooperative decision-making method for knowledge transfer organisations to enhance the efficiency of knowledge transfer in the big data environment.
Keywords: big data; knowledge transfer; multi-objective optimisation; organisation; decision making method.
On Parallelisation of image dehazing with OpenMP
by Tien-Hsiung Weng, Yi-Siang Chen, Huimin Lu
Abstract: In this paper, we present our learning experience on the design and implementation of image dehazing parallel code with OpenMP developed from the existing fast sequential version. The aim of this work is to present an analysis of a case study showing the development of parallel haze removal with practical and efficient use of shared memory multi-core servers. Implementation technique and result discussions in terms of program improvements that may be needed to support parallel application developers with similar high performance goals are presented. Preliminary studies, results and experiments on haze removal application program are executed on multi-core shared memory platforms, and results show that the performance of the proposed parallel code is promising.
Keywords: OpenMP; image haze removal; multicores; parallel programming.
Fast graph centrality computation via sampling: a case study of influence maximisation over online social networks
by Rui Wang, Yongkun Li, Yinlong Xu
Abstract: Graph centrality computation, e.g., asking for the most important vertices in a graph, may incur a high time cost with the increasing size of graphs. To address this challenge, this paper presents a sampling-based framework to speed up the computation of graph centrality. As a use case, we study the problem of influence maximisation, which asks for the k most influential nodes in a graph to trigger the largest influence spread, and presents an IM-RWS algorithm. We experimentally compare IM-RWS with the state-of-the-art influence maximisation algorithms IMM and IM-RW, and the results show that our solution can bring a significant improvement in efficiency as well as some improvement in empirical accuracy. In particular, our algorithm can solve the influence maximisation problem in graphs containing millions of nodes within tens of seconds, with an even better performance result in terms of influence spread.
Keywords: random walk; sampling; graph centrality; online social networks; influence maximisation.
Non-intrusive load monitoring and its challenges in a non-intrusive load monitoring system framework
by Qi Liu, Min Lu, Xiaodong Liu, Nigel Linge
Abstract: With the increasing energy demand and electricity price, researchers show more and more interest among the residential load monitoring. In order to feed back the individual appliances energy consumption instead of the whole-house energy consumption, Non-Intrusive Load Monitoring (NILM) is a good choice for residents to respond to the time-of-use price and achieve electricity saving. In this paper, we discuss the system framework of NILM and analyse the challenges in every module. Besides, we study and compare the public datasets and accuracy metrics of NILM techniques.
Keywords: NILM; system framework; data acquisition; event detection; feature extraction; load disaggregation.
A similarity algorithm based on Hamming distance used to detect malicious users in cooperative spectrum sensing
by Libin Xu, Pin Wan, Yonghua Wang, Ting Liang
Abstract: Spectrum sensing (SS) is a key technology in cognitive radio networks (CRNs). Many statistical methods have been proposed to improve the sensing performance. However, rare studies take security into account. The collaborative spectrum sensing (CSS) is vulnerable to the potential attacks from malicious users (MUs). Most existing MU detection methods are reputation-based, it is incapable when the MUs dominate the network. Although Markov model characterises the spectrum state behaviour more precisely, there is a scarcity of MU detection in the literature. In this paper, a Hamming distance check (HDC) is proposed to detect MUs. The Hamming distance between all the sensing nodes is calculated. For the reports from MUs is different from honest users (HUs), we can find the MUs and exclude in the fusion process. A new trust factor (TF) is proposed to increase the effects of trustworthy nodes in the final decision. The simulation results show that the impact of MU to CSS cannot be ignored. The proposed algorithm can effectively detect the MUs without prior knowledge. In addition, our proposed method can perform better than the existing approaches.
Keywords: malicious user; attack; Hamming distance; cognitive radio networks.
A self-adaptive quantum steganography algorithm based on QLSb modification in watermarked quantum images
by Qu Zhiguo
Abstract: As one of important research branches of quantum information hiding, quantum steganography embeds secret information into quantum images for covert communication by integrating quantum secure communication technology and classical steganography. In this paper, based on the Novel Enhanced Quantum Representation (NEQR), a novel quantum steganography algorithm is proposed to transfer secret information by virtue of quantum watermarked images. In order to achieve this goal, the Least Significant Qubit (LSQb) of the quantum carrier image is replaced with the secret information by implementing a quantum circuit. Compared with the previous quantum steganography algorithms, the communicating parties can recover the secret information tampered, meanwhile the tamperers can be located effectively. In the experiment result, the Peak Signal-to-Noise Ratios (PSNRs) are calculated for different quantum watermarked images and quantum watermarks, which demonstrate the imperceptibility of the algorithm is good and the secret information embedded can be recovered by virtue of its self-adaptive mechanism.
Keywords: quantum steganography; quantum least significant bit; watermarked quantum carrier image.
Distributed continuous KNN query over moving objects
by Xiaolin Yang, Zhigang Zhang, Yilin Wang, Cheqing Jin
Abstract: The continuous k-nearest neighbour(CKNN) queries over moving objects have been widely researched in many fields. However, existing centralised works cannot work anymore and distributed solutions suffer the problem of index maintainance, high communication cost and query latency. In this paper, we firstly propose a distributed hybrid indexing strategy that combines the SSGI (Spatial-temporal Sensitive Grid Index) and the DQI (Dynamic Quad-tree Index). The SSGI is proposed to locate the spatial range that contains the final results, and the DQI is used for data partitioning. Then, we introduce an algorithm named HDCKNN to implement the CKNN queries. In comparison of existing work, HDCKNN can achieve the final result in one round iteration, while existing methods require at least two rounds of iteration. Extensive experiments show that the performance of the proposed method is more efficient than state-of-the-art algorithms.
Keywords: moving objects;continuous k-nearest neighbour query; distributed query processing; hybrid indexing.
A cross-layer QoS model for space DTN
by Aiguo Chen, Xuemei Li, Guoming Lu, Guangchun Luo
Abstract: Delay/disruption tolerant networking (DTN) technology is considered as a new solution for highly stressed communications in space environments. An IP-based DTN network can support more flexible communication services, which is one of the research hotspots of space networking. At the same time, it is still an important and difficult problem to optimise the limited network resources allocation and guarantee the QoS of different services in a DTN network. For this challenge, a cross-layer QoS model that considers application, network, and node layer QoS requirements and resource limitations is proposed in this paper. In addition, a comprehensive admission control scheme that ensures productivity and fairness is employed. The results of experiments and analyses conducted demonstrate the benefits to be derived from this approach.
Keywords: DTN; QoS model; cross-layer; space network.
Industrial software abnormal behaviour analysis based on multi-granularity error propagation
by Cheng Peng
Abstract: Bugs in software are inevitable. This study on industrial software abnormal behaviour propagation mechanism trigged by bugs provides the way for people to grasp the execution rule and to adopt corresponding pinning measurements. According to the situation of abnormal behaviour propagation at different granularity software entities, the factor of error propagation probability affecting the abnormal behaviour propagation is proposed; the corresponding definition and calculation method are also investigated. The software abnormal behaviour propagation process model is constructed with reference to the compartment model and individual models and the factor mentioned above, which improves the model expression ability and enhances the model competence and accuracy. Then, an abnormal behaviour propagation analytical method is applied to the online electronic shopping system. The results verify the correctness and feasibility of the propagation mechanism.
Keywords: industrial software; abnormal behaviour; propagation model; system bug.
Modelling the propagation of soft errors in programs
by Lixing Xue, Decheng Zuo, Zhan Zhang
Abstract: Soft errors are a category of typical transient errors caused by multiple external factors. Owing to the continuous exponential growth of transistors in processors, the mainstream computing paradigm and increasingly complex operational environment, they have become an urgent challenge in ground-level systems. To handle these errors, making clear the error propagation is the key step. Fault injection campaigns that are used to study error propagation generate only limited results of injected soft errors and cannot deduce other errors. As the injection space becomes vast, this traditional method appears to be powerless. To attack this problem, this paper proposes a method to study and model the propagation of soft errors within a program. Based on dynamic instructions traced in an error-free program execution, the ACE analysis classifies soft errors in architectural registers into benign ones and non-benign ones. The benign errors are considered to be derated in the propagation. Then we build a crash model and an improved DDG to analyse the propagation of each non-benign error and to predict its consequence (crash or silent data corruption). If the error is considered to cause a crash, the crash latency and the propagation path are also predicted. The method can be used to predict outcomes of programs under soft errors as well as occurrence of a certain category of error consequences. Extensive fault-injection experiments are provided to validate the proposed method from multiple perspectives.
Keywords: soft error; error propagation; silent data corruption; crash; architecturally correct execution; dynamic dependency graph.
Mimicry honeypot: an evolutionary decoy system
by Leyi Shi, Yuwen Cui, Han Xu, Honglong Chen, Deli Liu
Abstract: Motivated by the mimic-and-evolve phenomenon for species rivalry, we present a novel concept of mimicry honeypot, which can bewilder the adversaries through comprehensively exploiting the protective coloration, warning coloration and mimicry evolution according to the changes of network circumstance. The paper firstly gives the definition of the protective coloration and warning coloration for cyber defence, formalises the mimicry honeypot model, discusses the critical issues of environment perception and mimicry evolution, and implements a mimicry prototype through web service platform and genetic algorithm. Afterwards, we perform experiments with the mimicry prototype deployed both in our private campus network and on the internet. Our empirical study demonstrates that the mimicry honeypot has better efficiency than the traditional decoy system.
Keywords: mimicry honeypot; warning coloration; protective coloration; evolution; genetic algorithm.
Protection of location private privacy with recommendation from a decision tree algorithm
by Hongjun Dai
Abstract: Location-based services (LBSs) have been widely used in apps on phones. They help apps to provide better service for us, for example, a navigation app guides us during a trip. However, LBSs can also cause the problem of privacy disclosure. For some apps, like weather apps, a coarse location accuracy (LA) is enough to achieve their function, while they get our location within several metres. In this paper, we propose a framework for LBS Management (LBSM) to protect users personal location message by providing multi-level LA to different apps. The framework will create two decision trees for apps and users to analyse the properties, and then use the decision trees to make the classification. According to the classification result, the LA level of an app is finally determined. For better service, the decision trees will adjust constantly when the cloud servers obtain new data of apps and users or their properties have changed. Furthermore, some improvements have been made on the basis of traditional Iterative Dichotomiser 3 (ID3) algorithm to make a more accurate classification result. Experiments show that users personal information can be better protected while also obtaining satisfactory service with our proposed LBSM.
Keywords: privacy protection; LBS; ID3 algorithm; location accuracy; app classification; user classification.
Evaluation of directive-based performance portable programming models
by M. Graham Lopez, Wayne Joubert, Veronica Vergara Larrea, Oscar Hernandez, Azzam Haidar, Stanimire Tomov, Jack Dongarra
Abstract: We present an extended exploration of the performance portability of directives provided by OpenMP 4 and OpenACC to program various types of node architecture with attached accelerators, both self-hosted multicore and offload multicore/GPU. Our goal is to examine how successful OpenACC and the newer offload features of OpenMP 4.5 are for moving codes between architectures, and we document how much tuning might be required and what lessons we can learn from these experiences. To do this, we use examples of algorithms with varying computational intensities for our evaluation, as both compute and data access efficiency are important considerations for overall application performance. To better understand fundamental compute vs. bandwidth bound characteristics, we add the compute-bound Level 3 BLAS GEMM kernel to our linear algebra evaluation. We implement the kernels of interest using various methods provided by newer OpenACC and OpenMP implementations, and we evaluate their performance on various platforms including both x86_64 and Power8 with attached NVIDIA GPUs, x86_64 multicores, self-hosted Intel Xeon Phi KNL, as well as an x86_64 host system with Intel Xeon Phi coprocessors. We update these evaluations with the newest version of the NVIDIA Pascal architecture (P100), Intel KNL 7230, Power8+, and the newest supporting compiler implementations. Furthermore, we present in detail what factors affected the performance portability, including how to pick the right programming model, its programming style, its availability on different platforms, and how well compilers can optimise and target multiple platforms.
Keywords: OpenMP 4; OpenACC; performance portability; programming models.
Genuine and secure public auditing scheme for the outsourced data
by Jianhong Zhang
Abstract: The most common concerns for users in cloud storage are data integrity, confidentiality and availability, so various data integrity auditing schemes for cloud storage have been proposed in the past few years, some of which achieve privacy-preserving public auditing, data sharing and group dynamic, or support data dynamic. However, as far as we know, until now yet there doesn't exist a practical auditing scheme which can simultaneously realise all the functions above; in addition, in all the existing schemes, block authentication tag (BAT) is adopted by the data owner to achieve data integrity auditing. Nevertheless, it's a arduous task to compute BATs for the resource-constrained data owner. In this paper, we propose a novel privacy-preserving public auditing scheme for shared data in the cloud, which can also support data dynamic operations and group dynamic. Our scheme has the following advantages: (1) we introduce proxy signature into the existing auditing scheme to reduce the cloud user's computation burden; (2) by introducing a Lagrange interpolating polynomial, our scheme realises the identity's privacy-preserving without increasing computation cost and communication overhead, moreover it makes group dynamic simple; (3) it can realise the practical and secure dynamic operations of shared data by combining the Merkle Hash Tree and index-switch table which is built by us; (4) to protect the data privacy and resist the active attack, the cloud storage server hides the actual proof information by inserting its private key in producing proof information process. After theoretical analysis demonstrates our scheme's security,experiment results show that our scheme not only has the low computational and communication overhead for data verification but also can complete the group dynamics with great speed.
Keywords: cloud computing; self-certified cryptography; integrity checking; security proof; provably secure; random oracle model; cryptography.
A new mobile opportunity perception network strategy and reliability research in coal mines
by Ping Ren, Jing-Zhao Li, Da-Yu Yang
Abstract: Aiming at the characteristics of application of Internet of Things in underground coal mines, this paper proposes a new method of mobile opportunity perception in coal mines, which makes full use of underground mine mobile resources. The method mainly uses the existing resources, without increasing or adding sensing devices, by deploying wired nodes, wireless fixed nodes and wireless mobile nodes, to establish the collaborative working mechanism of mobile nodes and fixed nodes, and accordingly builds environment opportunity perception and information transmission of the whole place in the mine and achieves a comprehensive perception of the mine. A novel sparse heterogeneous fusion network and its mixed node arrangement strategy and low-order error detection model between nodes are established, and the loss probability and redundancy perception between the mobile and fixed nodes are analysed, which improves the reliability of the system. The moving speed, the average moving distance and the communication coverage time of mobile nodes are analysed, and an interactive data quantity is obtained. On this basis, the system hardware design and experiment are carried out. Experimental results show the superiority of this method, which provides a new method and idea for industrial and mining enterprises in the sensing layer and transport layer application of the Internet of Things.
Keywords: mobile sensing; opportunistic routing; heterogeneous fusion; low false alarm; lost probability.
Distributed admission control algorithm for random access wireless networks in the presence of hidden terminals
by Ioannis Marmorkos, Costas Constantinou
Abstract: We address the problem of admission control for wireless clients in WLANs taking into account collisions between competing access points and considering explicitly the effect of hidden terminals, which play a prominent role in optimised client association. We propose an efficient, distributed admission control algorithm, where the wireless client node decides locally on which access point it will associate with in order to maximise its link throughput. The client can choose to optimise either its uplink or downlink throughput, depending on the type of traffic it predominantly intends to exchange with the network. The proposed approach takes into account the full contention resolution of the RTS/CTS IEEE802.11 medium access control protocol and leads towards an increase of the total throughput for the whole network. Finally, an algorithm is proposed, which can serve also as the basis for the development of efficient traffic offloading protocols in heterogeneous 5G networks.
Keywords: admission control; IEEE802.11; multi-cell; WiFi offloading.
Modelling geographical effect of user neighborhood on collaborative web service QoS prediction
by Zhen Chen, Limin Shen, Dianlong You, Huihui Jia, Chuan Ma, Feng Li
Abstract: QoS prediction is a task to predict the unknown QoS value of an active user to a web service that he/she has not accessed previously for supporting appropriate web service recommendation. Existing studies adopt collaborative filtering methods for QoS prediction, while the inherent issues of data sparsity and cold-start in collaborative filtering have not been resolved satisfactorily, and the role of geographical context is also underestimated. Through data analysis on a public real-world dataset, we observe that there exists a positive correlation between a users QoS values and geographical neighborhoods ratings. Based on the observation, we model the geographical effect of user neighborhood on QoS prediction and propose a unified matrix factorisation model by capitalising the advantages of geographical neighborhood and latent factor approaches. Experimental results exhibit the significance of geographical context on modelling user features and demonstrate the feasibility and effectiveness of our approach on improving QoS prediction performance.
Keywords: web service; QoS prediction; collaborative filtering; geographical effect; matrix factorisation.
Researches on data encryption scheme based on CP-ASBE of cloud storage
by Xiaohui Yang, Wenqing Ding
Abstract: The ciphertext-policy attribute-set based encryption (CP-ASBE) based on a single authorisation centre is easy to become the security bottleneck of the system. With the support of trusted measurement technology, a novel method of CP-ASBE based on multiple attribute authority (AA) is proposed to solve this problem, and an encryption scheme is designed for cloud storage, which includes data storage, data access, data encryption and trusted measurement scheme for AA. The security performance and time cost of the encryption scheme are simulated, and the results show that the scheme can improve the security of users' data in the cloud storage environment.
Keywords: cloud storage; CP-ASBE; authorisation centre; attribute authority; trusted measurement.
Deferred virtual machine migration
by Xiaohong Zhang, Jianji Ren, Zhizhong Liu, Zongpu Jia
Abstract: The rapid growth in demand for cloud services has led to the establishment of huge virtual machines. Online Virtual Machine (VM) migration techniques offer cloud providers a means to reduce power consumption while keeping quality of service. However, the efficiency of online migration is suboptimal since it is degraded by the execution of redundant migrations. To alleviate this problem, we introduce a load-aware VM migration technique. The key idea is to defer these migrations and perform a quick analysis of the loads on target servers before launching any migration. This consolidation is applied in two steps: first, all migrations are tested and a set of candidates that are suspected to be redundant is formed and postponed for a short time. Then, the servers are analysed and only a subset of the migration candidates is activated. Our selection mechanism is conservative in the sense that it avoids selecting and activating VM migrations that are unlikely to cause a harmful overloading on the target servers. Taking this conservative migration policy leads to an overall effective execution of virtual machines. Our experiment results demonstrate the usefulness and effectiveness of our method.
Keywords: cloud computing; virtual machine migration; virtual machine consolidation; power consumption.
Impact of using multi-levels of parallelism on HPC applications performance hosted on Azure cloud computing
by Hanan Hassan, Mona Kashkoush, Mohamed Azab, Walaa Sheta
Abstract: The use of High-Performance Computing (HPC) applications has increased progressively in scientific research and industry. Cloud computing attracts HPC users because of its extreme cost efficiency. The reduced cost is the result of the successful employment of multilayer-virtualisation, enabling dynamic elastic resource-sharing between different tenants. In this paper, we evaluate the impact of using multi-levels of parallelism on computationally intensive parallel tasks hosted on a cloud virtualised HPC cluster. We use multi-levels of parallelism through a set of experiments employing both message passing and multi-threading techniques. Our evaluation addresses two main perspectives, the performance of applications and cost of running HPC applications on clouds. We use Millions of Operation per Seconds (MOPS) and speed-up to evaluate the computational performance. To evaluate the cost we use United States Dollar/MOPS (USD/MOPS). The experiments on two different clouds are compared against each other and with published results for Amazon EC2 cloud. Results show that balancing the workload between processes and threads per process is the key factor to maintain high performance with reasonable cost.
Keywords: HPC performance on cloud computing; Azure cloud computing; hybrid MPI+OPENMP; NPB and NPB-MZ benchmarks.
NFV deployment strategies in SDN network
by Chia-Wei Tseng, Po-Hao Lai, Bo-Sheng Huang, Li-Der Chou, Meng-Chiou Wu
Abstract: The emergence of the internet has resulted in the relative expansion of complicated network architectures. Accordingly, the traditional network architecture can no longer meet the demands of a new and rapidly changing network service. Given the emergence of software defined network (SDN) and network functions virtualisation (NFV), technologies can now transform the current and complicated network architecture into a programmable, virtualised, and standardised managed architecture. This study aims to design and implement the rapid deployment strategies of the NFV services in SDN. Six different rapid deployment technologies are addressed based on the linked clone and the full clone cases. These technologies can accelerate the speed of deployment by enabling the NFV technology with an intelligent configuration. Experimental results show that the proposed parallel clone strategy in the linked clone scenario exhibited a better performance in terms of time efficiency compared with the other strategies.
Keywords: software defined network; network functions virtualisation; rapid deployment; network management.
A new localisation strategy with wireless sensor networks for tunnel space model
by Ying Huang, Yezhen Luo
Abstract: Because the three-dimensional distance vector hop (DV-Hop) localisation has some large error phenomena in the location of tunnel space model, an improved three-dimensional DV-Hop fixed node location strategy is proposed based on the wireless sensor network for the relationship between the geometric model of tunnel space. This strategy analysed the deficiency of traditional three-dimensional DV-Hop algorithm about counts hop and calculations distance on two WSN nodes, and it employed the relationship between beacon node and distance to modify the hop count of DV-Hop, then using the differential method corrects distance error from the unknown node to the beacon node. According to the tunnel space model, the selection mechanism is introduced, and three optimal beacon nodes are selected to locate the unknown nodes and further improve the location accuracy. The traditional three-dimensional DV-Hop calculation hops neglect the distance between nodes, and the improved strategy is used to correct the distance between unknown nodes and beacon nodes. The experimental results show that the improved DV-Hop localisation strategy has greatly improved the location accuracy, it can be widely used to solve the internet of things problems.
Keywords: tunnel space model; three-dimensional distance vector hop; received signal strength indicator; localisation algorithm.
Special Issue on: Wireless Network Technologies and Applications
Checkpointing distributed application running on mobile ad hoc networks
by Houssem Mansouri, Nadjib Badache, Makhlouf Aliouat, Al-Sakib Khan Pathan
Abstract: Mobile Ad hoc NETwork (MANET) is a type of wireless network consisting of a set of self-configured mobile hosts that can communicate with each other using wireless links without the assistance of any fixed infrastructure. This has made it possible to create distributed mobile computing applications and has also brought several new challenges in the field of distributed algorithm design. Checkpointing is a well explored fault tolerance technique for the wired and cellular mobile networks. However, it is not directly applicable to MANET owing to its dynamic topology, limited availability of stable storage, partitioning and the absence of fixed infrastructure. In this paper, we propose an adaptive, coordinated and non-blocking checkpointing algorithm to provide fault tolerance in cluster-based MANET, where only a minimum number of mobile hosts in the cluster should take checkpoints. The performance analysis and simulation results show that the proposed scheme requires less coordinating-message cost and performs well compared with the related previous works.
Keywords: MANET; distributed mobile computing; fault tolerance; clustering; checkpointing.
Cluster-based routing protocol using traffic information
by Hamza Toulni, Benayad Nsiri
Abstract: In recent years, vehicular ad-hoc networks (VANET) have gained much attention from researchers and the different actors in the field of transport, because of their crucial role in inter-vehicular communication and road safety. However, VANET still faces many challenges in terms of implementation. Furthermore, the routing in VANET is the most critical issue, because of its major role in the communication between vehicles. So the routing strategy must consider the periodic change in the topology and other characteristics of VANET. Moreover, the method adopted in the routing must provide the best flow of data in view of various performance metrics. In this paper, we propose a cluster-based routing protocol using the road traffic information to ensure the packet transmission in the most reliable manner and in record time. The proposed routing protocol is simulated in a city environment, and the experimental results show that our protocol is very effective.
Keywords: ontology, vehicular ad-hoc networks, traffic information, cluster-based routing protocols
Detection and mitigation of pulse-delay attacks in pairwise-secured wireless sensor networks
by Tooska Dargahi, Hamid H.S. Javadi, Hosein Shafiei, Payam Mousavi
Abstract: With the advances in technology, there has been an increasing interest
in the use of Wireless Sensor Networks (WSNs). WSNs are vulnerable to a
wide class of attacks among which pulse-delay attack puts severe threats to the
dependability of such networks. This paper proposes a distributed approach to
detect and mitigate such attack in pairwise-secured WSNs. It provides a lower
bound on the radio range under which the distributed approach can be performed.
Our simulation experiments validate the correctness and efficiency of the proposed
Keywords: wireless sensor networks; security; time synchronisation; pulse-delay attacks; self-healing systems.
PSCAR: a proactive-optimal-path selection with coordinator agents assisted routing for vehicular ad hoc networks
by Souaad Boussoufa-Lahlah, Fouzi Semchedine, Louiza Bouallouche-Medjkoune, Nadir Farhi
Abstract: In this paper, we propose the Proactive-optimal-path Selection with Coordinator Agents Assisted Routing (PSCAR) protocol for Vehicular Ad hoc NETworks (VANETs) in an urban environment. The main idea of PSCAR is to contribute static nodes as coordinator agents placed at each intersection, in order to improve the routing performance and to deal with radio obstacles (buildings, trees...) and voids as encountered in urban environments. Since the coordinator agents are static nodes, each one knows all the paths to any other coordinator agent in the network. Thus, instead of searching an optimal path toward the destination node, PSCAR will determine an optimal path to the nearest coordinator agent to the destination node so as to better anticipate any change of the destinations position. The optimal path is selected according to two criteria: the total physical distance and the vehicle density on the path. The vehicle density is estimated based on a fundamental diagram of the traffic that allows estimating the vehicular traffic density on each road segment. To evaluate the performance of PSCAR we used the Network Simulator 2 (ns-2) and the mobility simulator SUMO. We compare our scheme with some existing solutions with the aim of showing its effectiveness, in terms of packet delivery ratio, end-to-end delay, and network overhead.
Keywords: vehicular ad hoc networks; position-based routing; greedy forwarding; carry and forward; radio obstacles; urban environments; traffic density; ns-2; SUMO.
EAHKM+: energy-aware secure clustering scheme in wireless sensor networks
by Mohamed-Lamine Messai, Hamida Seba
Abstract: The effective technique to save energy in Wireless Sensor Networks (WSNs) is clustering. However, organising WSNs into clusters securely is a challenging task in regard to the vulnerabilities of these networks. In this paper, we provide a secured cluster formation by proposing a new symmetric key management scheme for hierarchical WSNs (HWSNs). The new scheme is called EAHKM+ (Energy Aware Hierarchical Key Management in WSNs). Only three keys are pre-distributed in each sensor node before deployment. EAHKM+ ensures the establishment of a pairwise key between each sensor node and its cluster head, thus the establishment of a broadcast key in each cluster in the network. Through comparison with other proposed key management schemes in hierarchical WSNs, we show that EAHKM+ is an energy-efficient, exible, robust and scalable solution to the key management problem in clustered WSNs.
Keywords: secure clustering; hierarchical wireless sensor networks; key management; energy saving
RPSE: reenactment of data using polynomial-variant in sensor
by Ambika Nagaraj
Abstract: Sensor environment relieves the hassle by monitoring the environment under study and providing results to the investigator in time. The users need be concerned only about the whereabouts of the installations of these tiny elements in the environment under study. Such networks without any management and control can be intruded by adversaries. As a precautionary measure, encryption can be adopted. These encrypted messages are self-insured from being exposed unless the intruders have acquired the decryption keys. In this study, a polynomial-variant is generated and used to encrypt the transmitted data. The variant is calculated using various properties of the nodes deployed. Adopting this protocol helps the base station to originate the manipulated data. The study preserves backward and forward secrecy. The work minimises wormhole attack and sinkhole attack in the network.
Keywords: prevention measure, wireless sensor network, polynomial-variant, wormhole attack, reconstruction of data, key generation, location-based keys, sinkhole attack.
Compact UWB BPF with notch-band using SIR and DGS
by Hassiba Louazene, Mouloud Challal, M’hamed Boulakroune
Abstract: This paper presents a novel ultra-wideband (UWB) bandpass filter (BPF) with a notch-band at specified frequency. Moreover, its equivalent circuit model (ECM) is also investigated. The proposed filter consists of a stepped impedance resonator (SIR), a parallel coupled feed-line on the top and a rectangular shaped defected ground structure (DGS) on the bottom of the structure. The notch-band can be shifted to any other desired frequency by tuning the length of the parallel coupled feed-line. Good performances in terms of wide stop-band rejection, low insertion loss, high return loss and more compact size (7.99 x 6 mm2) than those reported in literature are achieved. In addition, simulation results accomplished by ECM are in good agreement with the full-wave EM.
Keywords: ultra-wideband; bandpass filter; stepped rnimpedance resonator; defected ground structure; notch-band.
Special Issue on: Advances in Parallel Methods for Scientific Computing
Parallel solution of the discretised and linearised G-heat equation
by Pierre Spiteri, Amar Ouaoua, Ming Chau, Hacene Boutabia
Abstract: The present study deals with the numerical solution of the G-heat equation. Since the G-heat equation is defined in an unbounded domain, we firstly state that the solution of the G-heat equation defined in a bounded domain converges to the solution of the G-heat equation when the measure of the domain tends to infinity. Moreover, after time discretisation by an implicit time marching scheme, we define a method of linearisation of each stationary problem, which leads to the solution of a large scale algebraic system. Then we analyse in a unified approach the convergence of the sequential and parallel relaxation methods. Finally we present the results of numerical experiments.
Keywords: G-heat equation; relaxation methods; parallel computing; asynchronous iteration; financial application.
Special Issue on: Advances in Information Security and Networks
Full secure identity-based encryption scheme over lattices for wireless sensor networks in the standard model
by Jizhong Wang, Chunxiao Wang
Abstract: This paper proposes a lattice-based identity-based encryption scheme that can be used in wireless sensor networks to achieve secure communication. More precisely, a known lattice-based chosen plain-text secure encryption scheme and the Bonsai trees primitive are used to design a identity-based encryption (IBE) scheme. To reduce the public key size of the proposed IBE scheme, a public matrices chosen rule is used that reduces the public key size of proposed scheme to be close to half of that of a known lattice-based scheme. Under the hardness of the decision variant of the learning with errors (LWE) problem, we prove that the proposed IBE scheme is indistinguishable against the adaptive chosen identities and chosen-plaintext attack in the standard model. Moreover, the message-to-ciphtertext expanse factor of this scheme is also controlled efficiently, which is close to that of Gentrys scheme. Owing to the quantum intractability of the LWE problem on which the scheme is based, the proposed IBE scheme is secure even in the quantum era.
Keywords: wireless sensor networks, lattice-based cryptography, learning with errors problem, standard model
The fault diagnosis method of RVM based on FOA and improved multi-class classification algorithm
by Kun Wu, Jianshe Kang, Kuo Chi, Xuan Wang
Abstract: In order to solve the current problem of fault diagnosis and improve the traditional diagnosis models, an intelligent fault diagnosis approach of the relevance vector machine (RVM) based on the fruit fly optimisation algorithm (FOA) and improved multi-class classification algorithm is proposed. The optimal parameter values of the RVM kernel function are determined by the FOA, and an improved multi-class classification algorithm of the RVM, based on the traditional one-against-one (OAO) and one-against-rest (OAR) algorithm, is presented. The above classification method translates the multi-class classification problem into multiple three-classification problems to accelerate the running speed with high classification accuracy. Theoretical analysis and experiment results both demonstrate that the proposed method performs better than traditional methods in terms of diagnosis accuracy and running time, with more model sparsity and higher diagnosis efficiency.
Keywords: fault diagnosis; relevance vector machine; fruit fly optimisation algorithm; multi-class classification
Secure data outsourcing scheme in cloud computing with attribute-based encryption
by Shuaishuai Zhu, Yiliang Han
Abstract: In our IT society, cloud computing is clearly becoming one of the dominating infrastructures for enterprises as well as end users. As more cloud-based services become available to end users, their oceans of data are outsourced in the cloud as well. Without any special mechanisms, the data may be leaked to a third party for unauthorised use. Most presented works on cloud computing emphasise computing utility or new types of application. But in the view of cloud users, such as traditional big companies, data in cloud computing systems tends to be out of control and privacy fragile. So most of data they outsource is less important. A mechanism to guarantee the ownership of data is required. In this paper, we analyse a couple of recently presented scalable data management models to describe the storage patterns of data in cloud computing systems. Then we define a new tree-based data set management model to solve the storage and sharing problems in cloud computing. A couple of operation strategies including data encryption, data boundary maintenance, and data proof are extracted from the view of different entities in the cloud. The behaviour of different users is controlled by view management on the tree. Based on these strategies, a flexible data management mechanism is designed in the model to guarantee entity privacy, data availability and secure data sharing.
Keywords: cloud computing; outsourcing data; data privacy; database management
An optimisation research on two-dimensional age-replacement interval of two-dimensional product
by Yutao Wu, Wenyuan Song, Yongsheng Bai, Yucheng Han, Wenbin Cao
Abstract: For the failures of two-dimensional product affected by the calendar time and usage time, traditional one-dimensional preventive maintenance cant meet the actual maintenance demand of two-dimensional product, therefore, this paper proposes a two-dimensional preventive maintenance strategy for two-dimensional product, and makes a decision for two-dimensional product through calendar time and usage time. Firstly, we establish failure rate function expression of two-dimensional product and analyse the process of two-dimensional preventive maintenance in detail, then propose a cost model and availability model from the view of economy and mission in finite span. According to the cost and availability model, through examples, we derive the optimal two-dimensional age-replacement interval by numerical algorithm and particle swarm optimisation, and verify the models applicability and validity through contrast and analysis. Lastly, we give a brief discussion of future research area about two-dimensional maintenance.
Keywords: two-dimensional product; two-dimensional age-replacement; math model; maintenance interval
Extensional schemes of multipartite non-interactive key exchange from multilinear maps and their applications
by Huiwen Jia, Yupu Hu
Abstract: The question of generalising the celebrated two-party non-interactive key exchange (NIKE), Diffie-Hellman protocol, to a multipartite setting was left as an important open problem. In 2003, Boneh and Silverberg put forward a theoretical construction of multipartite NIKE protocol from a new notion called multilinear maps. In their protocol, however, the number of users N and the multilinearity k are related by N = k+1, resulting in the system initialising another multilinear map when the number of users who want to exchange a session key changes. In this paper, we describe two extensional schemes of multipartite NIKE, which enables any less than or equal to N users derive a common shared key from an (N-1)-multilinear map. In addition, using our extensional schemes, we show a concrete scenario: The establishment of any discussion group in a user group and its privacy version. Furthermore, we analyse its security.
Keywords: multipartite non-interactive key exchange; multilinear maps; MCDH assumption
Dynamic combined with static analysis for mining network protocols' hidden behaviour
by YanJing Hu
Abstract: Unknown protocols' hidden behaviour is becoming a new challenge in network security. This paper takes both the captured messages and the binary code that implement the protocol as the studied objects. Dynamic Taint Analysis combined with Static Analysis is used for protocol analysing. Firstly, we monitor and analyse the process of protocol program that parses the message in the virtual platform HiddenDisc prototype system developed by ourselves, and record the protocols public behaviour, then based on our proposed hidden behaviour perception and mining algorithm, we perform static analysis of the protocols hidden behaviour trigger conditions and hidden behaviour instruction sequences. According to the hidden behaviour trigger conditions, new protocol messages with the sensitive information are generated, and the hidden behaviours are executed by dynamic triggering. HiddenDisc prototype system can sense, trigger and analyse the protocols hidden behaviour. According to the statistical analysis results, we propose the evaluation method of protocol execution security. The experimental results show that the present method can accurately mining the protocols hidden behaviour, and can evaluate an unknown protocols execution security.
Keywords: protocol reverse analysis; protocols' hidden behaviour; protocol message; protocol software
Special Issue on: Algorithm and Technology Advances for Future Internet
UOSM: user-oriented semi-automatic method of constructing domain ontology
by Chao Qu, Fagui Liu, Dacheng Deng, Ruifen Yuan
Abstract: Based on the analysis of the existing main ontology construction methods and tools, we propose a user-oriented method of semi-automatic domain ontology construction. The method builds a descriptor set according to the user's requirements, and establishes a hi-erarchy structure based on it. Users can construct ontologies by using this method cases without the participation of experts. Additionally, the method has advantages for expansion and the collaborative development of ontologies.
Keywords: Domain Ontology; Ontology Construction; Semi-automatic; User Oriented
Cloud workflow scheduling algorithm based on reinforcement learning
by Delong Cui, Zhiping Peng, Wende Ke, Jinglong Zuo
Abstract: How to fairly schedule the multiple workflow with multiple priorities submitted at different time has become an increasing concern in Workflow Management Systems (WMS). To solve the problem, a novel workflow scheduling algorithm based on reinforcement learning is proposed in this study. In our scheme, we first define some basic concepts of reinforcement learning in cloud computing, such as state space, action space and immediate reward. Then single DAG and multiple DAG cloud workflow scheduling algorithm based on reinforcement learning are designed respectively. Reinforcement learning sets up a policy to maximise the cumulative rewards in the long term through the repetition of trial-and-error interactions in cloud computing environment. Finally, we analyse algorithm performance by using queuing theory. We use real cloud workflow to test the proposed scheme. Our results, on the one hand, demonstrate the proposed scheme can reasonably schedule multiple DAGs with multiple priorities and improve use rate of resources better and, on the other hand, show optimisation object function achieves fair workflow scheduling in cloud computing environment.
Keywords: multiple DAGs; reinforcement learning; workflow scheduling; cloud computing
Quick convergence algorithm of ACO based on convergence grads expectation
by Zhongming Yang, Yong Qin, Yunfu Jia
Abstract: While the ACO (ant colony optimisation) can find the optimal path of a network, there are too many iterative times and the convergence speed is also very slow. This paper proposes the Q-ACO QoSR based on convergence expectation to meet the requirement of OoS routing for real-time and highly efficient networks. This algorithm defines index expectation function of a link, and proposes convergence expectation and convergence grads. As for the multi-constrain QoS routing model, the algorithm controls the iteration and searches for the optimal path that meets the QoS restriction condition under the condition of higher convergence. This algorithm can find the optimal path by comparing the convergence grads in a faster and larger probability. This algorithm improves the ability of routing and convergence speed. This rapid path-finding algorithm can be applied not only to routing algorithm but also to optimisations, such as solving the problem of task scheduling and load balancing (17-18) by using ant colony algorithm in the field of cloud computing. Besides, during the process of task scheduling, the function that forms into the expectation function is decided by constraint conditions, such as the speed of execution, scheduling time and cost calculation. It realises the optimal solution to algorithm convergence through controlling the number of iterations.
Keywords: ACO; QoS; QoSR; CG expectation; Q-ACO
Short-term vegetable prices forecast based on improved gene expression programming
by Lei Yang, Kangshun Li, Wensheng Zhang, Yaolang Kong
Abstract: Gene expression programming (GEP) is an evolutionary algorithm based on genotype and phenotype, which is commonly used in network optimisation and prediction. Aiming at the problem that the traditional (GEP) is susceptible to noise interference, leading to premature convergence and falling into local solutions, this paper proposes an improved GEP algorithm, which increases 'inverted series' and 'extraction' operator. The improved algorithm can effectively increase the rate of use of genes, with convergence speed and solution precision becoming higher, and can avoid the premature phenomenon. Taking the price change trends of Chinese vegetables mooli, scallion, white gourd, eggplant, green pepper and potato as examples, this paper discusses the way to solve the forecasting problems by adopting gene expression programming, by constructing time interval unified time series data and normalised processing. Through the training and construction model, it realises the simulation and forecasting of price trends. The experimental results show that the improved GEP algorithm has fast calculation speed and high forecasting precision.
Keywords: gene expression programming; improved gene expression programming; short-term vegetable prices forecast; use of genes; gene extraction
High performance energy Saving policy based on IND
by Lin Liu, JingGuo Dai
Abstract: The IND (Intelligent Network Disk) storage solution has the advantages of intelligence, low cost, and the ability to be widely used in mass storage and portable devices; however, its energy cost is a key problem. Traditional energy saving policy leads to performance delays because of frequent changes in the component status. This research focuses on the performance loss resulting from RAM due to current energy saving policy and designs an optimised IND RAM energy management policy. In our method, a replica space is established in the memory of NIC and a copy of the high frequent read-only data is put into replica space, thereby reducing the frequency of status changes in INDs memory bank during the I/O process, and extending the duration of low-cost status of IND memory and BUS. Thus, more energy is saved. Meanwhile, a copy of high frequent read-only data in replica space can reduce the frequent read-only data transfer via BUS and the time of BUSs data transfer, which can improve the system performance. The experimental results show that the optimised new policy can save more than 20% of memorys energy compared with traditional IND energy saving policy, and could improve the I/O performance of the IND system by more than 10%. Additionally, this strategy can be extended to similar structures with good theoretical and practical value.
Keywords: IND; energy saving; high performance; dynamic power management
A privacy-preserving and fine-grained access control scheme in DaaS based on efficient DSP re-encryption
by Jin Yang
Abstract: Database as a service (DaaS) is an important cloud service, and preserving privacy is one of the most crucial needs in DaaS. To achieve more efficient and safer preservation of privacy in DaaS, a new access control scheme is proposed. Firstly, some definitions of data privacy value are introduced, which are used to formulate a data table. Then, an efficient DSP re-encryption mechanism is constructed based on the proposed data privacy value. Furthermore, the traditional DaaS service framework is improved to achieve more flexible and fine-grained access control. Finally, by combining the proposed mechanism with access control policy, our new access control scheme is carefully designed. Moreover, the correctness and security are analysed, which show that the new scheme is correct and IND-CCA is secure under random oracle mode. That is, it can prevent conspiracy attacks effectively and protect the privacy of data owner and data user.
Keywords: database as a service, proxy re-encryption, privacy preserving, access control, ND-CCA security
A parallel immune genetic algorithm for community detection in complex networks
by Lu Xiong, Kangshun Li, Lei Yang
Abstract: This paper proposes a complex network community discovery method based on a parallel immune genetic algorithm for the problem of low efficiency, slow convergence rate and population degradation in community mining method based on a genetic algorithm. The algorithm makes use of the principle of parallel immune system to ensure the diversity of population, and enhance the searching ability by using a single path crossover operator in the initial population and crossover operation to enhance the ability to find the best use of the single path crossover operator, while using the improved character encoding and adaptive mutation operator to further reduce the search space and improve the population degradation phenomenon. Experiments show that the improved parallel immune genetic algorithm can be used to find the problem of complex network community with high accuracy, effectiveness and efficiency.
Keywords: genetic algorithm; community discovery; parallel immune; data mining
An automatic traffic-congestion detection method for bad weather based on traffic video
by Jieren Cheng
Abstract: In order to solve the problem that the result of automatic traffic-congestion detection method in bad weather is inaccurate, we analyse current vehicle identification algorithms and image processing algorithms. After that, we propose a detection method of traffic congestion based on histogram equalisation and discrete-frame difference. Firstly, this method uses discrete-frame difference algorithm to extract the images that have vehicle information. It can save resources and make the method achieve real-time performance. Then, the method employs histogram equalisation algorithm to eliminate the noise of the vehicle images. We also propose a calculation method for traffic congestion index of the detection method in the paper. After that, this method recognises vehicles from the video and computes the traffic congestion index. Finally, the method transforms the dimension of traffic congestion index and obtains the state of traffic congestion. It is proved by experiments and theoretical analysis that the method decreases false negative rate and increases the accuracy rate of automatic traffic congestion detection in bad weather. The method has a wonderful prospect in some areas where there is usually bad weather.
Keywords: traffic congestion; histogram equalization; video processing; inter-frame difference.
Constrained evolution algorithm based on adaptive differential evolution
by Kangshun Li, Liang Zhong, Lei Zuo, Zhaopeng Wang
Abstract: Solving constrained optimisation is widely used in science and engineering, but the slow convergence speed and premature convergence are the biggest problems researchers face. Research on a constrained evolution algorithm (CO-JADE) based on adaptive differential evolution (JADE) for solving the constrained optimisation problems is proposed in this paper. According to features of the Gaussian distribution, the Cauchy distribution and the mutation factor, we exploited the crossover probability of each individual to improve the search strategy. Aiming to effectively evaluate the relationship between the value of the objective function and the degree of constraint violation, the paper uses an improved adaptive tradeoff model to evaluate the individuals of a population. This tradeoff model uses different treatment schemes for different stages of the population and is implemented on night standard test functions. The experiments show that the CO-JADE has better accuracy and stability than the COEA/ODE and the HCOEA.
Keywords: adaptive differential evolution; adaptive tradeoff model; constrained optimization.
Special Issue on: CloudTech'2015 Cloud Computing Technologies and Applications
Swift personal emergency help facilitated by the mobile cloud
by Hazzaa Alshareef, Dan Grigoras
Abstract: Many emergency cases require the swiftest possible response from an appropriate medical service if they are not to become life-threatening. In the medical emergency field, response times to emergency cases are a major concern and gain a high degree of attention. Many of the systems proposed in the literature are either intended to replace the existing emergency system with a fully automated one, or build on unreliable or less efficient frameworks that are based on some sort of social media application, such as redirecting emergency requests to Facebook friends. We have designed a mobile cloud service that works side by side with the existing emergency system and is aimed at reducing the time spent waiting for emergency help to arrive, as well as making the best use of medical professionals who may be located in close proximity to the medical case. The experimental results show that the amount of time needed to find a medical professional and establish communication was between 4 and 25 seconds, depending on the communication method used. This result means that no extra time is added to the total medical response time and could enhance the chance of a better outcome.
Keywords: component; mobile; cloud; healthcare; medical emergency; help;
doctors; nurses; SMS; medical response time
Special Issue on: ICNC-FSKD 2015 Parallel Computing and Signal Processing
1.25 Gbits/s-message experimental transmission using chaos-based fibre-optic secure communications over 143 km
by Hongxi Yin, Qingchun Zhao, Dongjiao Xu, Xiaolei Chen, Ying Chang, Hehe Yue, Nan Zhao
Abstract: Chaotic optical secure communications (COSC) are a kind of fast-speed hardware encryption techniques at the physical layer. Concerning to the practical applications, high-speed long-haul message transmission is always the goal to pursue. In this paper, we reported experimentally a scheme of long-haul COSC, where the bit rate reaches 1.25 Gbits/s and the transmission distance up to 143 km. Besides, a distinct advantage of low-cost is guaranteed with the off-the-shelf optical components, and no dispersion compensating fiber (DCF) or forward-error correction (FEC) is required. To the best of our knowledge, this is the first experimental evidence of the longest transmission distance in the COSC system. Our results show that high-quality chaotic synchronization can be maintained both in time- and frequency-domain, even after 143 km transmission; the bandwidth of the transmitter is enlarged by the external optical injection, which leads to the realization of 2.5 Gbits/s-message secure transmission up to 25 km. In addition, the effects of device parameters on the COSC are discussed for supplementary details.
Keywords: long-haul, high-speed, chaotic optical secure communications, semiconductor laser
Optimisation of ANFIS using mine blast algorithm for predicting strength of Malaysian small and medium enterprises
by Kashif Hussain, Mohd. Najib Mohd. Salleh, Abdul Mutalib
Abstract: Adaptive Neuro-Fuzzy Inference System (ANFIS) is a popular fuzzy inference system, as it is widely applied in business and economics. Many have trained ANFIS parameters using metaheuristic algorithms, but very few have tried optimising its rule-base. The auto-generated rules, using grid partitioning, comprise both the potential and the weak rules, increasing the complexity of ANFIS architecture as well as computational cost. Therefore, pruning less or non-contributing rules would optimise the rule-base. However, reducing the complexity and increasing the accuracy of the ANFIS network needs an effective training and optimisation mechanism. This paper proposes an efficient technique for optimising the ANFIS rule-base without compromising on accuracy. A newly developed Mine Blast Algorithm (MBA) is used to optimise ANFIS. The ANFIS optimised by MBA is employed to predict the strength of Malaysian small and medium enterprises (SMEs). Results prove that the MBA optimised ANFIS rule-base and trained parameters are more efficient than Genetic Algorithm (GA) and Particle Swarm Optimisation (PSO).
Keywords: ANFIS; neuro-fuzzy; fuzzy system; mine blast algorithm; rule optimisation; SME
Spectrum prediction and aggregation strategy in multi-user cooperative relay networks
by Yifei Wei, Qiao Li, Xia Gong, Da Guo, Yong Zhang
Abstract: In order to meet the constantly increasing demand by mobile terminals for higher data rate with limited wireless spectrum resources, cooperative relay and spectrum aggregation technologies have attracted much attention owing to their capacity in improving spectrum efficiency. Combining cooperative relay and spectrum aggregation technologies, in this paper, we propose a spectrum aggregation strategy based on the Markov prediction of the state of spectrum for the cooperatively relay networks on a multi-user and multi-relay scenario aiming at ensuring the user channel capacity and maximising the network throughput. The spectrum aggregation strategy is executed through two steps. First, the state of the spectrum is predicted through Markov prediction. Based on the prediction results of state of spectrum, a spectrum aggregation strategy is proposed. Simulation results show that the spectrum prediction process can observably lower the outage rate, and the spectrum aggregation strategy can greatly improve the network throughput.
Keywords: Markov model; spectrum aggregation; multi-user; cooperative relay; outage probability; network throughput
An optimised multicast data transmission scheme in IEEE 802.16j wireless relay networks
by Hanwu Wang
Abstract: Relay-enabled wireless networks play a significant role in data transmission, as it not only extends the BS coverage range but also improves the transmission efficiency of those MSs with poor channel conditions. Multicast transmission is an efficient transmission mode and quite suitable for relay-based networks. However, how to perform the desired multicast transmission in a wireless relay network has not been specified. To realise the effective multicast-based data delivery over the whole relay network, we propose an optimised transmission scheme by formulating an objective problem to achieve the desired performance goals. Specifically, an efficient multicast data transmission mechanism with respect to two schedule hops is proposed in detail to implement the optimisation process. Moreover, we also incorporate the spatial reuse into the transmission mechanism to improve the scheduling efficiency. The simulation results verify the efficacy and efficiency of the proposed transmission scheme.
Keywords: relay network; multicast; spatial reuse; optimisation
Special Issue on: ICICS 2016 Next Generation Cloud, Mobile Cloud, Mobile Edge Computing and Internet of Things Systems and Networking
Classifying environmental monitoring data to improve wireless sensor network management
by Emad Alsukhni, Shayma Almallahi
Abstract: Wireless sensor networks are considered as the most useful way for collecting data and monitoring the environment. Owing to the large amount of data that are produced from these networks, data-mining techniques are required to get interesting knowledge. This paper presents the effectiveness of using data-mining techniques to discover knowledge that can improve the management of wireless sensor networks in environmental monitoring. Data reduction in a network increases the networks lifetime. The classification model can predict the effect of sensed data, which is used to reduce the number of readings that are reported to the sink, in order to improve wireless sensor network management. In this paper, we demonstrate the efficiency and accuracy of using data-mining classifiers in predicting the effect of sensed data. The results show that the accuracy of the J48 classification model, Multilayer Perceptron and REPTree classifiers reached 90%. Using the classification model, the results show that the number of reported readings decreased by 37%. Hence, this significant reduction increases the wireless sensor network's lifetime by reducing the consumed energy, i.e., the total energy dissipated.
Keywords: classifying environmental monitoring data; data mining; wireless sensor network management; data reduction; energy consumption
A power-controlled interference management mechanism for femtocell-based networks
by Haythem Bany Salameh, Rawan Shabbar, Ahmad Al-shamali
Abstract: Femtocell deployment is one of the promising solutions to meet the increasing demands on wireless services and applications. Based on deploying femtocells within the already-existing macrocells, femtocells can improve indoor coverage and increase both spectral efficiency and data rate. This improvement can be achieved by reusing the available spectrum assigned for the macrocell, and being closer to the user. However, femtocell deployment faces many challenges. One of the most challenging issues is the interference management issue. In this paper, a power-controlled channel assignment algorithm to manage the interference between femtocell and macrocell networks is proposed. The proposed algorithm allows frequency reuse among femtocell users to provide better throughput performance. Simulation results have shown an improvement in throughput by up to 90% compared with previous models.
Keywords: femtocell; macrocell.
Exploring the relationships between web accessibility, web traffic, and university rankings: a case study of Jordanian universities
by Mohammed Al-Kabi
Abstract: Most statistics state that disabled people constitute approximately 20% of the world population; therefore, webmasters should consider this part of the world population when they design and implement their websites by following the Web Content Accessibility Guidelines. This paper aims to explore the relationships between web accessibility, web traffic, and university rankings using the metrics of 27 Jordanian university websites as a case study. Objective evaluations are conducted using a number of online tools. An extensive analysis of the tool outputs is presented, and the relationships of these tool outputs are explored. This study represents a longitudinal overview of the accessibility of Jordanian university websites and re-explores the different relationships between web accessibility, web traffic, and university rankings, as well as the differences between the accessibility of private and public Jordanian universities. This paper explores the relationships between three parameters: accessibility, web traffic, and university rankings. Therefore, its scope is larger than that of earlier publications.
Keywords: university websites; accessibility guidelines; WCAG 1.0 and WCAG 2.0; web accessibility; web traffic; website popularity; university rankings; automated evaluation.
Context-aware latency reduction protocol for secure encryption and decryption
by Muder Almiani, Abdul Razaque, Toufik Aidja, Ayman Al-Dmour
Abstract: Security is one of the biggest challenges for secure data communication. As the need of mobile phones increased, people started using several mobile applications to handle the internet of things. On the other hand, the frequent use of mobile phones in business caught the attention of researchers to protect the manufacturers and customers. Handling these issues, several paradigms for encrypting and decrypting the data-outsourcing have been proposed. However, there are still new threats and challenges for researchers owing to launch of new attacking methods and malicious actions of adversaries. Though the manufacturers have implemented an industry standard to protect the customers' privacy, launched standards do not provide a comfortable zone for the customers, particularly reducing the latency of the emerging applications being used in the internet of things. In this paper, we introduce a context-aware security (ConSec) protocol to support the internet of things applications to reduce the latency while encrypting and decrypting the applications. Furthermore, elliptic curve cryptography is used to fully secure the encryption and decryption processes. The proposed method is implemented using the Java platform, and results are verified and compared with full disk method from the latency perspective.
Keywords: security; elliptic curve cryptography; mobile platform; encryption; decryption; context aware system.
Special Issue on: AINA'2016 Cloud Computing Projects and Initiatives
Transaction management across data stores
by Marta Patino
Abstract: Companies have evolved from a world where they only had SQL databases to a world where they use different kinds of data stores, such as key-value data stores, document-oriented data stores and graph databases. The reason why they have started to introduce this diversity of persistency models is because different NoSQL technologies bring different data models with associated query languages and/or APIs. However, they are confronted now with a problem in which they have the data scattered across different data stores. This problem lies in that when a business action requires to update the data, the data reside in different data stores, and they are subject to inconsistencies in the event of failure and/or concurrent access. These inconsistencies appear due to the lack of transactional consistency that was guaranteed in traditional SQL databases but is not guaranteed either within the NoSQL data stores or across data stores and databases. CoherentPaaS comes to remedy this need. CoherentPaaS provides an ultra-scalable transactional management layer that can be integrated with any data store with multi-versioning capabilities. The layer has been integrated with six different data stores, three NoSQL data stores and three SQL-like databases. In this paper, we describe this generic ultra-scalable transactional management layer and focus on its API and how it can be integrated in different ways with different data stores and databases.
Keywords: data management; NoSQL; consistency.
Enabling and monitoring platform for cloud-based application
by Silviu Panica, Bogdan Irimie, Dana Petcu
Abstract: The deployment of applications able to consume cloud infrastructure services
is currently a tedious process, especially when it should be often repeated in the application development phase. Even further, the deployed applications need to be monitored in order to detect and take action against possible anomalies.
We present a technical solution that simplifies the deployment process, as being almost fully automated and capable of offering monitoring services for the deployed applications. Its integration as a module of an open-source platform for enforcing security controls by its users is further discussed.
Keywords: cloud computing; enabling platform; automatic deployment; distributed monitoring.
Exploring the complete data path for data interoperability in cyber-physical systems
by Athanasios Kiourtis, Argyro Mavrogiorgou, Dimosthenis Kyriazis, Ilias Maglogiannis, Marinos Themistocleous
Abstract: The amount of digital information increases tenfold every year, owing to the exponential increase of Cyber-Physical Systems (CPS), real and virtual internet-connected sources. Most researches are focused on data processing and inter-connection fields, leading to the question concerning the interoperable use of data: if data is efficiently processed, how can unknown data be used in a different natures application? A three-stepped approach is presented in this paper, addressing this question, where following the data-lifecycle, a known CPSs dataset is firstly stored into domain-specific language, then translated into domain-agnostic language, and finally, using the fitting function of an ANN, it is compared with an unknown dataset, resulting in the translation of the unknown dataset into the first datasets domain. A scenario of that approach is provided, analysing the data interoperability challenges and needs, emerging from todays Internet of Everything evolution, studying the fields of data annotation, semantics, modelling, and characterisation.
Keywords: cyber-physical systems; semantic; interoperability; domain agnostic; domain specific; data path; data lifecycle;.
Monitoring and management of a cloud application within a federation of cloud providers
by Rocco Aversa, Luca Tasquier
Abstract: Cloud federation is an emerging computing model where multiple resources from independent cloud providers are leveraged to create large-scale distributed virtual computing clusters, operating as within a single cloud organisation. This concept of service aggregation is characterised by interoperability features, which can address different problems about inter-cloud collaboration, such as vendor lock-in. Furthermore, it approaches challenges like performance and disaster-recovery through methods such as co-location and geographic distribution. One of the main issues within a cloud federation is related to the monitoring of the application deployed on resources coming from different vendors belonging to the federation. In this work we present an agent-based architecture and its prototypal implementation that aims at monitoring the user's cloud environment provided by the federation: the elasticity of the proposed architecture allows the configuration and customisation of the monitoring infrastructure to adapt it to the specific cloud application. A multi-layer architecture is proposed, where each part monitors different aspects of the multi-cloud infrastructure, starting from the detection of critical conditions on low level parameters for the computational units and composing different monitoring levels in order to check the federated SLA. The agent-based approach will introduce fault-tolerance and scalability to the monitoring architecture, while the agents' reactivity and proactivity capabilities will allow a deep and intelligent monitoring, where each agent can focus on different aspects of the monitoring activity, from low level performance indexes to the checking of the federated SLA compliance. Agents will be strengthened by algorithms and rules used to monitor QoS parameters that are critical for the specific application; the configuration of the adaptive monitoring environment will be made easier by an interface that will help the user in describing his/her application's deployment. The prototypal implementation of the proposed framework will be applied on a testbed application to validate the monitoring architecture.
Keywords: cloud federation; agent-based monitoring; multi-cloud environment.
New architecture for virtual appliance deployment in the cloud
by Amel Haji, Asma Ben Letaifa, Sami Tabbane
Abstract: Cloud computing is a model that offers a convenient access to a set of configurable resources which could be provisioned and released with less of administration.
The main lock is to assume interoperability and portablity between cloud services.
We deal with special type of services called "Virtual Appliance". These services offers pre-packaged applications, preconfigured and encapsulated in virtual machine images running on virtualisation platforms. This type of service enables flexible and easy deployment, while ensuring interoperability, reuse and migration of services.
This paper aims to offer architecture to deploy "Virtual Appliance" in cloud environment.
Keywords: cloud computing; virtual appliance; SDN; NFV; OpenStack.
From business process models to the cloud: a semantic approach
by Beniamino Di Martino, Antonio Esposito, Giuseppina Cretella
Abstract: The fast evolution of IT services and the shift from server-centred applications to widely distributed frameworks and platforms, of which cloud computing is the best representative, have highly conditioned the way enterprise managers deal with their businesses. Several tools for the design and execution of business processes have been proposed, together with standards for their formal representation, such as BPMN. Despite the expressivity of such tools and representations, there is still a lack of integration with cloud services, mostly owing to the high variety of currently available offers, which create confusion among customers. Moreover, portability and interoperability issues tend to hinder the adoption of cloud solutions for the immediate deployment of Business processes. In order to overcome such issues and support the adoption of cloud services for the deployment of business processes, in this paper a semantic approach is proposed. The presented representation aims at easing the mapping process and to suggest users the best suitable Cloud services to compose. A case study that demonstrates the applicability and efficacy of the approach is also described.
Keywords: business process; semantics; BPMN; OWL-S; cloud computing; ontologies; modelling; cloud services.
Migrating mission-critical application in federated cloud: a case study
by Alba Amato, Rocco Aversa, Massimo Ficco, Salvatore Ventcinque
Abstract: Although, virtualisation, elasticity and resource sharing enable new levels of flexibility, convenience and economy benefits, they also add new challenges and more areas for potential failures and security vulnerabilities, which represent major concerns for companies and public organisations that want to shift their business
and mission-critical applications and sensitive-data to the cloud. This paper discusses some technical issues that must be addressed to migrate mission-critical applications on public clouds. Moreover, by using a case study, an approach to broker cloud infrastructure needs to satisfy more restrictive critical requirements is presented.
Keywords: mission-critical applications; cloud federation; brokering; security; dependability.
Intelligent medical record management: a diagnosis support system
by Flora Amato, Giovanni Cozzolino, Antonino Mazzeo, Sara Romano
Abstract: The increasing life expectation and low birth rates have radically changed the demographic structure of the European Union. The number of elderly people is growing as is the number of chronic diseases: this implies higher and higher healthcare costs, a reduction of the number of healthcare personnel, and requests for better care services. E-health has led to a growth of the health organisation, providing innovative and non-intrusive systems together with value-added services to the healthcare actors that contribute to enhance the efficiency and reduce the costs of complex informative systems. Many research efforts lead to innovative and non-intrusive e-health systems. In this work, we present a system for supporting medical decisions. It is based on semantic analysis of available medical data. The system implements an innovative methodology, which combines different semantic approaches in order to extract the representation of a given document expressed in natural language, and to associate it to a set of RDF triples.
Keywords: e-health; knowledge management; data integration; natural language processing; decision support system.
Semi-supervised PSO clustering algorithm based on self-adaptive parameter optimization
by Xiuqin Pan, Wenmin Zhou, Yong Lu, Dongyin Sun
Abstract: Particle swarm optimisation (PSO) based on semi-supervised learning (SSPSO) is known for its higher clustering accuracy than other classical clustering algorithms. However, a fixed parameter, representing the use ratio of labelled sample and unlabelled sample for the clustering, is selected. Consequently, the determination method of this parameter means that it is difficult to reach the best clustering result. In this paper, we propose an improved clustering algorithm to solve the parameter optimisation problem for PSO based on semi-supervised learning. The new approach is called APO_SSPSO, which employs an adaptive strategy based on PSO to dynamically adjust the usage ratio of labelled and unlabelled samples for the clustering. Experiments are conducted on two sets of test samples. Simulation results show that the proposed algorithm is effective and valid.
Keywords: particle swarm optimisation; clustering; semi-supervised; self-adaptive; parameter optimisation.
Special Issue on: WACCPD High-level Programming Approaches for Accelerators
Acceleration of unstructured implicit low-order finite-element earthquake simulation using OpenACC on Pascal GPUs
by Takuma Yamaguchi, Kohei Fujita, Tsuyoshi Ichimura, Muneo Hori, Lalith Maddegedara
Abstract: We accelerate CPU-based unstructured implicit low-order finite-element simulations by porting to a GPU-CPU heterogeneous compute environment by OpenACC. We modified the algorithm of performance-sensitive parts, such as sparse matrix-vector multiplication and MPI communication, so that computations are suitable for GPUs. Other parts of the earthquake simulation code are ported by directly inserting OpenACC directives into the CPU code. This porting approach enables high performance with relatively low development costs. When comparing eight K computer nodes and eight NVIDIA Pascal P100 GPUs, we achieve 20.8 times speedup for the 3-by-3 block Jacobi preconditioned conjugate gradient finite-element solver. We show the effectiveness of the proposed method through many-case crust-deformation simulations and a large-scale computation using a finite element model with billion degrees-of-freedom on a GPU cluster.
Keywords: MPI; element-by-element; conjugate gradient method.
PACC: a directive-based programming framework for out-of-core stencil computation on accelerators
by Nobuhiro Miki, Fumihiko Ino, Kenichi Hagihara
Abstract: We present a directive-based programming framework, i.e., the pipelined accelerator (PACC), to accelerate large-scale stencil computation on an accelerator device, such as a graphics processing unit (GPU). PACC provides a collection of extended OpenACC directives to facilitate out-of-core stencil computation accelerated using temporal blocking. The proposed framework includes a source-to-source translator capable of generating out-of-core OpenACC code from PACC code, i.e., large data is automatically decomposed into smaller chunks that are processed using limited capacity device memory. The generated code is optimized using a temporal blocking technique to minimise CPU-GPU data transfer. Furthermore, code is accelerated using a multithreaded pipeline engine that maximizes data copy throughput and overlaps GPU execution and data transfer. In experiments, we applied the proposed translator to three stencil computation codes. The out-of-core performance for 107 GB data on an NVIDIA Tesla K40 GPU with 12 GB memory reached 69.3 GFLOPS, which is 17% less than the in-core performance for 8 GB data. We believe that the proposed directive-based approach can be used to facilitate out-of-core stencil computation on a GPU.
Keywords: accelerator; directive-based programming; out-of-core execution; OpenACC; GPU.
Efficient implementation of OpenACC cache directive on NVIDIA GPUs
by Ahmad Lashgar, Amirali Baniasadi
Abstract: OpenACC's programming model presents a simple interface to programmers, offering a trade-off between performance and development effort. OpenACC relies on compiler technologies to generate efficient code and optimise for performance. The cache directive is among the challenging to implement directives. The cache directive allows the programmer to use the accelerator's hardware- or software-managed caches by passing hints to the compiler. In this paper, we investigate the implementation aspect of cache directive under NVIDIA-like GPUs and propose optimisations for the CUDA backend. We use CUDA's shared memory as the software-managed cache space. We first show that a straightforward implementation can be very inefficient, and undesirably downgrade performance. We investigate the differences between this implementation and hand-written CUDA alternatives and introduce the following optimisations to bridge the performance gap between the two: i) improving occupancy by sharing the cache among several parallel threads and ii) optimising cache fetch and write routines via parallelisation and minimising control flow. Investigating three test cases, we show that the best cache directive implementation can perform very close to hand-written CUDA equivalent and improve performance up to 2.4X (compared with the baseline OpenACC.)
Keywords: OpenACC; cache memory; CUDA; software-managed cache; performance.
Performance evaluation of OpenMPs target construct on GPUs
by Akihiro Hayashi, Jun Shirako, Ettore Tiotto, Robert Ho, Vivek Sarkar
Abstract: OpenMP is a directive-based shared memory parallel programming model and has been widely used for many years. From OpenMP 4.0 onwards, GPU platforms are supported by extending OpenMPs high-level parallel abstractions with accelerator programming. This extension allows programmers to write GPU programs in standard C/C++ or Fortran languages, without exposing too many details of GPU architectures. However, such high-level programming models generally impose additional program optimisations on compilers and runtime systems. Otherwise, OpenMP programs could be slower than fully hand-tuned and even naive implementations with low-level programming models such as CUDA. To study potential performance improvements by compiling and optimising high-level programs for GPU execution, in this paper, we (1) evaluate a set of OpenMP benchmarks on two NVIDIA Tesla GPUs (K80 and P100) and (2) conduct a comparable performance analysis among hand-written CUDA and automatically generated GPU programs by the IBM XL and clang/LLVM compilers.
Keywords: OpenMP; GPUs; performance evaluation; compilers; target constructs; LLVM; XL compiler; Kepler; Pascal.
Developments in memory management in OpenMP
by Jason Sewall, John Pennycook, Alejandro Duran, Christian Terboven, Xinmin Tian, Ravi Narayanaswamy
Abstract: Modern computers with multi-/many-core processors and accelerators feature a sophisticated and deep memory hierarchy, potentially including distinct main memory, high-bandwidth memory, texture memory and scratchpad memory. The performance characteristics of these memories are varied, and studies have demonstrated the importance of using them effectively. In this article, we explore some of the major issues in developing software to effectively and portably implement these technologies and describe enhancements being added to the OpenMP language to bridge this software-hardware gap. Our proposal separately exposes the characteristics of memory resources (such as kind) and the characteristics of allocations (such as alignment), and is fully compatible with existing OpenMP constructs.
Keywords: memory management; programming languages; compiler directives; heap allocation.
Using the loop chain abstraction to schedule across loops in existing code
by Ian Bertolacci, Michelle Strout, Stephen Guzik, Jordan Riley, Eddie Davis, Catherine Olschanowsky
Abstract: Exposing opportunities for parallelisation while explicitly managing data locality is the primary challenge to porting and optimising existing computational science simulation codes to improve performance and accuracy. OpenMP provides many mechanisms for expressing parallelism, but it primarily remains the programmer's responsibility to group computations to improve data locality. The Loop Chain abstraction, where a summary of data access patterns is included as pragmas associated with parallel loops, provides compilers with sufficient information to automate the parallelism versus data locality tradeoff. In this paper, we present the syntax and semantics of loop chain pragmas for indicating information about the loops that belong to the chain and orthogonal specification of a high-level schedule (i.e., transformation plan) for the whole loop chain. We show example usage of the pragmas, detail attempts to automate the transformation of a legacy scientific code written with specific language constraints to loop chain codes, describe the compiler implementation for loop chain pragmas, and exhibit performance results for a computational fluid dynamics benchmark.
Keywords: loop chain abstraction; loop optimisation; source-to-source compiler; performance optimisation.
Creating a portable, high-level graph analytics paradigm for compute and data-intensive applications
by Robert Searles, Stephen Herbein, Travis Johnston, Michela Taufer, Sunita Chandrasekaran
Abstract: High performance computing (HPC) offers tremendous potential to process large amounts of data, commonly referred to as big data. Owing to the immense computational requirements of big data applications, the HPC and big data communities are converging. As a result, heterogeneous and distributed systems are becoming commonplace. In order to take advantage of the immense computing power of these systems, distributing data efficiently and leveraging specialised hardware (e.g. accelerators) is critical. In this paper, we develop a portable, high-level paradigm that can be used to run big data applications on existing and future HPC systems. More specifically, we will target graph analytics applications, since these types of application are becoming increasingly popular in the big data and machine learning communities. Using our paradigm, we accelerate three real-world, compute- and data-intensive, graph analytics applications: a function call graph similarity application, a triangle enumeration subroutine, and a graph assaying application. Our paradigm uses the popular MapReduce framework, Apache Spark, in conjunction with CUDA, in order to simultaneously take advantage of automatic data distribution and specialised hardware present on each node of our HPC systems. We demonstrate scalability with regard to compute-intensive portions of the code that are parallelisable, as well as an exploration of the parameter space for each application. We show that our method yields a portable solution that can be used to leverage almost any legacy, current, or next-generation HPC or cloud-based system.
Keywords: graph analytics; GPU; distributed systems; high performance computing; cloud; heterogeneous systems; cluster; big data; Apache Spark; CUDA.
Special Issue on: IEEE TrustCom-16 Trust Computing and Communications
A trust-based evaluation model for data privacy protection in cloud computing
by Wang Yubiao, Wen Junhao, Zhou Wei
Abstract: For high quality and privacy protection problems, this paper proposes a trust-based evaluation model for data privacy protection in cloud computing (TEM-DPP). In order to make the final trust evaluation values more practical, the model introduces the comprehensive trust evaluation. The comprehensive trust is composed of direct trust and recommend trust. Services attribute and combining weights-based method are used to calculate the direct trust, reflecting the direct trust timeliness and rationality. In order to protect data security, we propose a data protection method based on a normal cloud model for data privacy protection. Then, the customer satisfaction, decay time, transaction amount and penalty factor will be used to update the direct trust. Simulation results showed that cloud services trust evaluation model can not only adapt to the dynamic changes in the environment, but also ensure the actual quality of service. It can improve the service requesters' satisfaction and has certain resilience to fraud entities.
Keywords: cloud service; trust; privacy-aware; evaluation model.
A novel algorithm for TOP-K optimal path on complex multiple attribute graph
by Kehong Zhang, Keqiu Li
Abstract: In the rapidly-changing information world, the various users and personalised requirements lead to an urgent need for complex multiple attribute decision-making. In addition, the optimal solution of a single attribute decision cannot meet the actual needs. The TOP-K optimal path is an effective way to solve the above problem. The TOP-K mainly has non-repeatable vertex algorithm, repeatable vertex algorithm, index and other algorithm. But these techniques are mainly based on the single attribute. There are few documents introducing the complex multiple attribute decision-making problem so for. Therefore, a Tdp algorithm is presented in this paper. Firstly, it uses the technology of interval number and extreme value to solve the uncertain attribute value. Then TOPSIS technique solves the complex multiple attribute decision-making problems. In this way the comprehensive score was achieved. Secondly, by analysing the Yen algorithm, the paper proposes blocking and bidirectional shortest path algorithm for TOP-K optimal path. Finally, comparison and analysis between Tdp and the Yen were made. Results confirm that the Tdp algorithm improves the TOP-K optimal technology.
Keywords: multiple attribute; optimal path; blocking; bidirectional; deviation vertex; TOP-K; decision-making.