International Journal of Computational Intelligence Studies (22 papers in press)
Wolf : A framework for digital workplace - Architecture and models -
by Khadija ELAMRANI, Noureddine Chenfour, Mohamed LAHMER, Ghita Daoudi
Abstract: The main purpose of the digital workplace (DW) is to ensure to the organizations different contributors or actors a portal of digital services, which are accessible through a virtual desktop covering all its business services. During our studies, we were able to identify five major problems. First of all, we note a great confusion in the related definitions because most of them are restricted to the teaching sector. Secondly, most existing DWs are summarized as a simple gateway to pre-existing digital tools collection that covers the organizations business domains, without any means of communication between them. Another problem is the lack of a reference architecture. Moreover, we could not identify any logical or physical model to represent the different DWs entities. Lastly, there is a total absence of a standard or even an appropriate vocabulary.rnFaced with these shortfalls, we propose in this paper a set of fundamentals that is composed by a definition encapsulating the different domains, as well as a naming system and a vocabulary that identify both the entities that compose the virtual desktop and their connections and flows. Based on these fundamentals, we also propose our framework WOLF (Digital Workplace based on Open and Light architecture Framework) that generate automatically customized digital workplaces, and is distinguished from other existing DWs solutions by its generic and extensible character. The generated DW encapsulates all of the organizations domains, services, flows and a collaboration system between the different actors. Our proposed frameworks architecture allows us to classify and organize the various entities into a tree representation whilst data nodes are modelled using XML files.
Keywords: Digital workplace; Digital workspace; Collaboration; Digital work environment.
Application of Computational Intelligence Techniques for Internet of
Things: An Extensive Survey
by Shreyas J, Anand Jumnal, Dilip Kumar S M, Venugopal K R
Abstract: The application of computational intelligence (CI) techniques to Internet of Things (IoT) is gaining popularity due to its capability of providing human-like knowledge, such as cognition, recognition, understanding, learning, and others. This paper attempts to provide an exhaustive survey of the available literature on IoT using CI techniques. In addition, detailed categorization has been provided on the basis of different CI tools and their hybridizations used to tackle different problems of IoT. The potential benefits and utility of CI techniques in IoT are highlighted. The possible mapping of CI techniques to the real-world IoT problems is presented. The advantages and disadvantages of CI algorithms over traditional IoT solutions are discussed. A general evaluation of CI algorithms is presented, which will serve as a guide for using CI algorithms for IoT. Finally, some considerations regarding the recent trends and potential research directions are presented. An extensive bibliography is also included.
Keywords: Artificial Immune System; Computational Intelligence; Fuzzy Logi; Genetic Algorithms; Internet of Things; Neural Networks; Swarm Intelligence.
Alzheimer's disease prediction using Regression models and SVM
by Rohini M
Abstract: Alzheimer's disease (AD) and cognitive impairment due to aging are the recently prevailing diseases among aged inhabitants because of an increase in the aging population. Several demographic characters, structural and functional neuroimaging investigations, cardio-vascular studies, neuropsychiatric symptoms, cognitive performances and biomarkers in cerebrospinal fluids are the various predictors for AD. We can consider these input features for the prediction of symptoms whether they belong to AD or normal cognitive impairment for aging. In the proposed study, the hypothesis is derived for supervised learning methods such as multivariate linear regression, logistic regression, and SVM. We perform feature scaling and normalization with features as an initial step for applying the parameters to derive the hypothesis. We analyze performance metrics with the implementation results. The present work is applied to 1000 baseline assessment data from Alzheimers disease Neuro-Imaging Initiative studies (ADNI) that give conversion prediction. The comparison of results in literature studies suggests that the efficiency of the proposed study is highly helpful in differentiating AD pathology from cognitive impairment because of aging.
Keywords: Multivariate linear regression; logistic regression; Support Vector Machine(SVM);Feature scaling; Normalization;ADNI.
Detection of various categories of Fruit and vegetable through various descriptor using machine learning techniques
by Mukesh Tripahthi
Abstract: An accurate and efficient recognition system for fruits and vegetables is one major challenges. To solve this challenges, we have examined various feature descriptor based on colour and texture such as RGB, CMH, CCV, CDH, LBP, CSLBP and SEH. All process of proposed framework consists three phase 1) background subtraction 2) Feature extraction 3) training and classification. In this paper, Otsus thresholding is used for background subtraction. Further all segmented image is used in feature extraction phase. Finally, C4.5 and KNN is used for training and classification. The various performance metric such as CA, Precision, Recall, F-measure, MCC, PRC and FPR are used to evaluate the proposed system for recognition problem. We, also analysis performance accuracy of both classifiers. In that C4.5, KNN classifier produce CA with value 94.63 %, 90.25 % respectively.
Keywords: Detection; fruits; vegetables; descriptor; Performance metric; C4.5; KNN.
FEATURE SELECTION ON EDUCATIONAL DATA USING BORUTA ALGORITHM
by Neeyati Anand, Riya Sehgal, Sanchit Anand, Ajay Kaushik
Abstract: Data Mining in education deals with formulating strategies for students with the aim to increase the parameters affecting the learning and employability. It also helps the educational institutes in maintaining their reputation as it is directly linked to the students grades. We need to identify the parameters involved in learning and the relationship among those parameters. In EDM, Feature Selection (FS) is one of the most important and needed method in EDM. As it removes the features which has no direct link with the students performance. For example, Date of Birth of a student doesnt impact his/her performance. In this paper an attempt has been made to improve the performance of the classifiers for undergraduate students at Maharaja Agrasen Institute Of Technology. We have applied several techniques of Data Mining to make some rules that increases the learning and employability of students. The results of our study has shown a significant increase of accuracy, recall, precision and F-measure for Na
Keywords: Classifiers; Educational Data mining; Feature Selection; Performance.
Deep Learning for Apple Diseases: Classification and Identification
by Asif Iqbal Khan, S.M.K. Quadri, Saba Banday
Abstract: Diseases and pests cause huge economic loss to the apple industry every year. Precise identification of various apple diseases is challenging for the farmers due to their similar appearance in some stages. This paper is an attempt to provide the timely and accurate detection and identification of apple diseases. In this study, we propose a deep learning based approach for identification and classification of apple diseases. The first part of the study is dataset creation which includes data collection and data labeling. Next, we train a Convolutional Neural Network (CNN) model on the prepared dataset for automatic classification of apple diseases. CNNs are end-to-end learning algorithms which perform automatic feature extraction and learn complex features directly from raw images, making them suitable for wide variety of tasks like image classification, object detection, segmentation etc. We applied transfer learning to initialize the parameters of the proposed deep model. Data augmentation techniques like rotation, translation, reflection and scaling was also applied to prevent overfitting. The proposed CNN model obtained encouraging results, reaching around 97.18% of accuracy on our prepared dataset. The results validate that the proposed method is effective in classifying various types of apple diseases and can be used as a practical tool by farmers
Keywords: Apple Diseases; Deep Learning; Apple disease classification; Convolutional Neural Networks.
Special Issue on: GUCON-2018 Intelligent Models for Emerging Technologies
Empirical Estimation of Various Data Stream Mining Methods
by Ritesh Srivastava, Veena Tayal
Abstract: The online learning is done in order to work in dynamic environments in which concept tend to change with time and the accuracy of classifiers decreases. The current and previous research is done in static environments, but there is a need of real time data streaming due to the potentially larger number of applications available in the scientific and business domains. There are several methods used in learning in the presence of dynamic environments like single classifier methods such as batch and incremental learning approaches, classification methods with explicit drift detection method, windowing techniques and ensemble approaches. This paper, investigates these approaches for determining the best suitable method among them. We utilized light emitting diode (LED) data generator for evaluating the performance of the methods.
Keywords: Concept drifts; online learning; data stream mining; machine learning; classification; drift detection methods.
Special Issue on: IEEE IWCIA2019 Innovative Computational Intelligence for Deep Learning and Knowledge Acquisition
Code Completion for Programming Education based on Deep Learning
by Kenta Terada, Yutaka Watanobe
Abstract: In solving programming problems, it is difficult for beginners to create a program from scratch. One way to navigate this difficulty is to suggest the next word following an incomplete program. In the present study, we propose a method for code completion characterized by two principal elements: the prediction of the next within-vocabulary word and the prediction of the next referenceable identifier. For the prediction of within-vocabulary words, a neural language model based on an LSTM network with an attention mechanism is proposed. Additionally, for the prediction of referenceable identifiers, a model based on a pointer network to a given incomplete program is proposed. For evaluation of the proposed method, source code accumulated in an online judge system is used. The results of the experiment demonstrate that in both statically and dynamically typed languages, the proposed method can predict the next word to a high degree of accuracy.
Keywords: Programming education; Code completion; Deep learning; LSTM; Pointer network.
Multi-objective optimization of allocations and locations of incineration facilities with Voronoi diagram and genetic algorithm: Case study of Hiroshima city and Aki county
by Taketo Kamikawa, Takashi Hasuike
Abstract: This research focuses on the two objectives of maximizing the amount of heat generated by incineration and minimizing the waste collection distances divided by population densities, in determining allocations and locations of general waste incineration facilities as a case study of Hiroshima city and Aki county in Japan. For these objectives, we propose the version 2 of Multi-Objective optimization with Voronoi diagram and Genetic Algorithm (MOVGA2). As for the maximization of the amount of generated heat, we predict the amount by the regression equation of multiple linear regression analysis using 2013 to 2017 panel data and formulate it as the set partitioning problem (SPP) to maximize the prediction value. As for the minimization of waste collection distances divided by population densities, we formulate it as the multi-Weber problem. To solve these two problems, we use MOVGA2, which has the seeds and weights of the Laguerre Voronoi diagram as a gene. As a result of the survey using data of the year 2017 of Hiroshima city and Aki county, in the case of 2 existing facilities, 2 new facilities and 3 closed facilities it was found that the calorific value increased enough to cover the power of 1,233 households (converted to housing complex) per year despite the increase of only 9% t-km for waste collection distances.
Keywords: genetic algorithm; Voronoi diagram; thermal energy; incineration facility; combustible waste.
Efficient Parameter-Free Adaptive Penalty Method with Balancing the Objective Function Value and the Constraint Violation
by Takeshi Kawachi, Jun-ichi Kushida, Akira Hara, Tetsuyuki Takahama
Abstract: Real world problems are often formularized as constrained optimization problems (COPs). Constraint handling techniques are important for efficient search, and various approaches such as penalty methods or feasibility rule have been studied. The penalty methods deal with a single fitness function by combining an objective function value and a constraint violation with a penalty factor. Moreover, the penalty factor can be flexibly adapted by feeding back information on search process in adaptive penalty methods. However, setting parameters properly and keeping the good balance between the objective function value and the constraint violation are difficult. In this paper, we propose a new parameter-free adaptive penalty method with balancing the objective function value and the constraint violation. L-SHADE is adopted as a base search algorithm, and the optimization results of 28 benchmark functions provided by the CEC2017 and CEC2018 competitions on constrained single-objective numerical optimizations are compared with other methods.
Keywords: Evolutionary Algorithms; Differential Evolution; Constraint Handling Techniques; Penalty Method;.
Detecting Audio Adversarial Examples for Protecting Speech-to-Text Transcription Neural Network
by Keiichi Tamura, Akitada Omagari, Hajime Ito, Shuichi Hashida
Abstract: With the increasing use of deep learning techniques in real-world applications, their vulnerabilities have received significant attention from deep-learning researchers and practitioners. In particular, adversarial examples for deep neural networks and protection methods against them have been well-studied in recent years because they have serious vulnerabilities that threatens safety in real-world. Audio adversarial examples, which are targeted attacks, are designed such that the deep neural network-based speech-to-text systems misunderstand input voice sound. In this study, we propose a new protection method against audio adversarial examples. The proposed protection method is based on a sandbox approach, where an input voice sound is checked in the system to determine if it is an audio adversarial example. To evaluate the proposed protection method, we used actual audio adversarial examples created on Deep Speech, which is a typical speech-to-text transcription neural network. The experimental results show that our protection method can detect audio adversarial examples with high accuracy.
Keywords: Adversarial Example; Deep Learning; Computer Security; Data Representation; Speech-to-Text; Sandbox Method.
Special Issue on: IEEE IWCIA2019 Innovative Computational Intelligence for Deep Learning and Knowledge Acquisition
Using Term Similarity Measures for Classifying Short Document Data
by Hirohisa Seki, Shuhei Toriyama
Abstract: Term expansion (a.k.a. document expansion), proposed by Carpineto et al., is a method used for text classification. When handling short text data like social media and blogs, we can apply the term expansion method to expand the sparse information in them. While the prior works on term expansion use an FCA (Formal Concept Analysis)-based similarity measure defined between terms (or words), this paper studies the effectiveness of using two kinds of measures for term expansion: one is weighted similarity measures studied in FCA, and the other is some correlation measures, like cosine and all-conf, often employed in data mining. We also present some properties on the relationship between these term similarity/correlation measures and the notion of relevancy in classification. We show empirically that cosine correlation measure outperforms the prior methods in our two short document data. We also make a comparison of our approach with an LDA (Latent Dirichlet Allocation)-based term expansion approach by Rogers et al.
Keywords: term expansion; similarity measure,; correlation; formal concepts; LDA; short document data; classification.
A Video Prediction Method by using Long Short Term Memory based Adaptive Structural Learning of Deep Belief Network and its Investigation of Input Sequence Length for Data Structure
by Shin Kamada, Takumi Ichimura
Abstract: Artificial Intelligence (AI) with sophisticated technologies has become an essential technique in our life. Especially, deep learning has been a successful model which can effectively represent several features of input space and remarkably improve image recognition performance on the deep architectures. In our research, Adaptive RBM and Adaptive DBN have been developed as a deep learning model. The adaptive structural learning can find a suitable size of network structure for given input space during its training. This is, the neuron generation and annihilation algorithms were implemented on Restricted Boltzmann Machine (RBM), and layer generation algorithm was implemented on Deep Belief Network (DBN). Moreover, the learning algorithm of Adaptive RBM and Adaptive DBN was extended to the time-series prediction by using the idea of LSTM (Long Short Term Memory). Our previous research tackled the problems for supervised learning, in this paper, we challenge to reveal the power of our proposed method in the video recognition research field by using Moving MNIST for unsupervised learning, since video includes rich source of visual information. Moving MNIST is a benchmark data set for video recognition where two digits are moving for randomly determined direction in time sequence. Adaptive LSTM-DBN trained the time-series movement or features from the video data and it showed higher future prediction performance than the existing LSTM models (more than 90\ for test data). Moreover, the detailed prediction results were investigated by training Adaptive LSTM-DBN with various length of input sequence. As a result, we found some cases that the models represented multiple time-series movements of not only short-term simple movements but also long-term complicated features, according to the length of input sequence.
Keywords: Deep learning; Deep Belief Network; Adaptive structural learning method; Video recognition
Special Issue on: BDCA'18 Data Science and Applications
Information Technology performance management by Artificial Intelligence in Microfinance Institutions: An overview
by Kaicer Mohammed
Abstract: This paper presents an overview of the use of new information technology to improve the management of microfinance institutions, experiencing a gap due to the growth of the microfinance sector and the diversity of products and services they offer to the target populations. We will show that artificial intelligence could play a role to ensure reliable management information systems in MFIs.
Keywords: Management of Informatics Technology; Artificial intelligence; Microfinance Institution; Central risk.
Intelligent intrusion detection system using multilayer perceptron optimized by genetic algorithm
by Mehdi Moukhafi, Khalid El Yassini, Bri Seddik
Abstract: This paper presents a neural network-based intrusion detection method for the attacks on a computer network. Neural networks are used to predict unusual activities in the system. In particular, feedforward neural networks with the back propagation training algorithm were employed in this study. we propose a method of intrusion detection based on a combination of GA(Genetic algorithm) and MLP (Multilayer Perceptron) Neural Network to develop a model for intrusion detection system. All tests were realized with the kdd99 data set. The performance of the proposed method of intrusion detection was evaluated on all KDD99 data set, 10% of the KDD99 data set were used for training the GA-MLP model. This system achieves a top accuracy of up to 93.05%.
Keywords: Machine Learning Based Intrusion Detection; Parameters optimization; Genetic algorithm; Multilayer Perceptron Neural Network.
QoE in Video Streaming over Ad-hoc Networks: Comparison and Analysis of AODV and OLSR Routing Protocols
by Hind ZIANI, Nourddine ENNEYA
Abstract: Video Streaming services are easily among of the most consumed services on the internet. Indeed, they are single-handedly accountable for up to 85% of overall internet traffic. And yet, despite the multiple modern infrastructure networks and high-end technologies which remain in constant evolution, network masters still assess Quality levels by its dependent and independent factors. Furthermore, new venues of marketing strategies are constantly witnessing the emergence of ever-novel, ever-revolutionary quality horizons which centralize Human perception. And so, in addition to the Quality of Service (QoS) which is built upon network-oriented metrics, we are now faced with stakes bearing on the Quality of Experience (QoE). In MANET, guaranteeing good quality and performance, be it objective or subjective, is a challenge to be reckoned with. In fact, The extant routing protocols are generally network-oriented and are, as such, chiefly dependent upon objective quality parameters, whence they seldom correlate with the QoE standards as averred by the users perception of the received service. This article purports to analyze and experiment on video transmission, through an Ad-hoc network, based on two emblematic routing protocols -AODV and OLSR- in view of identifying the one most relevant to, and optimal for the subjective quality (QoE) we are focused on.
Keywords: Ad-hoc networks; Video Streaming; Routing protocols; Quality of Experience.
Efficient of Bitmap Join Indexes for optimizing Star Join Queries in Relational Data Warehouses
by Mohammed YAHYAOUI, Souad AMJAD, Lamia BENAMEUR, Ismail JELLOULI
Abstract: Data warehouses are dedicated to analysis and decision-making applications. They are often schematized as star relational models or variants for on-line analysis. Typically, the analysis process is conducted via OLAP (On-Line Analytical processing) type queries. These queries are usually complex, characterized by multiple selections operations, joins, grouping, and aggregations on large tables. Which require a lot of calculation time and thus a very high response time. The performance of these queries depends directly on the use of the secondary memory. Indeed, each input-output on disk requiring up to ten milliseconds. In order to reduce and minimize the cost of executing these queries, the data warehouse administrator must make a good physical design during the physical design and tuning phase by optimizing access to the secondary memory. We focus on bitmap join indexes that share the same resource, that is, the selection attributes extracted from the business intelligence queries. To optimize star join queries.
Keywords: Data Warehouse; OLAP; Indexes; Optimization Query; Star join query; Bitmap join indexes.
Special Issue on: Intelligent Systems for Cyber Security Current Trends, Applications and New Challenges
Intrusion Detection using Data Mining
by Shubha Puthran, Ketan Shah
Abstract: Intrusion Detection plays very important role in securing Information Servers. Classification and Clustering Data Mining algorithms are very effective to deal with Intrusion Detection. However, classification (supervised) results with false negative detection and Clustering (unsupervised) results with false positive detection. This paper introduces a unique framework consisting of Pre-processing unit, Intrusion detection using quad split(IDTQS), Intrusion Detection using Correlation based quad split (IDTCA) and Intrusion Detection using Clustering (IDTC). In this proposed framework, IDTQS and IDTCA shows accuracy improvement for University of New South Wales (UNSW) dataset is in the range 4%-34% for DoS, Probe, R2L, U2R and Normal classes. IDTC Clustering algorithm performs with 97% accuracy. Training and testing time is improved by 14% for IDTCA in comparison with IDTQS.
Keywords: Quad split; Decision Tree; Correlated Attributes; UNSW dataset.
An Integrated Approach for Multimodal Biometric Recognition System using Pearson Type-II (Beta) Distribution
by Naga Jagadesh Bommagani, A.V.S.N Murty
Abstract: Biometric recognition plays an important role in personnel identity authentication. Usually biometric recognition protocols which involve single source of information are called unimodal systems. Such systems suffer from the problems like noisy sensor data, performance, collectability and non-universality. To have an accurate recognition it is needed to develop a system with multimodal biometrics. Hence, in this paper a new approach is proposed with the combination of multiple biometric traits such as face, fingerprint and palm vein. Region of Interest (ROI) is used to consider the valuable information from the images. The 2D Discrete Cosine Transform is used for extracting the feature vector from face, fingerprint and palm vein and fusion at feature extraction level. Here the feature vector is modelled with Pearson Type-II distribution and the model parameters are estimated using the EM algorithm. The initialization of model parameters is done through moment method of estimators and K-means algorithm. The performance of the proposed algorithm is carried by experimentation with CASIA biometric database. Through experimentation the proposed model performs more effectively than the algorithm with Gaussian mixture model.
Keywords: Multimodal biometric recognition; Discrete Cosine Transform; EM algorithm; Pearson mixture model.
IbPaKdE: Identity-Based Pairing free Authenticated Key and Data Exchange Protocol for Wireless Sensor Networks
by Lakshmana Rao Kalabarige
Abstract: The security vulnerabilities in key distribution approaches of WSN
reveals important credentials. The secure distribution of keys without having
permanent storage of important credentials in the permanent memory part
(ROM) of a sensor node is a challenging task. Further, the design of a key
distribution approach with less computational complexity, energy efficient, low
communication overhead, and low memory overhead are some challenging
tasks for a resource constrained sensor nodes. This piece of work addresses
all these challenges by combining Identity-Based Cryptography(IBC) with
Symmetric Key Cryptography(SKC). The proposed Identity-Based Pairing free
Authenticated Key and Data Exchange Protocol(IbPaKdE) avails advantages
of both IBC and SKC to address the above challenges. This approach does
not require prior communication for the establishment of secret keys and it
supports pairing free key distribution. The proposed IbPaKdE uses IBC strategy
for secure exchange of keys and SKC to provide security to the data to be
transmitted. The on demand establishment of keys eliminates the permanent
storage of important credentials in the sensor nodes. Simulation results of the
proposed approach is compared with hashed identity based secure key and data
exchange(HISKDE). The results show that the IbPaKdE incur better results than
HISKDE in terms of energy efficiency, reduces memory, communication and
computational, overheads of a sensor node.
DCT statistics and pixel correlation based blind image steganalysis for identification of covert communication
by Madhavi B. Desai, S.V. Patel, Vipul H. Mistry
Abstract: In the last decade, the interconnection of systems through networks, access to information, different computer technologies and the combination of all these aspects have increased the use of image steganography techniques for illegal acts. Furtherance of image steganography techniques exploited to send secret information on social network builds the requisite of blind image steganalysis. Blind image steganalysis is one in which no prior information is available about the data hiding method used to embed the message. Existing image steganalysis methods are either domain specific or the one which requires a very high dimensional feature set. Considering the types of image steganography methods, embedding rates, image types and feature dimensionality there is an utmost need of a low dimensional blind image steganalysis method. This paper proposes a blind steganalysis method with a 32-D feature set comprising of DCT Statistics and Pixel Correlation (DSPC) algorithm with the aim to the reduced time complexity of feature extraction as well as the complexity of classifier. The experimental results evidence that the proposed feature set gives better results against state-of-art high dimensional image steganalysis methods. The performance of the proposed algorithm is evaluated using experiments with varying embedding message size, message types and image formats using Ensemble classifier. The algorithm is implemented in Matlab and all the experiments are performed on standard image datasets i.e. BSDS300, CorelDraw.
Keywords: Blind Image Steganalysis; Binary Similarity Measures; DCT Transform; Ensemble Classifier; Feature Extraction; Statistical Features; and SVM.
Adaptive QoS Constraint-based Service Differentiated Routing In Wireless Sensor Network
by Yogita Patil
Abstract: Achieving the best quality of service (QoS) as per user requirement is one of the important challenges. The time-critical applications in Wireless Sensor Networks (WSN) demand energy efficient transmission of data with limited resource availability. To resolve these issues AQSDR has been proposed. The proposed protocol support packet differentiation and selection of sensor node based on energy, delay, and congestion for path establishment to transfer normal data packets providing energy conservation. The multipath is chosen for transmission of emergency packets satisfying delay requirements of non-delay tolerant applications. AQSDR support adaptive path selection, according to application requirement. The proposed technique in this work outperforms when compared with the existing protocol in terms of minimized energy consumption, delay, control overhead, packet drop ratio, and high throughput.
Keywords: Clustering; Congestion Index; Delay; Energy Efficiency; Multipath; Service differentiation; WSN.