Welcome to CCSITA 2023

International Conference on Computer Science and Information Technology Advances (CCSITA 2023)

October 07 ~ 08, 2023, Virtual Conference



Accepted Papers
Potential Impact of Generative Artificial Intelligence(AI) on the Financial Industry

Professor Suman Kalia, Department of Computer Science and Information Sciences, Saint Peter's University, Jersey City, NJ

ABSTRACT

Presently, generative AI has taken center stage in the news media, educational institutions, and the world at large. Machine learning has been a decades-old phenomenon, with little exposure to the average person until very recently. In the natural world, the oldest and best example of a “generative” model is the human being - one can close one’s eyes and imagine several plausible different endings to one’s favorite TV show. This paper focuses on the impact of generative and machine learning AI on the financial industry. Although generative AI is an amazing tool for a discriminant user, it also challenges us to think critically about the ethical implications and societal impact of these powerful technologies on the financial industry. It requires ethical considerations to guide decision-making, mitigate risks, and ensure that generative AI is developed and used to align with ethical principles, social values, and in the best interests of communities.

KEYWORDS

Generative AI, Machine Learning, Deep Learning, Financial Industry.


An Improved Ctgan for Data Processing Method of Imbalanced Disk Failure

Jingbo Jia, Peng Wu, and Hussain Dawood, School of Information Science and Engineering, University of Jinan, Jinan, Department of Information Engineering Technology, National Skills University, Islamabad,Islamabad

ABSTRACT

To address the problem of insufficient failure data generated by disks and the imbalance between the number of normal and failure data. The existing Conditional Tabular Generative Adversarial Networks(CTGAN) deep learning methods have been proven to be effective in solving imbalance disk failure data. But CTGAN cannot learn the internal information of disk failure data very well. In this paper, a fault diagnosis method based on improved CTGAN, a classifier for specific category discrimination is added and a discriminator generate adversarial network based on residual network is proposed. We named it Residual Conditional Tabular Generative Adversarial Networks (RCTGAN). Firstly, to enhance the stability of system a residual network is utilized. RCTGAN uses a small amount of real failure data to synthesize fake fault data; Then, the synthesized data is mixed with the real data to balance the amount of normal and failure data; Finally, four classifier (multilayer perceptron, support vector machine, decision tree, random forest) models are trained using the balanced data set, and the performance of the models is evaluated using G-mean. The experimental results show that the data synthesized by the RCTGAN can further improve the fault diagnosis accuracy of the classifier.

KEYWORDS

Imbalanced dataset, CTGAN, Synthesis data, Disk failure data, Classification.


Data-level Hybrid Strategy Selection for Disk Fault Prediction Model Based on Multivariate GAN

Shuangshuang Yuan, Peng Wu, and Yuehui Chen, School of Information Science and Engineering, University of Jinan, 250022, Jinan, China

ABSTRACT

Data class imbalance is a common problem in classification problems, where minority class samples are often more important and more costly to misclassify in a classification task. Therefore, it is very important to solve the data class imbalance classification problem. The SMART dataset exhibits an evident class imbalance, comprising a substantial quantity of healthy samples and a comparatively limited number of defective samples. This dataset serves as a reliable indicator of the disc’s health status. In this paper, we obtain the best balanced disk SMART dataset for a specific classification model by mixing and integrating the data synthesised by multivariate generative adversarial networks (GAN) to balance the disk SMART dataset at the data level; and combine it with genetic algorithms to obtain higher disk fault classification prediction accuracy on a specific classification model.

KEYWORDS

Disk failure prediction, GAN, Genetic algorithm, Data imbalance, SMART data.


Disk Failure Prediction Based on Multi-layer Domain Adaptive Learning

Guangfu Gao, Peng Wu, and Hussain Dawood, School of lnformation Science and Engineering, University of Jinan, Jinan Department of Information Engineering Technology, National Skills University, Islamabad, Islamabad

ABSTRACT

Large scale data storage is susceptible to failure. As disks are damaged and replaced, traditional machine learning models, which rely on historical data to make predictions, struggle to accurately predict disk failures. This paper presents a novel method for predicting disk failures by leveraging multi-layer domain adaptive learning techniques. First, disk data with numerous faults is selected as the source domain, and disk data with fewer faults is selected as the target domain. A training of the feature extraction network is performed with the selected origin and destination domains. The contrast between the two domains facilitates the transfer of diagnostic knowledge from the domain of source and target. According to the experimental findings, it has been demonstrated that the suggested approach possesses the capability to produce a dependable predictive model and enhance the accuracy in forecasting disk failures, even when only a limited number of failure samples are available.

KEYWORDS

Disk failure prediction, Transfer learning, Domain adaptation, Distance metric.


A Survey of Methods for Handling Disk Data Imbalance

Shuangshuang Yuan, Peng Wu, Yuehui Chen, and Qiang Li, School of Information Science and Engineering, University of Jinan, Jinan 250022, China

ABSTRACT

Class imbalance exists in many classification problems, and since the data is designed for accuracy, imbalance in data classes can lead to classification challenges with a few classes having higher misclassification costs. The Backblaze dataset, a widely used dataset related to hard discs, has a small amount of failure data and a large amount of health data, which exhibits a serious class imbalance. In this paper, we summarise and sort out the research in the field of unbalanced data from three aspects, namely, data-level methods, algorithmic-level methods and hybrid methods, and summarise and analyse the problems, algorithmic ideas, strengths and weaknesses of each type of methods, as well as discussing the challenges of unbalanced data classification and strategies to deal with them.

KEYWORDS

Classification, Disk Failure Prediction, Imbalanced Dataset, Data Processing.


ESTIMATING CLOUD COMPUTING ROUND-TRIP TIME (RTT) USING FUZZY LOGIC FOR INTER-REGION DISTANCES

Ihab Sekhi, Institute of Information Technology, Miskolc University, 3515 Miskolc

ABSTRACT

Cloud computing is widely considered a transformative force in the computing world and is poised to replace the traditional office setup as an industry standard. However, given the relative novelty of these services and challenges such as the impact of physical distance on Round-Trip Time (RTT), questions have arisen regarding system performance and associated billing structures. The primary objective of this study is to address these concerns. We aim to alleviate doubts by leveraging a fuzzy logic system to classify distances between regions that support computing services and compare them with the conventional web hosting format. To achieve this, we analyse the responses of one of these services, like Amazon Web Services, across different distance categories (near, medium, and far) between regions and strive to conclude overall system performance. Our tests reveal that significant data is consistently lost during customer transmission despite exhibiting superior round-trip times. We delve into this issue and present our findings, which may illuminate the observed anomalous behaviour.

KEYWORDS

Round Trip Time, Wireless Network, SLA, cloud computing.


Implementation of Neural Key Generation Algorithm for Iot Devices

Zied Guitouni1, Aya Zairi2 and Mounir Zrigui2, 1Department of Electronic and Micro-Electronic Laboratory, Faculty of Sciences of Monastir, Tunisia, 2Informatics Department, Faculty of Sciences of Monastir, 5000, Tunisia

ABSTRACT

Cryptographic key generation plays a vital role in securing data transmission and device authentication in Internet of Things (IoT) systems. However, traditional random key generation methods face challenges in terms of security, efficiency and scalability, especially for resource-constrained IoT devices. Neural networks have emerged as a promising technique for generating cryptographic keys due to their ability to learn complex patterns. However, their architecture has a significant impact on performance. This paper provides a comparative analysis of three commonly used neural network architectures for cryptographic key generation on IoT devices. We propose a novel neural network-based key generation algorithm and implement it using each architecture. The models are trained to generate cryptographic keys of different sizes from random input data. Performance is evaluated based on key metrics like accuracy, loss, key randomness, and model complexity. Experimental results show that the FFNN architecture achieves over 99% accuracy and passes all randomness tests, outperforming the alternatives. CNNs underperform due to focusing on spatial features irrelevant for keys. RNNs struggle with keys' complex long-range dependencies.

KEYWORDS

Neural network architectures, cryptographic key generation algorithm, feedforward neural networks, convolutional neural networks, recurrent neural networks, evaluation for security in IoT devices .