Статьи журнала - International Journal of Information Technology and Computer Science
Все статьи: 1195
Enhanced Dynamic Algorithm of Genome Sequence Alignments
Статья научная
The merging of biology and computer science has created a new field called computational biology that explore the capacities of computers to gain knowledge from biological data, bioinformatics. Computational biology is rooted in life sciences as well as computers, information sciences, and technologies. The main problem in computational biology is sequence alignment that is a way of arranging the sequences of DNA, RNA or protein to identify the region of similarity and relationship between sequences. This paper introduces an enhancement of dynamic algorithm of genome sequence alignment, which called EDAGSA. It is filling the three main diagonals without filling the entire matrix by the unused data. It gets the optimal solution with decreasing the execution time and therefore the performance is increased. To illustrate the effectiveness of optimizing the performance of the proposed algorithm, it is compared with the traditional methods such as Needleman-Wunsch, Smith-Waterman and longest common subsequence algorithms. Also, database is implemented for using the algorithm in multi-sequence alignments for searching the optimal sequence that matches the given sequence.
Бесплатно
Enhanced Initial Centroids for K-means Algorithm
Статья научная
This paper focuses on the enhanced initial centroids for the K-means algorithm. The original k-means is using the random choice of initial seeds which is a major limitation of the original K-means algorithm because it produces less reliable result of clustering the data. The enhanced method of the k-means algorithm includes the computation of the weighted mean to improve the centroids initialization. This paper shows the comparison between K-Means and the enhanced K-Means algorithm, and it proves that the new method of selecting initial seeds is better in terms of mathematical computation and reliability.
Бесплатно
Enhanced PROBCONS for multiple sequence alignment in cloud computing
Статья научная
Multiple protein sequence alignment (MPSA) intend to realize the similarity between multiple protein sequences and increasing accuracy. MPSA turns into a critical bottleneck for large scale protein sequence data sets. It is vital for existing MPSA tools to be kept running in a parallelized design. Joining MPSA tools with cloud computing will improve the speed and accuracy in case of large scale data sets. PROBCONS is probabilistic consistency for progressive MPSA based on hidden Markov models. PROBCONS is an MPSA tool that achieves the maximum expected accuracy, but it has a time-consuming problem. In this paper firstly, the proposed approach is to cluster the large multiple protein sequences into structurally similar protein sequences. This classification is done based on secondary structure, LCS, and amino acids features. Then PROBCONS MPSA tool will be performed in parallel to clusters. The last step is to merge the final PROBCONS of clusters. The proposed algorithm is in the Amazon Elastic Cloud (EC2). The proposed algorithm achieved the highest alignment accuracy. Feature classification understands protein sequence, structure and function, and all these features affect accuracy strongly and reduce the running time of searching to produce the final alignment result.
Бесплатно
Статья научная
A text summarization system generates short and brief summaries of original document for given user queries. The machine generated summaries uses information retrieval techniques for searching relevant answers from large corpus. This research article proposes a novel framework for generating machine generated summaries using reinforcement learning techniques with Non-deterministic reward function. Experiments have exemplified with ROUGE evaluation metrics with DUC 2001, 20newsgroup data. Evaluation results of proposed system with hypothesis of automatic summarization from given datasets prove that statistically significant improvement for answering complex questions with f- actual vs. critical values.
Бесплатно
Enhancing Big Data Value Using Knowledge Discovery Techniques
Статья научная
The world has been drowned by floods of data due to technological development. Consequently, the Big Data term has gotten the expression to portray the gigantic sum. Different sorts of quick data are doubling every second. We have to profit from this enormous surge of data to convert it to knowledge. Knowledge Discovery (KDD) can enhance detecting the value of Big Data based on some techniques and technologies like Hadoop, MapReduce, and NoSQL. The use of Big Data value is critical in different fields. This survey discusses the expansion of data that led the world to Big Data expression. Big Data has distinctive characteristics as volume, variety, velocity, value, veracity, variability, viscosity, virality, ambiguity, and complexity. We will describe the connection between Big Data and KDD techniques to reach data value. Big Data applications that are applied by big organizations will be discussed. Characteristics of big data will be introduced, which represent a significant challenge for Big Data management. Finally, some of the important future directions in Big Data field will be presented.
Бесплатно
Статья научная
Brain tumors are among the deadliest forms of cancer, and there is a significant death rate in patients. Identifying and classifying brain tumors are critical steps in understanding their functioning. The best way to treat a brain tumor depends on its type, size, and location. In the modern era, Radiologists utilize Brain tumor locations that can be determined using magnetic resonance imaging (MRI). However, manual tests and MRI examinations are time-consuming and require skills. In addition, misdiagnosis of tumors can lead to inappropriate medical therapy, which could reduce their chances of living. As technology advances in Deep Learning (DL), Computer Assisted Diagnosis (CAD) as well as Machine Learning (ML) technique has been developed to aid in the detection of brain tumors, radiologists can now more accurately identify brain tumors. This paper proposes an MRI image classification using a VGG16 model to make a deep convolutional neural network (DCNN) architecture. The proposed model was evaluated with two sets of brain MRI data from Kaggle. Considering both datasets during the training at Google Colab, the proposed method achieved significant performance with a maximum overall accuracy of 96.67% and 97.67%, respectively. The proposed model was reported to have worked well during the training period and been highly accurate. The proposed model's performance criteria go beyond existing techniques.
Бесплатно
Статья научная
Hospitals are the primary hubs for healthcare service providers in Ethiopia; however, hospitals face significant challenges in adopting digital health information systems solutions due to disparate, non-interoperable systems and limited access. Information technology, especially via cloud computing, is crucial in healthcare for efficient data management, secure storage, real-time access to critical information, seamless provider communication, enhanced collaboration, and scalable IT infrastructure. This study investigated the challenges to standardizing smart and green healthcare information services and proposed a cloud-based model for overcoming them. We conducted a mixed-methods study in 11 public hospitals, employing quantitative and qualitative approaches with diverse stakeholders (N = 103). The data was collected through surveys, interviews, and technical observations by purposive quota sampling with the Raosoft platform and analyzed using IBM SPSS. Findings revealed several shortcomings in existing information systems, including limited storage, scalability, and security; impaired data sharing and collaboration; accessibility issues; no interoperability; ownership ambiguity; unreliable data recovery; environmental concerns; affordability challenges; and inadequate policy enforcement. Notably, hospitals lacked a centralized data management system, cloud-enabled systems for remote access, and modern data recovery strategies. Despite these challenges, 90.3% of respondents expressed interest in adopting cloud-enabled data recovery systems. However, infrastructure limitations, inadequate cloud computing/IT knowledge, lack of top management support, digital illiteracy, limited innovation, and data security concerns were identified as challenges to cloud adoption. The study further identified three existing healthcare information systems: paper-based methods, electronic medical catalog systems, and district health information systems2. Limitations of the paper-based method include error-proneness, significant cost, data fragmentation, and restricted remote access. Growing hospital congestion and carbon footprint highlighted the need for sustainable solutions. Based on these findings, we proposed a cloud-based model tailored to the Ethiopian context. This six-layered model, delivered as a Software-as-a-Service within a community cloud deployment, aims to improve healthcare services through instant access, unified data management, and evidence-based medical practices. The model demonstrates high acceptability and potential for improving healthcare delivery, and implementation recommendations are suggested based on the proposed model.
Бесплатно
Статья научная
We address the challenge of optimizing the interaction between medical personnel and treatment stations within mobile and flexible medical care units (MFMCUs) in conflict zones. For the analysis of such systems, a closed queuing model with a finite number of treatment stations has been developed, which accounts for the possibility of performing multiple tasks for a single medical service request. Under the assumption of Poisson event flows, a system of integro-differential equations for the probability densities of the introduced states has been compiled. To solve it, the method of discrete binomial transformations is employed in conjunction with production functions. Solutions were obtained in the form of finite expressions, enabling the transition from the probabilistic characteristics of the model to the main performance metrics of the MFMCU: the load factor of medical personnel, and the utilization rate of treatment stations. The results show the selection of the number of treatment stations in the medical care area and the calculation of the appropriate performance of medical personnel.
Бесплатно
Enhancing Jakarta Faces Web App with AI Data-Driven Python Data Analysis and Visualization
Статья научная
Python is widely used in artificial intelligence (AI) and machine learning (ML) because of its flexibility, adaptability, rich libraries, active community, and broad environment, which makes it a popular choice for AI development. Python compatibility has already been examined with Java using TCP socket programming on both non-graphical and graphical user interfaces, which is highly essential to implement in the Jakarta Faces web application to grab potential competitive advantages. Python data analysis library modules such as numpy, pandas, and scipy, as well as visualization library modules such as Matplotlib and Seaborn, and machine-learning module Scikit-learn, are intended to be integrated into the Jakarta Faces web application. The research method uses similar TCP socket programming for the enhancement process, which allows instruction and data exchange between Python and Jakarta Faces web applications. The outcome of the findings emphasizes the significance of modernizing data science and machine learning (ML) workflows for Jakarta Faces web developers to take advantage of Python modules without using any third-party libraries. Moreover, this research provides a well-defined research design for an execution model, incorporating practical implementation procedures and highlighting the results of the innovative fusion of AI from Python into Jakarta Faces.
Бесплатно
Enhancing the Performance in Generating Association Rules using Singleton Apriori
Статья научная
Association rule mining aims to determine the relations among sets of items in transaction database and data repositories. It generates informative patterns from large databases. Apriori algorithm is a very popular algorithm in data mining for defining the relationships among itemsets. It generates 1, 2, 3,…, n-item candidate sets. Besides, it performs many scans on transactions to find the frequencies of itemsets which determine 1, 2, 3,…, n-item frequent sets. This paper aims to eradicate the generation of candidate itemsets so as to minimize the processing time, memory and the number of scans on the database. Since only those itemsets which occur in a transaction play a vital role in determining frequent itemset, the methodology that is proposed in this paper is extracting only single itemsets from each transaction, then 2,3,..., n itemsets are generated from them and their corresponding frequencies are also calculated. Further, each transaction is scanned only once and no candidate itemsets is generated both resulting in minimizing the memory space for storing the scanned itemsets and minimizing the processing time too. Based on the generated itemsets, association rules are generated using minimum support and confidence.
Бесплатно
Ensemble approach for twitter sentiment analysis
Статья научная
Due to enlargement of social network and online marketing websites. The Blogs and reviews of the user are acquired from these websites. And these become useful for analysis and Decision making for various types of products, marketing and movie etc. with the extent of the usefulness of social Reviews. It is to be needed carefully analysis of that data. There are various techniques and methods are available that can accurately analyses the social information and provides greater accuracy for the analysis. But one of the major issues available with the social media data is that data is unstructured and noisy. It is to be required to solve this problem. So here in this paper a framework is proposed that includes latest data preprocessing techniques instead of noise removal like stemming, Lemmatization and Tokenization. After Pre-Processing of data ensemble methods is applied that increase the accuracy of previous classification algorithms. This method is inherent from bagging concept. First apply Decision Tree, Kneighbor and Naive Bayes classifier that not provide batter accuracy after that boosting concept is applied with the help of AdaBoost method that improves the accuracy of previous classical classifiers. At last our proposed ensemble method ExtraTree classifier is applied that inherent from bagging concept. Here we use the Extra Tree classifier that take the various sample are taken from training set and various random trees are created. It is also called as extremely randomized tree that provides extreme refined view. So that, it is to be conveying that The ExtraTree classifier of bagging ensemble method outperforms than all other techniques that are previously applied in this paper. with using some novel pre-processing techniques data that produced is more refined and that provides clean and pure base for the implementation of ensemble techniques. And also contributes in improving the accuracy of the applied methods.
Бесплатно
Enterprise architecture measurement: an extended systematic mapping study
Статья научная
A systematic mapping study (SMS) of proposed EA measurement solutions was undertaken to provide an in-depth understanding of the claimed achievements and limitations in evidence-based research of enterprise architecture (EA). This SMS reports on 22 primary studies on EA measurement solutions published up to the end of 2018. The primary studies were analyzed thematically and classified according to ten (10) mapping questions including, but not limited to, positioning of EA measurement solutions within EA schools of thought, analysis of consistency-inconsistency of the terms used by authors in EA measurement research, and an analysis of the references to the ISO 15939 measurement information model. Some key findings reveal that the current research on EA measurement solutions is focused on the “enterprise IT architecting” school of thought, does not use rigorous terminology as found in science and engineering, and shows limited adoption of knowledge from other disciplines. The paper concludes with new perspectives for future research avenues in EA measurement.
Бесплатно
Entity Extraction from Business Emails
Статья научная
Email still plays an important role in today's business communication thanks to its simplicity, flexibility, low cost, and compatibility of diversified types of information. However processing the large amount of emails received consumes tremendous time and human power for a business. In order to quickly deciphering information and locate business-related information from emails received from a business, a computerized solution is required. In this paper, we have proposed a comprehensive mechanism to extract important information from emails. The proposed solution integrates semantic web technology with natural language processing and information retrieval. It enables automatic extraction of important entities from an email and makes batch processing of business emails efficient. The proposed mechanism has been used in a Transportation company.
Бесплатно
Envisioning Skills for Adopting, Managing, and Implementing Big Data Technology in the 21st Century
Статья научная
The skills for big data technology provide a window of new job opportunities for the information technology (IT) professionals in the emerging data science landscape. Consequently, the objective of this paper is to introduce the research results of suitable skills required to work with big data technology. Such skills include Document Stored Database; Key-value Stored Database; Column-oriented Database; Object-oriented Database; Graph Database; MapReduce; Hadoop Distributed File System (HDFS); YARN Framework; Zookeeper; Oozie; Hive; Pig; HBase; Mahout; Sqoop; Spark; Flume; Drill; Programming Languages; IBM Watson Analytics; Statistical Tools; SQL; Project Management; Program Management; and Portfolio Management. This paper is part of an ongoing research that addresses the link between economic growth and big data.
Бесплатно
Error Detection in a Multi-user Request System Using Enhanced CRC Algorithm
Статья научная
Error and error related issues have been a challenge in the development and reliable usage of computing systems and application. The ability to detect minute error in a system improves the reliability of the system by aiding developers and users to know were challenges are so that they can be fixed during development and even when the system is already in use. In other to achieve that different algorithm have been used including the Cyclic Redundancy Check 16-bit, 32-bit and higher bits. In this paper, error detection schemes are examined and the way they check and detect error in multi-user request and transmitted system. The paper also offers improvement on a Cyclic Redundancy Checks 32-bit detection algorithm for the detection of error that can occur in transmitted data and on stored, backed-up and archived data in the system without consuming large resources as the higher bits.
Бесплатно
Estimating Software Reliability by Monitoring Software Execution through OpCode
Статья научная
Previous studies on estimating software reliability employed statistical functions for next system failure prediction. These models used parameters based on assumptions regarding the nature of software faults and debugging process. However, none of the existing models, attempted on ensuring reliable runtime system operation. To serve the current demand of autonomous, reliable, service-oriented software, we present a novel approach for runtime reliability estimation of executable software. The approach can help control software execution at runtime by monitoring software state-to-state transition at runtime. The approach involves representing executable software as an automata using opcode extracted from executable code. The extracted opcode is then used to learn stochastic finite state machine (SFSM) representation of executable software which is later employed to trace software state-to-state transition at each runtime instance. An evaluation of our approach on Java-based Chart generator application is also discussed to explain how we can ensure reliable software execution and prevent software failures at runtime with the proposed approach.
Бесплатно
Evaluating Design Patterns of Commercial Web Applications using Net Easy Score
Статья научная
Web interface design patterns provide solutions to recurring design problems. Many design patterns use various techniques, which have been proven to be significantly different, to solve the same design problem. Normally, web designers do not know whether users would be satisfied with their chosen choice until near or at the end of the web development process. To obtain user feedback, users are usually asked to interact with a web prototype or the finished web and give their opinion through standardized questionnaires. Net Promoter Score is one of such questionnaires. This scale categorizes users’ responses into promoters and detractors, which makes it easier for companies to understand user satisfaction towards their web. To enable the designers to obtain user feedback early in the design stage, Net Easy Score, a new metric based on Net Promoter Score, was proposed. With Net Easy Score (NES), ease-of-use scores on different design patterns will be divided into a positive and a negative group. The NES is a difference between percentages of positive responses and negative ones. This study examined ease-of-use scores on design patterns for five common tasks in commercial web applications. Results showed that NES and mean ease-of-use score were significantly correlated with an r of 0.965 (p < .000). Also, ranking the average ease-of-use scores and NES revealed the same design patterns identified as the best and the worst ones, which was consistent with the easiest-to-use design patterns voted by participants.
Бесплатно
Evaluating Web Services Functionality and Performance
Статья научная
Traditional distributed database transaction applications within large organizations often involve a large number of resources. In this case, people and DDBMSs distributed over a wide geographic area, may introduce conflict between heterogeneous systems. Web services (WS) provide solution for this problem since WS have an independent platform, independent language, and independent object model. This work presents WS application to access heterogeneous and distributed database via horizontal data fragments that is designed to be reliable, flexible and scalable. It describes the setup of SOAP server and applications based on the SOAP for end user client. In addition, it allows the publishing of WS descriptions to submit user requests (goal) to retrieve the required information. Here we evaluate the functional, behavior and performance of WS among possible different alternatives with real-time and execution parameters. Implementation details and case study experiments are presented along with the corresponding results.
Бесплатно
Статья научная
Most security and privacy issues in software are related to exploiting code vulnerabilities. Many studies have tried to find the correlation between the software characteristics (complexity, coupling, etc.) quantified by corresponding code metrics and its vulnerabilities and to propose automatic prediction models that help developers locate vulnerable components to minimize maintenance costs. The results obtained by these studies cannot be applied directly to web applications because a web application differs in many ways from a non-web application: development, use, etc. and a lot of evaluation of these conclusions has to be made. The purpose of this study is to evaluate and compare the vulnerabilities prediction power of three types of code metrics in web applications. There are a few similar studies that targeted non-web application and to the best of our knowledge, there are no similar studies that targeted web applications. The results obtained show that unlike non-web applications where complexity metrics have better vulnerability prediction power, in web applications the metrics that give better prediction are the coupling metrics with high recall (> 75%) and fewer costs in terms of inspection (<25%).
Бесплатно
Evaluating the Maintainability of a Software System by using Fuzzy Logic Approach
Статья научная
Maintainability is an important quality attribute for almost every quality model. Maintainability of the software is considered as most expensive phase in software development life cycle as it consumes almost major part of the total effort allocated to the software system. Maintainability evaluation is complex due to its imprecise output. This paper proposes a maintainability model by considering its fuzzy aspects. Since fuzzy modeling deals with uncertainty and impreciseness so this paper uses fuzzy methodology and AHP technique to evaluate the maintainability of the model. Object oriented system has taken as case study for maintainability evaluation purpose.
Бесплатно