An Evaluation of swine flu (Influenza A H3N2v)virus prediction using data mining and Conventional neural network techniques

Nowadays, The health care commercial enterprise collects huge amounts of healthcare data which, unfortunately, are not “mined” to discover hidden information. Data mining plays a significant role in predicting diseases. The database report of medical patient is not more efficient, currently we made an Endeavour to detect the most widely spread disease in all over the world named Swine flu. Swine flu is a respiratory disease which has Numeral number of tests must be requisite from the patient for detecting a disease. Advanced data mining techniques gives us help to remedy this situation. In this work we describes about a prototype using data mining techniques, namely Naive Bayes Classifier. The Data mining is an emerging research trend which helps in finding accurate solutions in many fields. This paper highlights the various data mining technique and Convolution Neural Network used for predicting swine flu diseases.


I. INTRODUCTION
Swine flu is one of the most infectious diseases which contain types of virus that causes thousands of deaths per year. Hence we have use a method of data mining to predict this disease using Naive Bayes classifier thus we can decrease the laboratory test cost as well as time also. "How can we turn useful data into information that can enable healthcare practitioners to make intelligent clinical decisions?" This is the main purpose for this work.
Swine flu is a respiratory disease caused by influenza viruses that infect the respiratory tract of pigs and result in a barking cough, decreased appetite, nasal secretions, and listless behavior the virus can be transmitted to humans. Swine flu viruses may mutate or change so that they are easily transmissible among humans.Swine influenza, is also called pig influenza or swine flu, hog flu and pig flu, is an infection caused by any one of several types of swine influenza viruses. Swine influenza virus (SIV) or swine origin influenza virus (S-OIV) is any strain of the influenza family of viruses that is endemic in pigs. In 2009, the known influenza include in strain and there are some subtypes of influenza A known as H1N1, H1N2, H2N1, H3N1, H3N2, and H2N3. Swine influenza virus is common throughout comes from pig populations. [1,2] The virus transmit from pigs to humans is not common and does not always tend toward human flu, often resulting only in the productivity of antibodies in the blood of human. If transmission does cause human flu, it is known as zoonotic swine flu. People with regular come in contact to pigs are at increased risk of swine flu infection. Around the mid-20th century, the checking of influenza subtypes became possible, allowing exactly right diagnosis of transmission to humans. Since then, only 50 such transmissions have been confirmed. Swine flu is transmitted from person to person by inhalation or ingestion of droplets containing virus from people sneezing or coughing; it is not transmitted by eating cooked pork products. The newest swine flu virus that has caused swine flu is influenza A H3N2v. [7] Some common symptoms of Swine flu are -• Cough • Fever • Sore throat • Runny nose • Headache • Chill • Fatigue • Nausea.

II. LITERATURE REVIEW
Thakkar, Hasan and Desai were inspired to do this research by the study of mortality rate of swine flu [3]. This work focus on the aspect of medical diagnosis by learning predictions through the data collected for Swine flu. We proposed a method to identify swine flu by study 110 symptoms to decrease the cost incurred in the test of the disease. We developed prototype intelligent swine flu prediction software,usedNaive Bayes classifier technique for classifying the patients of swine flu. Based on the possibility of the diseases and guaranteed the accuracy of almost 63.3 percent. In August 2010, the World Health Organization declared the swine flu pandemic officially over. Cases of swine flu have been reported in India, with over 25000 positive test cases and 1370 deaths till March 2015 [8] III. TECHNIQUES USED Concepts used are as follows:

B. Convolution Neural Networks
A. Data Mining • The cause of the 2009 swine flu was an influenza A virus type designated as H1N1. In 2011, a new swine flu virus was detected with new strain was named influenza A (H3N2) v. Only a few people (mainly children) were first infected, but officials from the U.S. Centers for Disease Control and Prevention (CDC) reported increased numbers of people infected in the 2012-2013 flu season. Currently, the swine flu virus kill more than 1500 people in India and it will rapidly spreading all over the world so recently there are not notice large numbers of people infected with H3N2v.Unfortunately, another virus termed H3N2 (note no "v" in its name) has been found and caused flu, but this strain is different from H3N2v. In general, all of the influenza A viruses have a structure similar to the H1N1 virus each type has a somewhat different H and/or N structure.

Fig.1: -Working of Data Mining
Technically, data mining is the process of finding correlations or patterns among different of fields in large relational databases. The step of the "Knowledge Discovery in Databases (KDD)" process, an interdisciplinary subfield of computer science, is the computational process of discovering patterns in large data sets involving methods at the intersection of artificial intelligence, machine learning, statistics, and database systems. The overall goal of the data mining process is to extract information from a data set and transform it into an understandable structure for further use. [ • A CNN convolves (not convolutes…) learned features with input data and uses 2D convolutional layers. This means that this type of network is ideal for processing 2D images. Compared to other image classification algorithms, CNNs actually use very little preprocessing. This means that they can learn the filters that have to be hand-made in other algorithms. CNNs can be used in tons of applications from image and video recognition, image classification, and recommender systems to natural language processing and medical image analysis.
• CNNs are inspired by biological processes. They're based on some cool research done by Hubel and Wiesel in the 60s regarding vision in cats and monkeys. The pattern of connectivity in a CNN comes from their research regarding the organization of the visual cortex. In a mammal's eye, individual neurons respond to visual stimuli only in the receptive field, which is a restricted region. The receptive fields of different regions partially overlap so that the entire field of vision is covered.

Fig. 2:-X-Ray image recognition
CNNs have an input layer, and output layer, and hidden layers. The hidden layers usually consist of convolutional layers, ReLU layers, pooling layers, and fully connected layers.
• Convolutional layers apply a convolution operation to the input. This passes the information on to the next layer.
• Pooling combines the outputs of clusters of neurons into a single neuron in the next layer.
• Fully connected layers connect every neuron in one layer to every neuron in the next layer.
In a convolutional layer, neurons only receive input from a subarea of the previous layer. In a fully connected layer, each neuron receives input from every element of the previous layer.
A CNN works by extracting features from images. This eliminates the need for manual feature extraction. The features are not trained. They are learned while the network trains on a set of images. This makes deep learning models extremely accurate for computer vision tasks. CNNs learn feature detection through tens or hundreds of hidden layers. Each layer increases the complexity of the learned features.

A CNN is
•Starts with an input image •Applies many different filters to it to create a feature map •Applies a ReLU function to increase non-linearity •Applies a pooling layer to each feature map •Flattens the pooled images into one long vector.
Inputs vector into a fully connected artificial neural network.

Pilla Srinivas,
•Processes the features through the network. The final fully connected layer provides the "voting" of the classes that we're after.
•Trains through forward propagation and back propagation for many, many epochs. This repeats until we have a well-defined neural network with trained weights and feature detectors.

A. Very Deep Convolutional Networks for Large
The VGG-16 is one of the most popular pre ILSVRC 2014 Conference, it was and remains THE model to beat even today. Developed at the Visual Graphics Group at the University of Oxford, VGG researchers and the industry for their image Classification Tasks.
As you can see, the model is sequential in nature and uses lots of filters. At each stage, small 3 * 3 filters are used to reduce the number of parameters all the hidden layers use the ReLU activation function. Even then, the number of parameters is 138 Billion Additionally, there are variations of the layers). You can find a detailed explanation

Srinivas, Debnath Bhattacharyya, Divya Midhun Chakkaravarthy
Processes the features through the network. The final fully connected layer provides the "voting" of the classes Trains through forward propagation and back propagation for many, many epochs. This repeats until we have a defined neural network with trained weights and feature detectors.

Very Deep Convolutional Networks for Large-Scale Image Recognition (VGG-16)
16 is one of the most popular pre-trained models for image classification. Introduced in the famous ILSVRC 2014 Conference, it was and remains THE model to beat even today. Developed at the Visual Graphics d, VGG-16 beat the then standard of AlexNet and was quickly adopted by researchers and the industry for their image Classification Tasks.
As you can see, the model is sequential in nature and uses lots of filters. At each stage, small 3 * 3 filters are d to reduce the number of parameters all the hidden layers use the ReLU activation function. Even then, the number of parameters is 138 Billion -which makes it a slower and much larger model to train than others.
Additionally, there are variations of the VGG16 model, which are basically, improvements to it, like VGG19 (19 layers). You can find a detailed explanation

1380
Processes the features through the network. The final fully connected layer provides the "voting" of the classes Trains through forward propagation and back propagation for many, many epochs. This repeats until we have a

16)
trained models for image classification. Introduced in the famous ILSVRC 2014 Conference, it was and remains THE model to beat even today. Developed at the Visual Graphics 16 beat the then standard of AlexNet and was quickly adopted by As you can see, the model is sequential in nature and uses lots of filters. At each stage, small 3 * 3 filters are d to reduce the number of parameters all the hidden layers use the ReLU activation function. Even then, the which makes it a slower and much larger model to train than others.

B.Inception
At only 7 million parameters, it was much smaller than the then prevalent models like VGG and AlexNet. Adding to it a lower error rate, you can see why it was a breakthrough model. Not only was this, but the major innovation in this work also another breakthrough -the Inception Module.
Fig5.: -Inception module with dimension reductions As can be seen, in simple terms, the Inception Module just performs convolutions with different filter sizes on the input, performs Max Pooling, and concatenates the result for the next Inception module. The introduction of the 1 * 1 convolution operation reduces the parameters drastically. Fig.6:-InceptionV2 Model The Inceptionv2 model was a major improvement on the Inceptionv1 model which increased the accuracy and further made the model less complex. In the same paper as Inceptionv2, the authors introduced the Inceptionv3 model with a few more improvements on v2.
The following are the major improvements included: •Introduction of Batch Normalization

D. EfficientNet
We finally come to the latest model amongst these 4 that have caused waves in this domain and of course, it is from Google. In EfficientNet, the authors propose a new Scal and short of it is this: The earlier models like ResNet follow the conventional approach of scaling the dimensions arbitrarily and by adding up more and more layers.
However, the paper proposes that if we scale uniformly, we achieve much better performance. The scaling coefficients can be in fact decided by the user.

Though this scaling technique can be used for any CNN baseline model called EfficientNetB0
MBConv stands for mobile inverted bottleneck Convolution (similar to MobileNetv2). They also propose the Compound Scaling formula with the following scaling coefficients: This formula is used to again build a family The following is a simple graph showing the comparative performance of this family vis models:

Architecture of ResNet family in terms of layers
We finally come to the latest model amongst these 4 that have caused waves in this domain and of course, it is from Google. In EfficientNet, the authors propose a new Scaling method called Compound Scaling. The long and short of it is this: The earlier models like ResNet follow the conventional approach of scaling the dimensions arbitrarily and by adding up more and more layers.
However, the paper proposes that if we scale the dimensions by a fixed amount at the same time and do so uniformly, we achieve much better performance. The scaling coefficients can be in fact decided by the user.
Though this scaling technique can be used for any CNN-based model, the authors started o baseline model called EfficientNetB0 This formula is used to again build a family of Efficient Nets -EfficientNetB0 to EfficientNetB7.
The following is a simple graph showing the comparative performance of this family vis

An Evaluation of swine flu (Influenza A H3N2v) virus prediction using data mining and Conventional neural network techniques
We finally come to the latest model amongst these 4 that have caused waves in this domain and of course, it is ing method called Compound Scaling. The long and short of it is this: The earlier models like ResNet follow the conventional approach of scaling the dimensions the dimensions by a fixed amount at the same time and do so uniformly, we achieve much better performance. The scaling coefficients can be in fact decided by the user. based model, the authors started off with their own stands for mobile inverted bottleneck Convolution (similar to MobileNetv2). They also propose the EfficientNetB0 to EfficientNetB7.
The following is a simple graph showing the comparative performance of this family vis-a-vis other popular

IV. THEOREMS USED
Here we use two types of theorems:

B.Naive Bayesian Classifier
A. Bayes Theorem :Bayes Theorem is a simple mathematical formula used for calculating conditional probabilities.
Conditional probability is a measure of the probability of an event occurring given that another event has (by assumption, presumption, assertion, or evidence) occurred.
The formula isCV

P B A P A A P B P B =
Which tells us: how often A happens given that B happens, written P(A|B) also called posterior probability, When we know: how often B happens given that A happens, written P(B|A) and how likely A is on its own, written P(A) and how likely B is on its own, written P(B).
In simpler terms, Bayes Theorem is a way of finding a probability when we know certain other probabilities.

B. Naive Bayesian Classifier:
Example: -Given all the patients we have seen their Chill Runny nose Headache Fever Swineflu Therefore, the argument of probability of P2 seems maximum than P1 so that patient not having the swine flu.
Types of Naive Bayes Classifier: •Multinomial Naive Bayes: Feature vectors represent the frequencies with which certain events have been generated by a multinomial distribution. This is the event model typically used for document classification.
•Bernoulli Naive Bayes: In the multivariate Bernoulli event model, features are independent Booleans (binary variables) describing inputs. Like the multinomial model, this model is popular for document classification tasks, where binary term occurrence features are used rather than term frequencies.
•Gaussian Naive Bayes: In Gaussian Naive Bayes, continuous values associated with each feature are assumed to be distributed according to a Gaussian distribution (Normal distribution). When plotted, it gives a bell-shaped curve which is symmetric about the mean of the feature values.
The likelihood of the features is assumed to be Gaussian. Hence, conditional probability is given by: Now, what if any feature contains numerical values instead of categories i.e. Gaussian distribution.
One option is to transform the numerical values to their categorical counterparts before creating their frequency tables. The other option, as shown above, could be using the distribution of the numerical variable to have a good guess of the frequency. For example, one common method is to assume normal or Gaussian distributions for numerical variables.
The probability density function for the normal distribution is defined by two parameters (mean and standard deviation). The next step is to preprocess text before splitting the dataset into a train and test set. The preprocessing steps involve Removing Numbers, Removing Punctuations in a string, Removing Stop Words, Stemming of Words and Lemmatization of Words.

Constructing a Naive Bayes Classifier:
Combine all the preprocessing techniques and create a dictionary of words and each word's count in training data.
1.Calculate probability for each word in a text and filter the words which have a probability less than threshold probability. Words with probability less than threshold probability are irrelevant.
2.Then for each word in the dictionary, create a probability of that word being in insincere questions and its probability insincere questions, then finding the conditional probability to use in naive Bayes classifier.

V.CONCLUSION
The Data mining technique can be used in collaboration with a naive bayes classifiers algorithm which used in diagnosing Swine flu disease. The proposed approach showed Promising results which may lead to further attempts to utilize information technology for diagnosing patients for Swine flu diseases. Here we used Naivebayes Classification rules which are easy to interpret. In future, we will try to get more the accuracy for the swine flu disease patient by increasing the various parameters suggested from the doctors by using different data mining techniques.