Sign In
Not register? Register Now!
Pages:
7 pages/≈1925 words
Sources:
No Sources
Style:
APA
Subject:
IT & Computer Science
Type:
Reaction Paper
Language:
English (U.S.)
Document:
MS Word
Date:
Total cost:
$ 41.58
Topic:

Mutual Information for Feature Selection

Reaction Paper Instructions:

Remember you wrote approx 2000 works last time and regards this one as a modification and 'sequal' of last one. https://rtm(dot)science(dot)unitn(dot)it/~battiti/archive/mutual-nn.pdf
This link is the main paper for this topic, don't worry about the last several pages he is doing. Just give a screen show of the equations and explain each in words (This address the mathematical explaination mentiones below). This time, please focus only on FILTER METHOD. I will attached a ppt file that includes some of ideas that might inspire you of what you want to write. And in side it, a short paragraph that I want you to write down in the paper.
Please a submit a first draft of your final paper. You should have a thorough and well-grounded introduction to your topic, and you should be able to submit working code for any empirical work that you are doing. This draft should be at least 4000 words, excluding references.
I will be looking for the following in your first draft:
A proper introduction and background section.
A reasonable amount of mathematical background and a connection with information theory.
Background material should be written in your own words, and you should give your intuitive understanding of the mathematics you use.
Talk about MI fits for what kind of feature selection (see the figure in ppt that shows for categorical-categorical) You can give an example to explain Numerical v.s Categorical variables
Discuss at least some of the measures for relevance and redundancy in detail, and discuss their relative strengths and drawbacks. (THESE PART IMPORTANT) There is a related paper that might help with this:
https://link(dot)springer(dot)com/chapter/10.1007/978-3-319-90509-9_4
If you are doing empirical work, please submit your source code separately. (No codnig for you, but leave a space for me to talk about the coding I will be doing)
A clear indication of what work you plan to do to complete the final paper. (Just write I will work on the coding using scikit-learn and demonstrate how does MI work in FS)
Also, I will attach the paper you wrote last time.
Please reach out to me if there's anything you're unsure of.

Reaction Paper Sample Content Preview:

Mutual Information for Feature Selection
Name
Institutional Affiliation
Mutual Information for Feature Selection
Introduction
While the model selection is critical in learning signals from the provided data, providing the right variables or data is of utmost importance. In machine learning, the model building requires the construction of relevant features or variables through feature engineering, and the resulting data set can then be employed as a statistical input to train a model. While these models are often assumed to be sophisticated and smart algorithms, they are easily fooled by unnecessary clutter and dependencies in the data. Data scientists often make the signals to be easily identifiable by performing feature selection, which is a necessary step in data pre-processing (Huijskens, 2017). According to Zhou, Wang, and Zhu (2022), feature selection is a fundamental pre-processing step in machine learning as it selects only the crucial features by eliminating redundant or irrelevant features from the primary dataset. Battiti (1994) recognizes this pre-processing stage as a critical step where the required number of appropriate features are selected from raw data to impact on learning phase complexity and the achievable generalization performance. While using mutual information (MI) for selecting features in supervised neural net learning, Battiti (1994) notes that although it is important that information in the input vector is sufficient in determining the output class, excess input results in the burdening of the training process and thus lead to the production of neural networks with excess connection weights compared to those needed by the problem at hand. Based on an application-oriented perspective, excessive features lengthen the duration for pre-processing and recognition, regardless of a satisfactory performance in learning and recognition (Battiti, 1994). Data scientists use algorithms that maximize only relevant information while minimizing unnecessary or redundant information.
One of these techniques that have been adopted by machine learning experts and data scientists is mutual information feature selection. In this algorithm, the in-filter feature selection approach is used in assessing the relevancy of a subset of features to predict the largest variable as well as the redundancy based on other variables. Nevertheless, Beraha et al. (2019) note that the existing algorithms are often heuristic and fail to provide any guarantee that they will resolve a proposed problem. This limitation has motivated the authors to propose a novel way of observing the theoretical results that indicate conditional mutual information may occur naturally when handling the ideal regression or classification errors that are achieved by various features or a subset of features. One thing to do before selecting is to remove words that appear only infrequently in one category because they are destined to have high mutual information with one category and low mutual information with the others. Ones show that low word frequency has a great influence on mutual information. If a word is not frequent enough but mainly appears in a certain category, it will have a high level of mutual information, which will bring noise to scree...
Updated on
Get the Whole Paper!
Not exactly what you need?
Do you need a custom essay? Order right now:

👀 Other Visitors are Viewing These APA Reaction Paper Samples:

HIRE A WRITER FROM $11.95 / PAGE
ORDER WITH 15% DISCOUNT!