Local means-based fuzzy k-nearest neighbor classifier with Minkowski distance and relevance-complementarity feature weighting

  • Abstract
  • Highlights & Summary
  • Literature Map
  • Similar Papers
Abstract
Translate article icon Translate Article Star icon
Take notes icon Take Notes

This paper introduces an enhanced fuzzy k-nearest neighbor (FKNN) approach called the feature-weighted Minkowski distance and local means-based fuzzy k-nearest neighbor (FWM-LMFKNN). This method improves classification accuracy by incorporating feature weights, Minkowski distance, and class representative local mean vectors. The feature weighting process is developed based on feature relevance and complementarity. We improve the distance calculations between instances by utilizing feature information-based weighting and Minkowski distance, resulting in a more precise set of nearest neighbors. Furthermore, the FWM-LMFKNN classifier considers the local structure of class subsets by using local mean vectors instead of individual neighbors, which improves its classification performance. Empirical results using twenty different real-world data sets demonstrate that the proposed method achieves statistically significantly higher classification performance than traditional KNN, FKNN, and six other related state-of-the-art methods.

Similar Papers
  • Research Article
  • Cite Count Icon 64
  • 10.1016/j.patrec.2020.10.005
A new fuzzy k-nearest neighbor classifier based on the Bonferroni mean
  • Oct 11, 2020
  • Pattern Recognition Letters
  • Mahinda Mailagaha Kumbure + 2 more

A new fuzzy k-nearest neighbor classifier based on the Bonferroni mean

  • Research Article
  • Cite Count Icon 21
  • 10.1007/s00500-020-05311-x
A new globally adaptive k-nearest neighbor classifier based on local mean optimization
  • Oct 3, 2020
  • Soft Computing
  • Zhibin Pan + 3 more

The k-nearest neighbor (KNN) rule is a simple and effective nonparametric classification algorithm in pattern classification. However, it suffers from several problems such as sensitivity to outliers and inaccurate classification decision rule. Thus, a local mean-based k-nearest neighbor classifier (LMKNN) was proposed to address these problems, which assigns the query sample with a class label based on the closest local mean vector among all classes. It is proven that the LMKNN classifier achieves better classification performance and is more robust to outliers than the classical KNN classifier. Nonetheless, the unreliable nearest neighbor selection rule and single local mean vector strategy in LMKNN classifier severely have negative effect on its classification performance. Considering these problems in LMKNN, we propose a globally adaptive k-nearest neighbor classifier based on local mean optimization, which utilizes the globally adaptive nearest neighbor selection strategy and the implementation of local mean optimization to obtain more convincing and reliable local mean vectors. The corresponding experimental results conducted on twenty real-world datasets demonstrated that the proposed classifier achieves better classification performance and is less sensitive to the neighborhood size $$k$$ compared with other improved KNN-based classification methods.

  • Conference Article
  • Cite Count Icon 3
  • 10.1109/ijcnn.2005.1556000
Iterative feature weighting for identification of relevant features with radial basis function networks
  • Dec 27, 2005
  • Baofu Duan + 1 more

This paper reports on advances in identification of relevant features through iterative feature weighting with radial basis function networks. It proceeds with a set of feature weights to scale the data which are used to train a radial basis function network model. Then from the learned model, the feature weights are updated via one-step gradient descent. The updated feature weights are then fed back to build a new model. The procedure continues until we find a satisfactory model and the feature weights converge. Experimental results for some benchmark datasets show that the approach is efficient and effective for selecting relevant features for data modeling and classification tasks.

  • Research Article
  • 10.3233/jifs-202779
A novel weighted fuzzy c-means based on feature weight learning
  • Dec 16, 2021
  • Journal of Intelligent & Fuzzy Systems
  • Xingguang Pan + 4 more

In feature weighted fuzzy c-means algorithms, there exist two challenges when the feature weighting techniques are used to improve their performances. On one hand, if the values of feature weights are learnt in advance, and then fixed in the process of clustering, the learnt weights might be lack of flexibility and might not fully reflect their relevance. On the other hand, if the feature weights are adaptively adjusted during the clustering process, the algorithms maybe suffer from bad initialization and lead to incorrect feature weight assignment, thus the performance of the algorithms may degrade the in some conditions. In order to ease these problems, a novel weighted fuzzy c-means based on feature weight learning (FWL-FWCM) is proposed. It is a hybrid of fuzzy weighted c-means (FWCM) algorithm with Improved FWCM (IFWCM) algorithm. FWL-FWCM algorithm first learns feature weights as priori knowledge from the data in advance by minimizing the feature evaluation function using the gradient descent technique, then iteratively optimizes the clustering objective function which integrates the within weighted cluster dispersion with a term of the discrepancy between the weights and the priori knowledge. Experiments conducted on an artificial dataset and real datasets demonstrate the proposed approach outperforms the-state-of-the-art feature weight clustering methods. The convergence property of FWL-FWCM is also presented.

  • Research Article
  • Cite Count Icon 86
  • 10.1093/comjnl/bxr131
A Local Mean-Based k-Nearest Centroid Neighbor Classifier
  • Jan 5, 2012
  • The Computer Journal
  • J Gou + 3 more

K-nearest neighbor (KNN) rule is a simple and effective algorithm in pattern classification. In this article, we propose a local mean-based k-nearest centroid neighbor classifier that assigns to each query pattern a class label with nearest local centroid mean vector so as to improve the classification performance. The proposed scheme not only takes into account the proximity and spatial distribution of k neighbors, but also utilizes the local mean vector of k neighbors from each class in making classification decision. In the proposed classifier, a local mean vector of k nearest centroid neighbors from each class for a query pattern is well positioned to sufficiently capture the class distribution information. In order to investigate the classification behavior of the proposed classifier, we conduct extensive experiments on the real and synthetic data sets in terms of the classification error. Experimental results demonstrate that our proposed method performs significantly well, particularly in the small sample size cases, compared with the state-of-the-art KNN-based algorithms.

  • Research Article
  • Cite Count Icon 1176
  • 10.1016/j.asoc.2019.105524
Investigating the impact of data normalization on classification performance
  • May 23, 2019
  • Applied Soft Computing
  • Dalwinder Singh + 1 more

Investigating the impact of data normalization on classification performance

  • Research Article
  • Cite Count Icon 3
  • 10.1007/s10462-019-09700-z
A comparative study of neural-network feature weighting
  • Mar 27, 2019
  • Artificial Intelligence Review
  • Tongfeng Sun + 3 more

Many feature weighting methods have been proposed to evaluate feature saliencies in recent years. Neural-network (NN) feature weighting, as a supervised method, is founded upon the mapping from input features to output decisions, and implemented by evaluating the sensitivity of network outputs to its inputs. Through training on sample data, NN implicitly embodies the saliencies of input features. The partial derivatives of the outputs with respect to the inputs in the trained NN are calculated to measure their sensitivities to input features, which means that implicit feature weighting of the NN is transformed into explicit feature weighting. The purpose of this paper is to further probe into the principle of NN feature weighting, and evaluate its performance through a comparative study between NN feature weighting method and state-of-art weighting methods in the same working conditions. The motivation of this study is inspired by the lack of direct and comprehensive comparison studies of NN feature weighting method. Experiments in UCI repository data sets, face data sets and self-built data sets show that NN feature weighting method achieves superior performance in different conditions and has promising prospects. Compared with the other existing methods, NN feature weighting method can be used in more complex conditions, provided that NN can work in those conditions. As decision data, output data can be labels, reals or integers. Especially, feature weights can be calculated without the discretization of outputs in the condition of continuous outputs.

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 32
  • 10.1109/access.2020.2968984
Class-Specific Deep Feature Weighting for Naïve Bayes Text Classifiers
  • Jan 1, 2020
  • IEEE Access
  • Shufen Ruan + 3 more

Feature weighting is used to alleviate the conditional independence assumption of Naïive Bayes text classifiers and consequently improve their generalization performance. Most traditional feature weighting algorithms use general feature weighting, which assigns the same weight to each feature for all classes. We focus on class-specific feature weighting approaches, which discriminatively assign each feature a specific weight for each class. This paper uses a statistical feature weighting technique and proposes a new class-specific deep feature weighting method for Multinomial Naïve Bayes text classifiers. In this deep feature weighting method, feature weights are not only incorporated into the classification formulas but they are also incorporated into the conditional probability estimates of Multinomial Naïve Bayes text classifiers. Experimental results for a large number of text classification datasets validate the effectiveness and efficiency of our method.

  • Research Article
  • Cite Count Icon 286
  • 10.1016/j.engappai.2016.02.002
Deep feature weighting for naive Bayes and its application to text classification
  • Feb 27, 2016
  • Engineering Applications of Artificial Intelligence
  • Liangxiao Jiang + 3 more

Deep feature weighting for naive Bayes and its application to text classification

  • Research Article
  • 10.15379/ijmst.v10i5.3624
Improving Kinnow Fruit Classification with Feature Weighting using Modified Levy Flight Distribution-Inspired Sunflower Optimization Technique
  • Dec 4, 2023
  • International Journal of Membrane Science and Technology
  • Sukhpreet Singh + 1 more

In this paper, Feature weighting is used to create an intelligent and effective classification method for Kinnow fruits. Feature weighting approach is used because it improves classification performance more than feature selection methods. The modified sunflower optimization algorithm (SFO) is proposed to search the optimal feature weights and parametric values of k-Nearest Neighbour (kNN).The levy flight distribution operator has been utilised to enhance the convergence speed of the sunflower optimization algorithm by improving the local and global search ability of the optimization algorithm. Also, the algorithmic parameter of the SFO algorithm has been adaptively selected using the linear time varying adaption method. In addition, tanh normalization technique is used for the data pre-processing to reduce the influence of outliers and dominating features before the feature weighting method. The findings suggest that the proposed wrapper based approach feature weighting technique is more capable of achieving higher accuracy than the existing strategies.

  • Research Article
  • Cite Count Icon 1
  • 10.1016/j.neucom.2024.128150
Safe dynamic sparse training of modified RBF networks for joint feature selection and classification
  • Jul 6, 2024
  • Neurocomputing
  • Xusheng Qian + 5 more

Safe dynamic sparse training of modified RBF networks for joint feature selection and classification

  • Research Article
  • Cite Count Icon 72
  • 10.1007/s00357-016-9208-4
A Survey on Feature Weighting Based K-Means Algorithms
  • Jul 1, 2016
  • Journal of Classification
  • Renato Cordeiro De Amorim

In a real-world data set, there is always the possibility, rather high in our opinion, that different features may have different degrees of relevance. Most machine learning algorithms deal with this fact by either selecting or deselecting features in the data preprocessing phase. However, we maintain that even among relevant features there may be different degrees of relevance, and this should be taken into account during the clustering process. With over 50 years of history, K-Means is arguably the most popular partitional clustering algorithm there is. The first K-Means based clustering algorithm to compute feature weights was designed just over 30 years ago. Various such algorithms have been designed since but there has not been, to our knowledge, a survey integrating empirical evidence of cluster recovery ability, common flaws, and possible directions for future research. This paper elaborates on the concept of feature weighting and addresses these issues by critically analyzing some of the most popular, or innovative, feature weighting mechanisms based in K-Means.

  • Single Report
  • 10.37686/ser.v1i2.79
A survey on feature weighting based K-Means algorithms
  • Dec 11, 2020
  • Renato Cordeiro De Amorim

In a real-world data set there is always the possibility, rather high in our opinion, that different features may have different degrees of relevance. Most machine learning algorithms deal with this fact by either selecting or deselecting features in the data preprocessing phase. However, we maintain that even among relevant features there may be different degrees of relevance, and this should be taken into account during the clustering process. With over 50 years of history, K-Means is arguably the most popular partitional clustering algorithm there is. The first K-Means based clustering algorithm to compute feature weights was designed just over 30 years ago. Various such algorithms have been designed since but there has not been, to our knowledge, a survey integrating empirical evidence of cluster recovery ability, common flaws, and possible directions for future research. This paper elaborates on the concept of feature weighting and addresses these issues by critically analysing some of the most popular, or innovative, feature weighting mechanisms based in K-Means

  • Research Article
  • Cite Count Icon 76
  • 10.1016/j.bbe.2019.12.004
Simultaneous feature weighting and parameter determination of Neural Networks using Ant Lion Optimization for the classification of breast cancer
  • Dec 25, 2019
  • Biocybernetics and Biomedical Engineering
  • Singh Dalwinder + 2 more

Simultaneous feature weighting and parameter determination of Neural Networks using Ant Lion Optimization for the classification of breast cancer

  • Research Article
  • Cite Count Icon 67
  • 10.1016/j.knosys.2014.07.020
Improved pseudo nearest neighbor classification
  • Jul 31, 2014
  • Knowledge-Based Systems
  • Jianping Gou + 5 more

Improved pseudo nearest neighbor classification

Save Icon
Up Arrow
Open/Close
  • Ask R Discovery Star icon
  • Chat PDF Star icon

AI summaries and top papers from 250M+ research sources.