Incorporating Connectivity in k-Nearest Neighbors Regression

dc.AffiliationOctober university for modern sciences and Arts MSA
dc.contributor.authorMahfouz, Mohamed A
dc.date.accessioned2023-09-28T12:02:39Z
dc.date.available2023-09-28T12:02:39Z
dc.date.issued2023-07
dc.description.abstractThe standard k-nearest neighbors' approach to regression analysis encounters a number of problems when used on datasets with various density distributions. This paper proposes a kNNR-relative ensemble regressor based on connectivity. In each cross-validation round, the pipeline starts by clustering the input data using any partitioning algorithm. Then, a random sample of the edges is selected from each partition favoring the edges with small distances. After that, the selected edges are transformed into a dataset in which feature values represent the amount of increase or decrease in each dimension compared to the source node's values and the label of each feature vector is the difference in the label of the source and the destination. Then a regressor is built for each cluster based on the output of the transformer. In order to predict a label for an unseen object, the nearest centroid is identified, and, k-nearest neighbors from the corresponding cluster are identified as the source nodes. Then, a vector representing the difference between the unseen object and each source node is computed and fed to the regressor model of the corresponding cluster, the output label is the predicted difference so it is added to the label of the source node. The diversity between the suggested decision model and the traditional kNN regressor, termed kNNR motivates us to include the kNNR in the suggested ensemble. The k-nearest neighbors of kNNR are also selected from the nearest cluster. The weighted average of the predicted labels offered by the base models serves as the final output label. The sample size, the number of neighbors to be used, and the number of clusters can all be fine-tuned via cross validation. The ensemble is evaluated, and the results showed that the ensemble achieved a significant increase in effectiveness compared to its base regressors and several related algorithms.en_US
dc.description.urihttps://08104euot-1103-y-https-ieeexplore-ieee-org.mplbci.ekb.eg/document/10217502/authors
dc.identifier.doi10.1109/IMSA58542.2023.10217502
dc.identifier.other10.1109/IMSA58542.2023.10217502
dc.identifier.urihttp://repository.msa.edu.eg/xmlui/handle/123456789/5734
dc.language.isoenen_US
dc.relation.ispartofseries1st International Conference of Intelligent Methods, Systems and Applications, IMSA 2023;Pages 551 - 5562023
dc.subjectclustering; Nearest Neighbors; Regressionen_US
dc.titleIncorporating Connectivity in k-Nearest Neighbors Regressionen_US
dc.typeArticleen_US

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
MSA avatar.jpg
Size:
49.74 KB
Format:
Joint Photographic Experts Group/JPEG File Interchange Format (JFIF)
Description:

License bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
51 B
Format:
Item-specific license agreed upon to submission
Description: