Implementation of k-NN algorithm in python (IRIS data)

Coke 2020-11-13 03:49:59
implementation k-nn nn algorithm python


Running environment :Jupyter

One 、K- Nearest neighbor algorithm

K- a near neighbor (K-NN) The algorithm can be said to be the simplest machine algorithm . To build the model, we only need to save the training data set . Want to predict new data points , The algorithm will find the nearest data point in the training data set , That's what it is “ Nearest neighbor ”.

What we implement here is a classification in supervised learning ( Two classification ) problem . We need to predict the category of test data .

Two 、 Implementation steps

1. Get data set

Import Numpy Easy to operate data ,pyplot For drawing
 Insert picture description here
explain :
(1)load_iris(): Import iris data
(2)x_data = datas[‘data’][0:150],y_data= datas[‘target’][0:150]:
Using slicing operations to get data sets .‘data’ The corresponding is the iris data .target The corresponding category is .

The final source dataset is :
x_data(150 Group source data ):
 Insert picture description here

… altogether 150 Group data
y_data( For categories ):
 Insert picture description here
It's also 150 Group data .

2. The data set is divided into training set and test set

Training set acquisition :
 Insert picture description here
explain : Using slicing operation, half of the source data is used as the training set . from 0 Start to 150, Left closed right away , In steps of 2. obtain 75 Group data .

Test set data acquisition :
 Insert picture description here
explain : Using slicing operation, half of the source data is used as the training set . from 1 Start to 150, Left closed right away , In steps of 2. obtain 75 Group data .

3. Build the model

1. Draw a scatter plot :
 Insert picture description here
 Insert picture description here
Be careful : The abscissa and ordinate here correspond to the length and width of the calyx of iris respectively , Because four dimensional vectors can't be drawn , So take the first two elements for illustration .
2.k-NN The process of :
 Insert picture description here
 Insert picture description here
explain : The black dots indicate the data to be predicted , We need to find the category of the points closest to it , Use voting to determine the category of black dots , According to the figure, the final category of black dots should be the same as that of green dots 0

Concrete algorithm :( Distance is the distance between two coordinate points )  Insert picture description here
 Insert picture description here
The final prediction is 0 This is the same category as the green dot, and this is the end of building the model .

4. Using the data of the test set to test the accuracy

 Insert picture description here
explain : This is actually to repeat the previous operation on the test set data , Then compare the predicted results of the test set data with its correct category , Record the number of correct predictions , Finally, the accuracy of this model is calculated by dividing the total test data .
 Insert picture description here

3、 ... and 、 Algorithm implementation

? The code is sequential

from sklearn.datasets import load_iris
import numpy as np
import matplotlib.pyplot as plt
# Load iris data 
datas = load_iris()
# Using slices to collect 150 Group Iris data (datas['data'] Express datas Of ‘data’key The data corresponding to the value , That is, the data of petals and calyx of iris )
x_data = datas['data'][0:150]
# Represents the category of iris source data ,0 representative setosa,1 representative versicolor,2 representative virginica(K-NN The algorithm deals with the classification problem based on supervised learning )
#datas['target'] Express datas Of ‘target’key The data corresponding to the value , The category of iris ( label )
y_data = datas['target'][0:150]
# Generating training sets ( Percent of iris source data 50)
x_train = x_data[0:150:2]
y_train = y_data[0:150:2]
# Generating test sets ( Percent of iris source data 50)
x_test = x_data[1:150:2]
y_test = y_data[1:150:2]
y_data
# Drawing training sets ( Here, we only use the first two of iris data as coordinates to illustrate the principle of the algorithm , Because four dimensions can't be drawn )
plt.scatter(x_train[y_train==0,0],x_train[y_train==0,1],color='g',label="symbol 0")
plt.scatter(x_train[y_train==1,0],x_train[y_train==1,1],color='r',label="symbol 1")
plt.scatter(x_train[y_train==2,0],x_train[y_train==2,1],color='b',label="symbol 2")
plt.title("k-NN view")
plt.xlabel("x axis")
plt.ylabel("y axis")
plt.legend()
plt.show()
# Add a new data ( This data is an example ) Judge its category as 0 or 1 or 2( According to distance )
x = np.array([4.5023242,3.03123123,1.3023123,0.102123123])
plt.scatter(x_train[y_train==0,0],x_train[y_train==0,1],color='g',label="symbol 0")
plt.scatter(x_train[y_train==1,0],x_train[y_train==1,1],color='r',label="symbol 1")
plt.scatter(x_train[y_train==2,0],x_train[y_train==2,1],color='b',label="symbol 2")
plt.scatter(x[0],x[1],color='black',label="symbol ?")
plt.title("k-NN view")
plt.xlabel("x axis")
plt.ylabel("y axis")
plt.legend()
plt.show()
#K-NN The process ( Calculated distance , And store it in a list )
from math import sqrt
distances = []
for x0 in x_train:
d = sqrt(np.sum((x-x0)**2))
distances.append(d)
# Sort the distances ( From big to small ) Returns the element subscript 
near = np.argsort(distances)
k = 3
# Before selection 5 Categories of the most recent elements 
topK_y = [y_train[i] for i in near[:k]]
topK_y
from collections import Counter
# Count the number of elements ( That is, the number of numbers representing the category of iris )
votes = Counter(topK_y)
votes
# Find the element with the most votes , This method returns a tuple , We just need key value ( That's the category )
# The result of the prediction ,0 representative setosa,1 representative versicolor,2 representative virginica
result = votes.most_common(1)[0][0]
result
# Note that the data above is used here , But the steps start again, because you have to traverse the test set one by one and compare the prediction results with the results in the source data to get the correct rate 
# The accuracy of statistical test data 
count = 0
index = 0
for j in x_test:
distance = []
x = j;
# Calculated distance 
for x1 in x_train:
t = sqrt(np.sum((x-x1)**2))
distance.append(t)
near = np.argsort(distance)
topK_y = [y_train[i] for i in near[:k]]
votes = Counter(topK_y)
result = votes.most_common(1)[0][0]
if y_test[index]==result:
count=count+1
index=index+1
else:
index=index+1
score=count/75
score

If the code is wrong, please refer to : Code

版权声明
本文为[Coke]所创,转载请带上原文链接,感谢

  1. 利用Python爬虫获取招聘网站职位信息
  2. Using Python crawler to obtain job information of recruitment website
  3. Several highly rated Python libraries arrow, jsonpath, psutil and tenacity are recommended
  4. Python装饰器
  5. Python实现LDAP认证
  6. Python decorator
  7. Implementing LDAP authentication with Python
  8. Vscode configures Python development environment!
  9. In Python, how dare you say you can't log module? ️
  10. 我收藏的有关Python的电子书和资料
  11. python 中 lambda的一些tips
  12. python中字典的一些tips
  13. python 用生成器生成斐波那契数列
  14. python脚本转pyc踩了个坑。。。
  15. My collection of e-books and materials about Python
  16. Some tips of lambda in Python
  17. Some tips of dictionary in Python
  18. Using Python generator to generate Fibonacci sequence
  19. The conversion of Python script to PyC stepped on a pit...
  20. Python游戏开发,pygame模块,Python实现扫雷小游戏
  21. Python game development, pyGame module, python implementation of minesweeping games
  22. Python实用工具,email模块,Python实现邮件远程控制自己电脑
  23. Python utility, email module, python realizes mail remote control of its own computer
  24. 毫无头绪的自学Python,你可能连门槛都摸不到!【最佳学习路线】
  25. Python读取二进制文件代码方法解析
  26. Python字典的实现原理
  27. Without a clue, you may not even touch the threshold【 Best learning route]
  28. Parsing method of Python reading binary file code
  29. Implementation principle of Python dictionary
  30. You must know the function of pandas to parse JSON data - JSON_ normalize()
  31. Python实用案例,私人定制,Python自动化生成爱豆专属2021日历
  32. Python practical case, private customization, python automatic generation of Adu exclusive 2021 calendar
  33. 《Python实例》震惊了,用Python这么简单实现了聊天系统的脏话,广告检测
  34. "Python instance" was shocked and realized the dirty words and advertisement detection of the chat system in Python
  35. Convolutional neural network processing sequence for Python deep learning
  36. Python data structure and algorithm (1) -- enum type enum
  37. 超全大厂算法岗百问百答(推荐系统/机器学习/深度学习/C++/Spark/python)
  38. 【Python进阶】你真的明白NumPy中的ndarray吗?
  39. All questions and answers for algorithm posts of super large factories (recommended system / machine learning / deep learning / C + + / spark / Python)
  40. [advanced Python] do you really understand ndarray in numpy?
  41. 【Python进阶】Python进阶专栏栏主自述:不忘初心,砥砺前行
  42. [advanced Python] Python advanced column main readme: never forget the original intention and forge ahead
  43. python垃圾回收和缓存管理
  44. java调用Python程序
  45. java调用Python程序
  46. Python常用函数有哪些?Python基础入门课程
  47. Python garbage collection and cache management
  48. Java calling Python program
  49. Java calling Python program
  50. What functions are commonly used in Python? Introduction to Python Basics
  51. Python basic knowledge
  52. Anaconda5.2 安装 Python 库(MySQLdb)的方法
  53. Python实现对脑电数据情绪分析
  54. Anaconda 5.2 method of installing Python Library (mysqldb)
  55. Python implements emotion analysis of EEG data
  56. Master some advanced usage of Python in 30 seconds, which makes others envy it
  57. python爬取百度图片并对图片做一系列处理
  58. Python crawls Baidu pictures and does a series of processing on them
  59. python链接mysql数据库
  60. Python link MySQL database