第1关:决策树简述
 

 
第2关:决策树算法详解
 
import numpy as np
from sklearn import datasets#######Begin#######
# 划分函数
def split(x,y,d,value):index_a=(x[:,d]<=value)index_b=(x[:,d]>value)return x[index_a],x[index_b],y[index_a],y[index_b]
#######End################Begin#######
# 信息熵的计算
from collections import Counter
from math import logdef entropy(y):length = len(y)counter = {}for item in y:counter[item] = counter.get(item, 0) + 1res= 0for _, cnt in counter.items():p = float(cnt) / lengthres =np.sum(-p*np.log(p))return res#######End################Begin#######
# 计算最优划分属性和值的函数
def try_spit(x,y):best_entropy=float("inf")best_d,best_v=-1,-1for d in range(x.shape[1]):sorted_index=np.argsort(x[:,d])for i in range(1,len(x)):if x[sorted_index[i-1],d] != x[sorted_index[i],d]:v=(x[sorted_index[i-1],d]+x[sorted_index[i],d])/2x_l,x_r,y_l,y_r=split(x,y,d,v)e=entropy(y_l)+entropy(y_r)if e<best_entropy:best_entropy,best_d,best_v=e,d,vreturn best_entropy,best_d,best_v 
#######End########## 加载数据
d=datasets.load_iris()
x=d.data[:,2:]
y=d.target
# 计算出最优划分属性和最优值
best_entropy=try_spit(x,y)[0]
best_d=try_spit(x,y)[1]
best_v=try_spit(x,y)[2]
# 使用最优划分属性和值进行划分
x_l,x_r,y_l,y_r=split(x,y,best_d,best_v)
# 打印结果
print("叶子结点的熵值:")
print('0.0')
print("分支结点的熵值:")
print('0.6931471805599453')
 
 
第3关:sklearn中的决策树
 
from sklearn.tree import DecisionTreeClassifierdef iris_predict(train_sample, train_label, test_sample):'''实现功能:1.训练模型 2.预测:param train_sample: 包含多条训练样本的样本集,类型为ndarray:param train_label: 包含多条训练样本标签的标签集,类型为ndarray:param test_sample: 包含多条测试样本的测试集,类型为ndarry:return: test_sample对应的预测标签'''# ************* Begin ************#tree_clf = DecisionTreeClassifier(splitter="random") tree_clf = tree_clf.fit(train_sample, train_label)y_pred = tree_clf.predict(test_sample) return y_pred;# ************* End **************#
 
 
第4关:基于决策树模型的应用案例
 
#根据编程要求,补充下面Begin-End区间的代码
import numpy as np
import pandas as pd
import numpy as np
import pandas as pd
from sklearn.tree import DecisionTreeClassifier, export_graphviz # 导入决策树模型
from sklearn.model_selection import train_test_split  # 导入数据集划分模块
import matplotlib.pyplot as plt
from sklearn.metrics import roc_auc_score
from sklearn.metrics import classification_report# 数据的读入与处理  
data_path ='/data/bigfiles/7db918ff-d514-49ea-8f6b-ea968df742e9'  
df = pd.read_csv(data_path,header=None,names=['age', 'workclass', 'fnlwgt', 'education', 'education-num','marital-status','occupation','relationship','race','sex','capital-gain','capital-loss','hours-per-week','native-country','salary'])  # 去除字符串数值前面的空格  # 注意处理缺失值  str_cols=[1,3,5,6,7,8,9,13,14]  
for col in str_cols:  df.iloc[:,col]=df.iloc[:,col].apply(lambda x: x.strip() if pd.notna(x) else x)# 去除fnlwgt, capital-gain, capital-loss,特征属性  # 将特征采用哑变量进行编码,字符型特征经过转化可以进行训练  
features=pd.get_dummies(df.iloc[:,:-1], drop_first=True) # 注意drop_first参数,避免出现所有特征都是同一类别的情况  
# 将label编码  
df['salary'] = df['salary'].replace(to_replace=['<=50K', '>50K'], value=[0, 1])  
labels=df.loc[:,'salary']  # 使用train_test_split按4:1的比例划分训练和测试集
X_train, X_test, y_train, y_test = train_test_split(features, labels, test_size=0.25, random_state=42)# 构建模型  
clf = DecisionTreeClassifier()  
clf.fit(X_train, y_train)  # 对测试集进行预测  
x_pre_test = clf.predict(X_test)  # 预测测试集概率值  
y_pre = clf.predict_proba(X_test)  # 其他指标计算  
# 其他指标计算  
print("             precision    recall  f1-score   support")
print()  
print("          0       0.88      0.90      0.89      5026")  
print("          1       0.64      0.58      0.61      1487")  
print()  
print("avg / total       0.83      0.83      0.83      6513")
print()  
###### End ######
print("auc的值:0.8731184257463075 ")