Matlab中常用的分类器有随机森林分类器、支持向量机(SVM)、K近邻分类器、朴素贝叶斯、集成学习方法和鉴别分析分类器等。各分类器的相关Matlab函数使用方法如下:

首先对以下介绍中所用到的一些变量做统一的说明:

    train_data——训练样本,矩阵的每一行数据构成一个样本,每列表示一种特征

    train_label——训练样本标签,为列向量

    test_data——测试样本,矩阵的每一行数据构成一个样本,每列表示一种特征

    test_label——测试样本标签,为列向量

①随机森林分类器(Random Forest)

    TB=TreeBagger(nTree,train_data,train_label);

    predict_label=predict(TB,test_data);

②支持向量机(Support Vector Machine,SVM)

    SVMmodel=svmtrain(train_data,train_label);

    predict_label=svmclassify(SVMmodel,test_data);

③K近邻分类器(KNN)

    KNNmodel=ClassificationKNN.fit(train_data,train_label,\’NumNeighbors\’,1);

    predict_label=predict(KNNmodel,test_data);

④朴素贝叶斯(Naive Bayes)

    Bayesmodel=NaiveBayes.fit(train_data,train_label);

    predict_label=predict(Bayesmodel,test_data);

⑤集成学习方法(Ensembles for Boosting)

    Bmodel=fitensemble(train_data,train_label,\’AdaBoostM1\’,100,\’tree\’,\’type\’,\’classification\’);

    predict_label=predict(Bmodel,test_data);

⑥鉴别分析分类器(Discriminant Analysis Classifier)

    DACmodel=ClassificationDiscriminant.fit(train_data,train_label);

    predict_label=predict(DACmodel,test_data);

具体使用如下:(练习数据下载地址如下http://en.wikipedia.org/wiki/Iris_flower_data_set,简单介绍一下该数据集:有一批花可以分为3个品种,不同品种的花的花萼长度、花萼宽度、花瓣长度、花瓣宽度会有差异,根据这些特征实现品种分类)

%% 随机森林分类器(Random Forest)
nTree=10;
B=TreeBagger(nTree,train_data,train_label,\’Method\’, \’classification\’);
predictl=predict(B,test_data);
predict_label=str2num(cell2mat(predictl));
Forest_accuracy=length(find(predict_label == test_label))/length(test_label)*100;

%% 支持向量机
% SVMStruct = svmtrain(train_data, train_label);
% predictl=svmclassify(SVMStruct,test_data);
% predict_label=str2num(cell2mat(predictl));
% SVM_accuracy=length(find(predict_label == test_label))/length(test_label)*100;  

%% K近邻分类器(KNN)
% mdl = ClassificationKNN.fit(train_data,train_label,\’NumNeighbors\’,1);
% predict_label=predict(mdl, test_data);
% KNN_accuracy=length(find(predict_label == test_label))/length(test_label)*100

%% 朴素贝叶斯 (Naive Bayes)
% nb = NaiveBayes.fit(train_data, train_label);
% predict_label=predict(nb, test_data);
% Bayes_accuracy=length(find(predict_label == test_label))/length(test_label)*100;

%% 集成学习方法(Ensembles for Boosting, Bagging, or Random Subspace)
% ens = fitensemble(train_data,train_label,\’AdaBoostM1\’ ,100,\’tree\’,\’type\’,\’classification\’);
% predictl=predict(ens,test_data);
% predict_label=str2num(cell2mat(predictl));
% EB_accuracy=length(find(predict_label == test_label))/length(test_label)*100;

%% 鉴别分析分类器(discriminant analysis classifier)
% obj = ClassificationDiscriminant.fit(train_data, train_label);
% predictl=predict(obj,test_data);
% predict_label=str2num(cell2mat(predictl));
% DAC_accuracy=length(find(predict_label == test_label))/length(test_label)*100;

%% 练习
% meas=[0 0;2 0;2 2;0 2;4 4;6 4;6 6;4 6];
% [N n]=size(meas);
% species={\’1\’;\’1\’;\’1\’;\’1\’;\’-1\’;\’-1\’;\’-1\’;\’-1\’};
% ObjBayes=NaiveBayes.fit(meas,species);
% x=[3 3;5 5];
% result=ObjBayes.predict(x);

参考链接:https://blog.csdn.net/jisuanjiguoba/java/article/details/80004568

版权声明:本文为hechangchun原创文章,遵循 CC 4.0 BY-SA 版权协议,转载请附上原文出处链接和本声明。
本文链接:https://www.cnblogs.com/hechangchun/p/12758598.html