Citycluster[label[i]].append (Cityname[i]) #将每个簇的城市输出For I in range (len (citycluster)):Print ("expenses:%.2f"% expenses[i]) #将每个簇的平均花费输出Print (Citycluster[i])Click to run, you can come out results.Where the N_clusters class, the consumption level of similar cities gathered in a classExpense: The numerical plus of the central point of the cluster, that is, the average consumption levelImplementation process:1, establish the project, import Sklearn related packageImport NumPy as NPFrom Sklearn.cl
)]=1 else:print "The word:%s is not in my vocabulary!" %word return returnvecdef TRAINNBC (trainsamples,traincategory): Numtrainsamp=len (Trainsamples) NumWords=len (train Samples[0]) pabusive=sum (traincategory)/float (numtrainsamp) #y =1 or 0 feature Count P0num=np.ones (numwords) P1NUM=NP.O NES (numwords) #y =1 or 0 category count P0numtotal=numwords p1numtotal=numwords for I in Range (Numtrainsamp): if Traincategory[i]==1:p0num+=trainsamples[i] P0numtotal+=sum (Trainsamples[i]) E
attribute in the data set. The general situation is somewhere between the two.D. High-dimensional mappingMap properties to high-dimensional space. This is the most precise approach, which completely retains all the information and does not add any additional information. For example, Google, Baidu's CTR Prediction model, pre-processing will be all the variables to deal with this, up to hundreds of millions of dimensions. The benefit of this is that the entire information of the original data is
:", X) - Print("Y:", Y) - innumiterations=100000 -alpha=0.0005 toTheta=np.ones (x.shape[1]) +Theta=graientdescent (x,y,theta,alpha,x.shape[0],numiterations) - Print(Theta)Operation Result:...... Too many output data to intercept only the next more than 10 linesIteration 99988/cost:3.930135Iteration 99989/cost:3.930135Iteration 99990/cost:3.930135Iteration 99991/cost:3.930135Iteration 99992/cost:3.930135Iteration 99993/cost:3.930135Iteration 99994/cost:3.930135Iteration 99995/cost:3.930135Iterat
* (XMAT.T * (Weights *Ymat)) returnTestPoint *SigmadefLwlrtest (Testarr,xarr,yarr,k = 1.0): M=shape (Testarr) [0] Yhat=zeros (m) forIinchRange (m): Yhat[i]=LWLR (testarr[i],xarr,yarr,k)returnYhatThe LWLR () function is the code for locally weighted linear regression, and the function of the lwlrtest () function is to make the LWLR () function traverse the entire data set. We also need to draw a picture to see how the results fit. def PlotLine1 (testarr,xarr,yarr,k = 1.0 = Mat (Xarr) ymat = Ma
Python3 Learning using the APIUsing the data set on the Internet, I downloaded him to a localcan download datasets in my git: https://github.com/linyi0604/MachineLearningCode:1 ImportNumPy as NP2 ImportPandas as PD3 fromSklearn.clusterImportKmeans4 fromSklearnImportMetrics5 6 " "7 K-Mean-value algorithm:8 1 randomly selected K samples as the center of the K category9 2 from the K sample, select the nearest sample to be the same category as yourself,
called the polynomial model, but its class conditional probability calculation formula is not accurate.Referencesalgorithm Grocer--naive Bayesian classification of classification algorithm (Naive Bayesian classification)study of naive Bayesian text classification algorithmThe author of this paper, Adan, derives from: The classical algorithm of machine learning and the implementation of
, or K nearest neighbor (Knn,k-nearestneighbor) classification algorithm, is one of the simplest methods in data mining classification technology. The so-called K nearest neighbor is the meaning of K's closest neighbour, saying that each sample can be represented by its nearest K-neighbor.The core idea of the KNN algorithm is that if the majority of the k nearest samples in a feature space belong to a category, the sample also falls into this category and has the characteristics of the sample on
!accuracy:87.07%******************* SVM ********************Training took3831. 564000s!accuracy:94.35%******************* GBDT ********************In this data set, because the cluster of data distribution is better (if you understand this database, see its T-sne map can be seen.) Since the task is simple, it has been considered a toy dataset in the deep learning boundary, so KNN has a good effect. GBDT is a very good algorithm, in Kaggle and other bi
Python3 Learning using the APIA sample of a data structure of a dictionary type, extracting features and converting them into vector formSOURCE Git:https://github.com/linyi0604/machinelearningCode:1 fromSklearn.feature_extractionImportDictvectorizer2 3 " "4 dictionary feature Extractor:5 pumping and vectorization of dictionary data Structures6 category type features vectorization with 0 12 values using prototype feature names7 numeric type features r
=true) # drop useless columns and create LABELSIDX = test.id.values.astype (int) test = Test.drop ([' id ', ' tube_assembly_id ', ' quote_date '), Axis = 1) labels = Train.cost.valuestrain = Train.drop ([' Quote_date ' , ' cost ', ' tube_assembly_id '], Axis = 1) # Convert data to NumPy Arraytrain = Np.array (train) test = Np.array (test)From:kaggle Copyright NOTICE: This article for Bo Master original article, without Bo Master permission not reproduced. Ma
Efficiency comparison:It's a cliché, but this time with a new module,Run Time Test Module Timeti:1 ImportTimeit2 3normal = Timeit.timeit ('sum (x*x for x in range )', number=10000)4NATIVE_NP = Timeit.timeit ('sum (na*na)',#Repeating part5setup="import numpy as np; na = Np.arange (+)",#Setup runs only once6number=10000)#Number of repetitions7GOOD_NP = Timeit.timeit ('Na.dot (NA)',8setup="import numpy as np; na = Np.arange (+)",9number=10000)Ten One Print('Native Run time:', Normal,'\ n', A
90avg/total 0.82 0.78 0.79 329The accuracy of gradient tree boosting is 0.790273556231 Precision recall f1-score support 0 0.92 0.78 0.84 239 1 0.58 0.82 0.68 90avg/total 0.83 0.79 0.80 329Conclusion:Predictive performance: The gradient rise decision tree is larger than the random forest classifier larger than the single decision tree. The industry often uses the stochastic forest c
#数据预处理方法, mainly dealing with the dimension of data and the problem of the same trend.Import NumPy as NPFrom Sklearn Import preprocessing#零均值规范Data=np.random.rand (3,4) #随机生成3行4列的数据Data_standardized=preprocessing.scale (data) #对数据进行归一化处理, that is, each value minus the mean divided by the variance is primarily used for SVM#线性数据变换最大最小化处理Data_scaler=preprocessing. Minmaxscaler (feature_range= (0,1)) #选定区间 (0,1), raw Data-min/(max-min)Data_scaled=data_scaler.fit (data)#数据标准化处理normalizeddata_normaliz
The language used for machine learning is python. Here's how to get started with Python for "machine learning." (Environment: CentOS 7)1, two important packagesNumPy and SciPy. (http://scipy.org/scipylib/download.html) mainly deal
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.