神經網路實現連續型變數的迴歸預測(python)

來源:互聯網
上載者:User

標籤:dde   python   啟用   int   網路   變數   mat   layer   sel   

 

轉至:50488727

輸入資料變為樓價預測:

105.0,2,0.89,510.0
105.0,2,0.89,510.0
138.0,3,0.27,595.0
135.0,3,0.27,596.0
106.0,2,0.83,486.0
105.0,2,0.89,510.0
105.0,2,0.89,510.0
143.0,3,0.83,560.0
108.0,2,0.91,450.0

 

 

最近寫論文時用到一個方法,是基於神經網路的最優組合預測,主要思想如下:在建立由迴歸模型、灰色預測模型、BP神經網路預測模型組成的組合預測模型庫的基礎上,利用以上三種單一預測模型的組合構成BP神經網路混合預測模型。(我是參考的參考這篇文章:路玉龍,韓靖,餘思婧,張鴻雁.BP神經網路混合預測在城市生活垃圾產量預測中應用)

我的目的

我需要用BP神經網路來做連續預測。關於BP神經網路的python實現網上有很多,但大多是用於分類決策,於是不得不搞清楚原理改代碼。 
以下是我參考的一篇BP神經網路的分類決策的實現(我的連續預測的代碼是基於下面這個連結改的,在此致謝一下): 
https://www.cnblogs.com/Finley/p/5946000.html

修改思路:

(1)最後一層不啟用,直接輸出。或者說把啟用函數看作f(x)=x 
(2)損失函數函數改為MSE

代碼

代碼中用兩個#——-圍起來的就是我更正的部分。

import math
import random

random.seed(0)
def rand(a, b):
return (b - a) * random.random() + a

def make_matrix(m, n, fill=0.0):
mat = []
for i in range(m):
mat.append([fill] * n)
return mat

def sigmoid(x):
return 1.0 / (1.0 + math.exp(-x))

def sigmoid_derivative(x):
return x * (1 - x)

class BPNeuralNetwork:
def __init__(self):
self.input_n = 0
self.hidden_n = 0
self.output_n = 0
self.input_cells = []
self.hidden_cells = []
self.output_cells = []
self.input_weights = []
self.output_weights = []
self.input_correction = []
self.output_correction = []

def setup(self, ni, nh, no):
self.input_n = ni + 1
self.hidden_n = nh
self.output_n = no
# init cells
self.input_cells = [1.0] * self.input_n
self.hidden_cells = [1.0] * self.hidden_n
self.output_cells = [1.0] * self.output_n
# init weights
self.input_weights = make_matrix(self.input_n, self.hidden_n)
self.output_weights = make_matrix(self.hidden_n, self.output_n)
# random activate
for i in range(self.input_n):
for h in range(self.hidden_n):
self.input_weights[i][h] = rand(-0.2, 0.2)
for h in range(self.hidden_n):
for o in range(self.output_n):
self.output_weights[h][o] = rand(-2.0, 2.0)
# init correction matrix
self.input_correction = make_matrix(self.input_n, self.hidden_n)
self.output_correction = make_matrix(self.hidden_n, self.output_n)

def predict(self, inputs):
# activate input layer
for i in range(self.input_n - 1):
self.input_cells[i] = inputs[i]#輸入層輸出值
# activate hidden layer
for j in range(self.hidden_n):
total = 0.0
for i in range(self.input_n):
total += self.input_cells[i] * self.input_weights[i][j]#隱藏層輸入值
self.hidden_cells[j] = sigmoid(total)#隱藏層的輸出值
# activate output layer
for k in range(self.output_n):
total = 0.0
for j in range(self.hidden_n):
total += self.hidden_cells[j] * self.output_weights[j][k]
#-----------------------------------------------
# self.output_cells[k] = sigmoid(total)
self.output_cells[k] =total#輸出層的激勵函數是f(x)=x
#-----------------------------------------------
return self.output_cells[:]

def back_propagate(self, case, label, learn, correct):#x,y,修改最大迭代次數, 學習率λ, 矯正率μ三個參數.
# feed forward
self.predict(case)
# get output layer error
output_deltas = [0.0] * self.output_n
for o in range(self.output_n):
error = label[o] - self.output_cells[o]
#-----------------------------------------------
# output_deltas[o] = sigmoid_derivative(self.output_cells[o]) * error
output_deltas[o] = error
#-----------------------------------------------
# get hidden layer error
hidden_deltas = [0.0] * self.hidden_n
for h in range(self.hidden_n):
error = 0.0
for o in range(self.output_n):
error += output_deltas[o] * self.output_weights[h][o]
hidden_deltas[h] = sigmoid_derivative(self.hidden_cells[h]) * error

# update output weights
for h in range(self.hidden_n):
for o in range(self.output_n):
change = output_deltas[o] * self.hidden_cells[h]
self.output_weights[h][o] += learn * change + correct * self.output_correction[h][o]#??????????
self.output_correction[h][o] = change

# update input weights
for i in range(self.input_n):
for h in range(self.hidden_n):
change = hidden_deltas[h] * self.input_cells[i]
self.input_weights[i][h] += learn * change + correct * self.input_correction[i][h]
self.input_correction[i][h] = change
# get global error
error = 0.0
for o in range(len(label)):
error += 0.5 * (label[o] - self.output_cells[o]) ** 2
return error

def train(self, cases, labels, limit=10000, learn=0.05, correct=0.1):
for j in range(limit):
error = 0.0
for i in range(len(cases)):
label = labels[i]
case = cases[i]
error += self.back_propagate(case, label, learn, correct)

def test(self):
cases = [
[10.5,2,0.89],
[10.5,2,0.89],
[13.8,3,0.27],
[13.5,3,0.27],
]
labels = [[0.51], [0.51], [0.595], [0.596]]
self.setup(3, 5, 1)
self.train(cases, labels, 10000, 0.05, 0.1)
for case in cases:
print(self.predict(case))

if __name__ == ‘__main__‘:
nn = BPNeuralNetwork()
nn.test()

實驗結果:

[0.5095123779256603]
[0.5095123779256603]
[0.5952606219141522]
[0.5939697670509705]

 

神經網路實現連續型變數的迴歸預測(python)

相關文章

聯繫我們

該頁面正文內容均來源於網絡整理,並不代表阿里雲官方的觀點,該頁面所提到的產品和服務也與阿里云無關,如果該頁面內容對您造成了困擾,歡迎寫郵件給我們,收到郵件我們將在5個工作日內處理。

如果您發現本社區中有涉嫌抄襲的內容,歡迎發送郵件至: info-contact@alibabacloud.com 進行舉報並提供相關證據,工作人員會在 5 個工作天內聯絡您,一經查實,本站將立刻刪除涉嫌侵權內容。

A Free Trial That Lets You Build Big!

Start building with 50+ products and up to 12 months usage for Elastic Compute Service

  • Sales Support

    1 on 1 presale consultation

  • After-Sales Support

    24/7 Technical Support 6 Free Tickets per Quarter Faster Response

  • Alibaba Cloud offers highly flexible support services tailored to meet your exact needs.