-
relu单元实现
def relu(x):
return np.max(x,0)
weight=np.array([-0.21,0.3,0.7])
np.dot([1,0,1], weight)
-
与AND,或OR,非NOT,与非NAND,异或XOR
-
softmax:计算出一个概率分布向量,所有分量之和为1
-
logit:与sigmoid互为反函数
-
MSE:自己写的版本
def MSE(x, y): return np.mean(np.square(x-y))/x.size a=np.array([72,94,79,83,65,82,73,67,85,82]) b=80 MSE(a,b)
-
CE:
def CE(x,y): return (-np.dot(x,np.log(y)))
This repository has been archived by the owner on Sep 5, 2021. It is now read-only.