Xavier:
The basic idea is that through the network layer , The variance of input and output is the same , Including forward propagation and backward propagation .
If the initialization value is very small , So with the transfer of layers , Variance tends to 0, Enter the value It's getting smaller and smaller , stay sigmoid It's in the 0 near , Close to linear , Without nonlinearity
If the initial value is large , So with the transfer of layers , The variance will increase rapidly , At this point, the input value becomes large , and sigmoid Writing a reciprocal at a large input tends to 0, Back propagation will encounter the problem of gradient disappearance
Sense and BN The purpose of using is similar
np.newaxis: Insert new dimension
list and ndarray Interturn :
list turn numpy:np.array(a)
ndarray turn list:a.tolist()
numpy.ndarray And string to string :
str=arr.tostring()
arr=np.frombuffer(string, dtype=np.float32)
pydub Learning reference for :https://blog.csdn.net/Debatrix/article/details/59058762
pyworld The use of reference :https://blog.csdn.net/m0_43395719/article/details/107930075
Im the son of Science . stay p
1. Preface My home is in the