Xavier:
The basic idea is that through the network layer , The variance of input and output is the same , Including forward propagation and backward propagation .
If the initialization value is very small , So with the transfer of layers , Variance tends to 0, Enter the value It's getting smaller and smaller , stay sigmoid It's in the 0 near , Close to linear , Without nonlinearity
If the initial value is large , So with the transfer of layers , The variance will increase rapidly , At this point, the input value becomes large , and sigmoid Writing a reciprocal at a large input tends to 0, Back propagation will encounter the problem of gradient disappearance
Sense and BN The purpose of using is similar
np.newaxis: Insert new dimension
list and ndarray Interturn :
list turn numpy:np.array(a)
ndarray turn list:a.tolist()
numpy.ndarray And string to string :
str=arr.tostring()
arr=np.frombuffer(string, dtype=np.float32)
pydub Learning reference for :https://blog.csdn.net/Debatrix/article/details/59058762
pyworld The use of reference :https://blog.csdn.net/m0_43395719/article/details/107930075
When using the shutil module in Python, there is a permissionerror: [errno 13] permission denied error. How to solve it
Im the son of Science . stay p
The gospel of friends in the north is coming. I wrote a boiler temperature control system in Python, so I dont have to worry about it any more.
1. Preface My home is in the