1.Summary: Apply the chain rule to compute the gradient of the loss function with respect to the inputs. ----cs231n 2.what problems to slove? 2.1introduction 神经网络的本质是一个多层的复合函数,图: 表达式为: 上面式中的Wij就是相邻两层神经元之间的权值,它们就是深度学习需要学习的参数,也就相当于直线拟合y=k*x+b中的待求参数k和b.…
Training LeNet on MNIST with Caffe We will assume that you have Caffe successfully compiled. If not, please refer to the Installation page. In this tutorial, we will assume that your Caffe installation is located at CAFFE_ROOT. Prepare Datasets You w…