Derivative Hadamard Product
Answer
Answers can only be viewed under the following conditions:
- The questioner was satisfied with and accepted the answer, or
- The answer was evaluated as being 100% correct by the judge.
1 Attachment
3.5K
The answer is accepted.
Join Matchmaticians Affiliate Marketing
Program to earn up to a 50% commission on every question that your affiliated users ask or answer.
- answered
- 1235 views
- $58.00
Related Questions
- Rouche’s Theorem applied to the complex valued function $f(z) = z^6 + \cos z$
- Calculating the residues at the poles of $f(z) = \frac{\tan(z) }{z^2 + z +1} $
- Find $x$ so that $\begin{bmatrix} 2 & 0 & 10 \\ 0 & x+7 & -3 \\ 0 & 4 & x \end{bmatrix} $ is invertible
- Gauss-Legendre quadrature rule
- Compute $$\oint_{|z-2|=2} \frac{\cos e^z}{z^2-4}dz$$
- Prove that $\int_{-\infty}^{\infty}\frac{\cos ax}{x^4+1}dx=\frac{\pi}{2}e^{-\frac{a}{\sqrt{2}}}(\cos \frac{a}{\sqrt{2}}+\sin \frac{a}{\sqrt{2}} )$
- Show that $tr(\sqrt{\sqrt A B \sqrt A})\leq 1$ , where both $A$ and $B$ are positive semidefinite with $tr(A)=tr(B)=1.$
- Find $a,b,c$ so that $\begin{bmatrix} 0 & 1& 0 \\ 0 & 0 & 1\\ a & b & c \end{bmatrix} $ has the characteristic polynomial $-\lambda^3+4\lambda^2+5\lambda+6=0$
What do you mean with the derivative wrt W?
I did d some research and it seems to me that there is no commonly accepted definition. Please provide one.
The term is part of the backpropagation of my neural network. I need to calculate this derivative to update my weights in the neural network. Hope this helps.
The hadamard product is the elemt-wise product of two matrices of the same dimensionality https://en.m.wikipedia.org/wiki/Hadamard_product_(matrices)
One can view the product as a function from R^(|V|\times |d|) to R^(|d|). So the derivative may mean just the matrix of partial derivatives. Do you expect the result to be a a |d| by ( |V|\times |d| ) matrix?
a and b must be row vectors for this product to make sense.
Ideally the result would be in the dimenson of |V|
The preferred notation for a vector is as a column vector, so ideally aW would be written as a^t W , meaning a transpose times W. will keep with the notation you used, so that aW is a row vector times a matrix, but this is not a very common notation.
I will keep with the notation you used, so that aW is a row vector times a matrix, but this is not a very common notation*
I am sorry for any inconvenience with the notation. Of course, a and b would be transposed.