Research (연구 관련)
Retargeting and Optimization
홍돌
2025. 2. 17. 12:06
And solving the linear equation system when the system matrix is a triangular matrix is very efficient!
# Normal equations:
(J^TJ)x = J^Ty
# This is exactly our Ax = b where:
A = J^TJ # symmetric positive definite!
b = J^Ty
# Solve using Cholesky:
A = LL^T # Cholesky decomposition
Ly = b # Forward substitution
L^Tx = y # Backward substitution
# For n×n system:
General matrix: O(n³) # Using Gaussian elimination
Triangular matrix: O(n²) # Using forward/backward substitution
So why is it a bad idea to do Gaussian Newton to optimize a neural network instead of gradient descent? When should and should not I use Gaussian Newton?