Research Article
[Retracted] Gradient Descent Optimization in Deep Learning Model Training Based on Multistage and Method Combination Strategy
Figure 8
Combined strategy versus noncombined strategy (2). A combined strategy of SGD with Momentum, Adam, and RMSprop with cosine decay versus cosine decay. (a, d) The comparison results of SGD with Momentum and cosine decay, (b, e) the comparison results of Adam and cosine decay, and (c, f) the comparison results of RMSprop and cosine decay.
(a) |
(b) |
(c) |
(d) |
(e) |
(f) |