pub struct Adagrad {
pub learning_rate: f64,
pub accumulators: Tensor,
pub weight_decay: Option<f64>,
pub epsilon: Option<f64>,
}
Expand description
Adaptive Gradient (Adagrad):
Adapts the learning rate based on the history of gradients. Divides the learning rate by a running average of the magnitude of the gradients. This allows the learning rate to decrease for parameters that have consistently large gradients and increase for parameters that have consistently small gradients. Includes an option to apply weight decay regularization to the gradients.
Fields§
§learning_rate: f64
§accumulators: Tensor
§weight_decay: Option<f64>
§epsilon: Option<f64>
Implementations§
Trait Implementations§
Auto Trait Implementations§
impl Freeze for Adagrad
impl RefUnwindSafe for Adagrad
impl Send for Adagrad
impl Sync for Adagrad
impl Unpin for Adagrad
impl UnwindSafe for Adagrad
Blanket Implementations§
source§impl<T> BorrowMut<T> for Twhere
T: ?Sized,
impl<T> BorrowMut<T> for Twhere
T: ?Sized,
source§fn borrow_mut(&mut self) -> &mut T
fn borrow_mut(&mut self) -> &mut T
Mutably borrows from an owned value. Read more