Reduced Gradient

14,000,000 Leading Edge Experts on the ideXlab platform

Scan Science and Technology

Contact Leading Edge Experts & Companies

Scan Science and Technology

Contact Leading Edge Experts & Companies

The Experts below are selected from a list of 246 Experts worldwide ranked by ideXlab platform

Quanquan Gu - One of the best experts on this subject based on the ideXlab platform.

  • NeurIPS - Stochastic Nested Variance Reduced Gradient Descent for Nonconvex Optimization
    2018
    Co-Authors: Dongruo Zhou, Pan Xu, Quanquan Gu
    Abstract:

    We study finite-sum nonconvex optimization problems, where the objective function is an average of n nonconvex functions. We propose a new stochastic Gradient descent algorithm based on nested variance reduction. Compared with conventional stochastic variance Reduced Gradient (SVRG) algorithm that uses two reference points to construct a semi-stochastic Gradient with diminishing variance in each epoch, our algorithm uses K+1 nested reference points to build an semi-stochastic Gradient to further reduce its variance in each epoch. For smooth functions, the proposed algorithm converges to an approximate first order stationary point (i.e., ‖∇F(\xb)‖2≤ϵ) within \tO(n∧ϵ−2+ϵ−3∧n1/2ϵ−2)\footnote{\tO(⋅) hides the logarithmic factors} number of stochastic Gradient evaluations, where n is the number of component functions, and ϵ is the optimization error. This improves the best known Gradient complexity of SVRG O(n+n2/3ϵ−2) and the best Gradient complexity of SCSG O(ϵ−5/3∧n2/3ϵ−2). For Gradient dominated functions, our algorithm achieves \tO(n∧τϵ−1+τ⋅(n1/2∧(τϵ−1)1/2) Gradient complexity, which again beats the existing best Gradient complexity \tO(n∧τϵ−1+τ⋅(n1/2∧(τϵ−1)2/3) achieved by SCSG. Thorough experimental results on different nonconvex optimization problems back up our theory.

  • stochastic nested variance Reduced Gradient descent for nonconvex optimization
    Neural Information Processing Systems, 2018
    Co-Authors: Dongruo Zhou, Pan Xu, Quanquan Gu
    Abstract:

    We study finite-sum nonconvex optimization problems, where the objective function is an average of n nonconvex functions. We propose a new stochastic Gradient descent algorithm based on nested variance reduction. Compared with conventional stochastic variance Reduced Gradient (SVRG) algorithm that uses two reference points to construct a semi-stochastic Gradient with diminishing variance in each epoch, our algorithm uses K+1 nested reference points to build an semi-stochastic Gradient to further reduce its variance in each epoch. For smooth functions, the proposed algorithm converges to an approximate first order stationary point (i.e., ‖∇F(\xb)‖2≤ϵ) within \tO(n∧ϵ−2+ϵ−3∧n1/2ϵ−2)\footnote{\tO(⋅) hides the logarithmic factors} number of stochastic Gradient evaluations, where n is the number of component functions, and ϵ is the optimization error. This improves the best known Gradient complexity of SVRG O(n+n2/3ϵ−2) and the best Gradient complexity of SCSG O(ϵ−5/3∧n2/3ϵ−2). For Gradient dominated functions, our algorithm achieves \tO(n∧τϵ−1+τ⋅(n1/2∧(τϵ−1)1/2) Gradient complexity, which again beats the existing best Gradient complexity \tO(n∧τϵ−1+τ⋅(n1/2∧(τϵ−1)2/3) achieved by SCSG. Thorough experimental results on different nonconvex optimization problems back up our theory.

  • SVRG-LD$^+$: Subsampled Stochastic Variance-Reduced Gradient Langevin Dynamics
    2018
    Co-Authors: Pan Xu, Quanquan Gu
    Abstract:

    Stochastic variance-Reduced Gradient Langevin dynamics (SVRG-LD) was recently proposed to improve the performance of stochastic Gradient Langevin dynamics (SGLD) by reducing the variance in the stochastic Gradient. In this paper, we study a variant of SVRG-LD, namely SVRG-LD$^+$, which replaces the full Gradient in each epoch by a subsampled one. We provide a nonasymptotic analysis of the convergence of SVRG-LD$^+$ in $2$-Wasserstein distance, and show that SVRG-LD$^+$ enjoys a lower Gradient complexity than SVRG-LD, when the sample size is large or the target accuracy requirement is moderate. Our analysis directly implies a sharper convergence rate for SVRG-LD, which improves the existing convergence rate by a factor of $\kappa^{1/6}n^{1/6}$, where $\kappa$ is the condition number of the log-density function and $n$ is the sample size. Experiments on both synthetic and real-world datasets validate our theoretical results.

  • Stochastic Variance-Reduced Gradient Descent for Low-rank Matrix Recovery from Linear Measurements
    arXiv: Machine Learning, 2017
    Co-Authors: Xiao Zhang, Lingxiao Wang, Quanquan Gu
    Abstract:

    We study the problem of estimating low-rank matrices from linear measurements (a.k.a., matrix sensing) through nonconvex optimization. We propose an efficient stochastic variance Reduced Gradient descent algorithm to solve a nonconvex optimization problem of matrix sensing. Our algorithm is applicable to both noisy and noiseless settings. In the case with noisy observations, we prove that our algorithm converges to the unknown low-rank matrix at a linear rate up to the minimax optimal statistical error. And in the noiseless setting, our algorithm is guaranteed to linearly converge to the unknown low-rank matrix and achieves exact recovery with optimal sample complexity. Most notably, the overall computational complexity of our proposed algorithm, which is defined as the iteration complexity times per iteration time complexity, is lower than the state-of-the-art algorithms based on Gradient descent. Experiments on synthetic data corroborate the superiority of the proposed algorithm over the state-of-the-art algorithms.

Jungmin Hwang - One of the best experts on this subject based on the ideXlab platform.

Abdelkrim El Mouatasim - One of the best experts on this subject based on the ideXlab platform.

  • Nesterov Step Reduced Gradient Algorithm for Convex Programming Problems
    Big Data and Networks Technologies, 2019
    Co-Authors: Abdelkrim El Mouatasim, Yousef Farhaoui
    Abstract:

    In this paper, we proposed an implementation of method of speed Reduced Gradient algorithm for optimizing a convex differentiable function subject to linear equality constraints and nonnegativity bounds on the variables. In particular, at each iteration, we compute a search direction by Reduced Gradient, and line search by bisection algorithm or Armijo rule. Under some assumption, the convergence rate of speed Reduced Gradient (SRG) algorithm is proven to be significantly better, both theoretically and practically. The algorithm of SRG are programmed by Matlab, and comparing by Frank-Wolfe algorithm some problems, the numerical results which show the efficient of our approach, we give also an application to ODE, optimal control, image and video co-localization and learning machine.

  • implementation of Reduced Gradient with bisection algorithms for non convex optimization problem via stochastic perturbation
    Numerical Algorithms, 2018
    Co-Authors: Abdelkrim El Mouatasim
    Abstract:

    In this paper, we proposed an implementation of stochastic perturbation of Reduced Gradient and bisection (SPRGB) method for optimizing a non-convex differentiable function subject to linear equality constraints and non-negativity bounds on the variables. In particular, at each iteration, we compute a search direction by Reduced Gradient, and optimal line search by bisection algorithm along this direction yields a decrease in the objective value. SPRGB method is desired to establish the global convergence of the algorithm. An implementation and tests of SPRGB algorithm are given, and some numerical results of large-scale problems are presented, which show the efficient of this approach.

  • stochastic perturbation of Reduced Gradient grg methods for nonconvex programming problems
    Applied Mathematics and Computation, 2014
    Co-Authors: Abdelkrim El Mouatasim, Rachid Ellaia, Eduardo Souza De Cursi
    Abstract:

    In this paper, we consider nonconvex differentiable programming under linear and nonlinear differentiable constraints. A Reduced Gradient and GRG (generalized Reduced Gradient) descent methods involving stochastic perturbation are proposed and we give a mathematical result establishing the convergence to a global minimizer. Numerical examples are given in order to show that the method is effective to calculate. Namely, we consider classical tests such as the statistical problem, the octagon problem, the mixture problem and an application to the linear optimal control servomotor problem.

  • Stochastic perturbation of Reduced Gradient & GRG methods for nonconvex programming problems
    Applied Mathematics and Computation, 2014
    Co-Authors: Abdelkrim El Mouatasim, Rachid Ellaia, Eduardo Souza De Cursi
    Abstract:

    In this paper, we consider nonconvex differentiable programming under linear and nonlinear differentiable constraints. A Reduced Gradient and GRG (generalized Reduced Gradient) descent methods involving stochastic perturbation are proposed and we give a mathematical result establishing the convergence to a global minimizer. Numerical examples are given in order to show that the method is effective to calculate. Namely, we consider classical tests such as the statistical problem, the octagon problem, the mixture problem and an application to the linear optimal control servomotor problem.

  • Two-Phase Generalized Reduced Gradient Method for Constrained Global Optimization
    Journal of Applied Mathematics, 2010
    Co-Authors: Abdelkrim El Mouatasim
    Abstract:

    The random perturbation of generalized Reduced Gradient method for optimization under nonlinear differentiable constraints is proposed. Generally speaking, a particular iteration of this method proceeds in two phases. In the Restoration Phase, feasibility is restored by means of the resolution of an auxiliary nonlinear problem, a generally nonlinear system of equations. In the Optimization Phase, optimality is improved by means of the consideration of the objective function, on the tangent subspace to the constraints. In this paper, optimal assumptions are stated on the Restoration Phase and the Optimization Phase that establish the global convergence of the algorithm. Some numerical examples are also given by mixture problem and octagon problem.

Malgorzata P Kaleta - One of the best experts on this subject based on the ideXlab platform.

  • Efficient Solution of the Optimization Problem in Model-Reduced Gradient-based History Matching
    ECMOR XIII - 13th European Conference on the Mathematics of Oil Recovery, 2012
    Co-Authors: Slawomir P. Szklarz, Marielba Rojas, Malgorzata P Kaleta
    Abstract:

    Adjusting parameters in reservoir models by minimizing the discrepancy between the model's predictions and actual measurements is a popular approach known as history matching. One of the most effective techniques is Gradient-based history matching. For reservoir models, the number of grid blocks and therefore, the size of the problem can become very large. In recent years, model-order reduction techniques aiming to replace large, complex dynamic systems with lower-dimension models have been incorporated into history matching. In both Gradient-based history matching and model-Reduced Gradient-based history matching, first-order optimization methods are used in order to minimize the mismatch between simulated well-production data and observed production. In this work, we investigate the performance of some optimization methods on the minimization problem in model-Reduced Gradient-based history matching. The methods were tested on the history matching of a small reservoir model with synthetic measurements. Our results show that fast first-order techniques such as the spectral projected Gradient method can compete with the popular quasi-Newton BFGS approach.

  • The optimization problem in model-Reduced Gradient-based history matching
    IFAC Proceedings Volumes, 2012
    Co-Authors: Slawomir P. Szklarz, Marielba Rojas, Malgorzata P Kaleta
    Abstract:

    Abstract We present preliminary results of a performance evaluation study of several Gradient-based state-of-the-art optimization methods for solving the nonlinear minimization problem arising in model-Reduced Gradient-based history matching. The issues discussed also apply to other areas, such as production optimization in closed-loop reservoir management.

  • model Reduced Gradient based history matching
    Computational Geosciences, 2011
    Co-Authors: Malgorzata P Kaleta, R G Hanea, A W Heemink, J D Jansen
    Abstract:

    Gradient-based history matching algorithms can be used to adapt the uncertain parameters in a reservoir model using production data. They require, however, the implementation of an adjoint model to compute the Gradients, which is usually an enormous programming effort. We propose a new approach to Gradient-based history matching which is based on model reduction, where the original (nonlinear and high-order) forward model is replaced by a linear Reduced-order forward model and, consequently, the adjoint of the tangent linear approximation of the original forward model is replaced by the adjoint of a linear Reduced-order forward model. The Reduced-order model is constructed with the aid of the proper orthogonal decomposition method. Due to the linear character of the Reduced model, the corresponding adjoint model is easily obtained. The Gradient of the objective function is approximated, and the minimization problem is solved in the Reduced space; the procedure is iterated with the updated estimate of the parameters if necessary. The proposed approach is adjoint-free and can be used with any reservoir simulator. The method was evaluated for a waterflood reservoir with channelized permeability field. A comparison with an adjoint-based history matching procedure shows that the model-Reduced approach gives a comparable quality of history matches and predictions. The computational efficiency of the model-Reduced approach is lower than of an adjoint-based approach, but higher than of an approach where the Gradients are obtained with simple finite differences.

Dongruo Zhou - One of the best experts on this subject based on the ideXlab platform.

  • stochastic nested variance Reduced Gradient descent for nonconvex optimization
    Neural Information Processing Systems, 2018
    Co-Authors: Dongruo Zhou, Pan Xu, Quanquan Gu
    Abstract:

    We study finite-sum nonconvex optimization problems, where the objective function is an average of n nonconvex functions. We propose a new stochastic Gradient descent algorithm based on nested variance reduction. Compared with conventional stochastic variance Reduced Gradient (SVRG) algorithm that uses two reference points to construct a semi-stochastic Gradient with diminishing variance in each epoch, our algorithm uses K+1 nested reference points to build an semi-stochastic Gradient to further reduce its variance in each epoch. For smooth functions, the proposed algorithm converges to an approximate first order stationary point (i.e., ‖∇F(\xb)‖2≤ϵ) within \tO(n∧ϵ−2+ϵ−3∧n1/2ϵ−2)\footnote{\tO(⋅) hides the logarithmic factors} number of stochastic Gradient evaluations, where n is the number of component functions, and ϵ is the optimization error. This improves the best known Gradient complexity of SVRG O(n+n2/3ϵ−2) and the best Gradient complexity of SCSG O(ϵ−5/3∧n2/3ϵ−2). For Gradient dominated functions, our algorithm achieves \tO(n∧τϵ−1+τ⋅(n1/2∧(τϵ−1)1/2) Gradient complexity, which again beats the existing best Gradient complexity \tO(n∧τϵ−1+τ⋅(n1/2∧(τϵ−1)2/3) achieved by SCSG. Thorough experimental results on different nonconvex optimization problems back up our theory.

  • NeurIPS - Stochastic Nested Variance Reduced Gradient Descent for Nonconvex Optimization
    2018
    Co-Authors: Dongruo Zhou, Pan Xu, Quanquan Gu
    Abstract:

    We study finite-sum nonconvex optimization problems, where the objective function is an average of n nonconvex functions. We propose a new stochastic Gradient descent algorithm based on nested variance reduction. Compared with conventional stochastic variance Reduced Gradient (SVRG) algorithm that uses two reference points to construct a semi-stochastic Gradient with diminishing variance in each epoch, our algorithm uses K+1 nested reference points to build an semi-stochastic Gradient to further reduce its variance in each epoch. For smooth functions, the proposed algorithm converges to an approximate first order stationary point (i.e., ‖∇F(\xb)‖2≤ϵ) within \tO(n∧ϵ−2+ϵ−3∧n1/2ϵ−2)\footnote{\tO(⋅) hides the logarithmic factors} number of stochastic Gradient evaluations, where n is the number of component functions, and ϵ is the optimization error. This improves the best known Gradient complexity of SVRG O(n+n2/3ϵ−2) and the best Gradient complexity of SCSG O(ϵ−5/3∧n2/3ϵ−2). For Gradient dominated functions, our algorithm achieves \tO(n∧τϵ−1+τ⋅(n1/2∧(τϵ−1)1/2) Gradient complexity, which again beats the existing best Gradient complexity \tO(n∧τϵ−1+τ⋅(n1/2∧(τϵ−1)2/3) achieved by SCSG. Thorough experimental results on different nonconvex optimization problems back up our theory.