pdf
pdf |
|---|
Golden-Ratio Step Size Heuristic for Discrete Optimization-Curve Convexity
Author: Kevin T.N
Affiliation: Independent Researcher
Contact: jkdkr2439@gmail.com
Date: January 11, 2026
What this repository is
This repository contains a technical note proposing a conservative step-size heuristic for gradient descent on convex, L-smooth objectives, evaluated under a discrete optimization-curve convexity diagnostic.
This is not a theorem-driven work.
It is a heuristic + diagnostic + reproducible evaluation protocol.
Core idea (plain language)
Gradient descent guarantees monotone decrease of the objective when
[
0 < \eta < \frac{2}{L},
]
but this does not guarantee that the sequence of objective values looks “convex” or smooth when viewed as a discrete curve.
This repo introduces a simple diagnostic: [ \Delta^2 g_k = g_{k-1} - 2g_k + g_{k+1}, \quad g_k = f(x_k), ] and empirically studies how often this quantity becomes negative under different step sizes.
Based on experiments, we propose: [ \eta \approx \frac{\varphi}{L}, \quad \varphi = \frac{1+\sqrt{5}}{2}, ] as a conservative, memorable engineering heuristic.
Explicit scope and non-claims (important)
What is claimed
- A well-defined discrete diagnostic on the optimization trajectory.
- An empirically conservative step-size heuristic relative to more aggressive normalized step sizes.
- A reproducible evaluation protocol.
What is not claimed
- ❌ No universal guarantee for all convex L-smooth functions.
- ❌ No claim that the golden ratio is optimal.
- ❌ No claim about convergence rates or global optimality.
- ❌ No replacement for classical convergence theory.
If you are looking for a theorem, this is not it.
Contents
paper.pdf
Golden-Ratio Step Size as a Conservative Heuristic for Discrete Optimization-Curve Convexity in Gradient Descent(optional)
paper.tex
LaTeX source for transparency and reuse.
Diagnostic definition
We say the optimization curve is discretely convex if: [ \Delta^2 g_k \ge 0 \quad \text{for all } k \ge 1. ]
In numerical experiments, a tolerance ε ≥ 0 is used, and a violation is recorded if:
[
\Delta^2 g_k < -\varepsilon.
]
This diagnostic:
- is trajectory-based, not a property of the objective function,
- is independent of monotone descent,
- does not imply faster convergence.
Heuristic rule
Recommended default:
η = 0.99 · φ / L
- Downloads last month
- 14