@@ -54,8 +54,8 @@ _apply_scale_bias(x, scale, bias) = x .* scale .+ bias
5454
5555Shared code path for all built-in norm functions.
5656
57- `μ` and `σ²` should be calculated on the fly using [`NNlib. norm_stats`](@ref),
58- or extracted from an existing collection such as [`NNlib. RunningStats`](@ref).
57+ `μ` and `σ²` should be calculated on the fly using [`norm_stats`](@ref),
58+ or extracted from an existing collection such as [`RunningStats`](@ref).
5959`bias` and `scale` are consistent with cuDNN and Flux.Scale.
6060We opt for `scale` over `weight` to avoid confusion with dense layers.
6161If the size of the statistics and affine parameters differ,
@@ -79,7 +79,7 @@ Contains running mean and variance estimates for stateful norm functions.
7979If the parameters are mutable, they will be updated in-place.
8080Otherwise, they will be replaced wholesale.
8181
82- See also [`NNlib. update_running_stats!`](@ref).
82+ See also [`update_running_stats!`](@ref).
8383"""
8484mutable struct RunningStats{M <: AbstractArray , V <: AbstractArray , MT <: Real }
8585 mean:: M
@@ -129,10 +129,10 @@ end
129129 reduce_dims) where {N}
130130
131131Performs a moving average update for layers with tracked statistics.
132- `μ` and `σ²` are the sample mean and variance, most likely from [`NNlib. norm_stats`](@ref).
133- `reduce_dims` should also match the `dims` argument of [`NNlib. norm_stats`](@ref).
132+ `μ` and `σ²` are the sample mean and variance, most likely from [`norm_stats`](@ref).
133+ `reduce_dims` should also match the `dims` argument of [`norm_stats`](@ref).
134134
135- See also [`NNlib. RunningStats`](@ref).
135+ See also [`RunningStats`](@ref).
136136"""
137137function update_running_stats! (stats:: RunningStats , x, μ, σ², reduce_dims:: Dims )
138138 V = eltype (σ²)
@@ -168,7 +168,7 @@ Normalizes `x` along the first `S` dimensions.
168168
169169For an additional learned affine transform, provide a `S`-dimensional `scale` and `bias`.
170170
171- See also [`NNlib. batchnorm`](@ref), [`NNlib. instancenorm`](@ref), and [`NNlib. groupnorm`](@ref).
171+ See also [`batchnorm`](@ref), [`instancenorm`](@ref), and [`groupnorm`](@ref).
172172
173173# Examples
174174
@@ -205,14 +205,14 @@ Functional [Batch Normalization](https://arxiv.org/abs/1502.03167) operation.
205205Normalizes `x` along each ``D_1×...×D_{N-2}×1×D_N`` input slice,
206206where `N-1` is the "channel" (or "feature", for 2D inputs) dimension.
207207
208- Provide a [`NNlib. RunningStats`](@ref) to fix a estimated mean and variance.
208+ Provide a [`RunningStats`](@ref) to fix a estimated mean and variance.
209209`batchnorm` will renormalize the input using these statistics during inference,
210210and update them using batch-level statistics when training.
211211To override this behaviour, manually set a value for `training`.
212212
213213If specified, `scale` and `bias` will be applied as an additional learned affine transform.
214214
215- See also [`NNlib. layernorm`](@ref), [`NNlib. instancenorm`](@ref), and [`NNlib. groupnorm`](@ref).
215+ See also [`layernorm`](@ref), [`instancenorm`](@ref), and [`groupnorm`](@ref).
216216"""
217217function batchnorm (x:: AbstractArray{<:Any, N} ,
218218 running_stats:: Union{RunningStats, Nothing} = nothing ,
@@ -247,7 +247,7 @@ To override this behaviour, manually set a value for `training`.
247247
248248If specified, `scale` and `bias` will be applied as an additional learned affine transform.
249249
250- See also [`NNlib. layernorm`](@ref), [`NNlib. batchnorm`](@ref), and [`NNlib. groupnorm`](@ref).
250+ See also [`layernorm`](@ref), [`batchnorm`](@ref), and [`groupnorm`](@ref).
251251"""
252252function instancenorm (x:: AbstractArray{<:Any, N} ,
253253 running_stats:: Union{RunningStats, Nothing} = nothing ,
@@ -281,7 +281,7 @@ The number of channels must be an integer multiple of the number of groups.
281281
282282If specified, `scale` and `bias` will be applied as an additional learned affine transform.
283283
284- See also [`NNlib. layernorm`](@ref), [`NNlib. batchnorm`](@ref), and [`NNlib. instancenorm`](@ref).
284+ See also [`layernorm`](@ref), [`batchnorm`](@ref), and [`instancenorm`](@ref).
285285
286286# Examples
287287
0 commit comments