Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix typos #53

Merged
merged 1 commit into from
Apr 15, 2022
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions docs/src/examples/mnist.md
Original file line number Diff line number Diff line change
Expand Up @@ -56,7 +56,7 @@ corresponding to the model:
```julia
@time p = SimpleChains.init_params(lenet);
```
The convolutional layers are initialized with a Glorot (Xavier) unifirom distribution,
The convolutional layers are initialized with a Glorot (Xavier) uniform distribution,
while the dense layers are initialized with a Glorot (Xaviar) normal distribution.
Biases are initialized to zero.
Because the number of parameters can be a function of the input size, these must
Expand All @@ -75,7 +75,7 @@ G = SimpleChains.alloc_threaded_grad(lenetloss);
Here, we're estimating that the number of physical cores is half the number of threads
on an `x86_64` system, which is true for most -- but not all!!! -- of them.
Otherwise, we're assuming it is equal to the number of threads. This is of course also
likely to be wrong, e.g. recent Power CPUs may habe 4 or even 8 threads per core.
likely to be wrong, e.g. recent Power CPUs may have 4 or even 8 threads per core.
You may wish to change this, or use [Hwloc.jl](https://github.com/JuliaParallel/Hwloc.jl) for an accurate number.

Now that this is all said and done, we can train for `10` epochs using the `ADAM` optimizer
Expand Down