Skip to content
This repository has been archived by the owner on Nov 8, 2024. It is now read-only.

Fix a few typos #66

Merged
merged 1 commit into from
Oct 16, 2022
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion _literate/02_bayes_stats.jl
Original file line number Diff line number Diff line change
Expand Up @@ -537,7 +537,7 @@ savefig(joinpath(@OUTPUT, "mixture.svg")); # hide
# I believe Turing is the most **important and popular probabilistic language framework in Julia**. It is what PyMC3 and Stan
# are for Python and R, but for Julia. Furthermore, you don't have to do "cartwheels" with Theano backends and tensors like
# in PyMC3 or learn a new language to declare your models like in Stan (or even have to debug C++ stuff).
# Turing is **all** Julia. It uses Julia arrays, Julia distributions, Julia autodiff, Julia plots, Julia randon number generator,
# Turing is **all** Julia. It uses Julia arrays, Julia distributions, Julia autodiff, Julia plots, Julia random number generator,
# Julia MCMC algorithms etc. I think that developing and estimating Bayesian probabilistic models using Julia and Turing is
# **powerful**, **intuitive**, **fun**, **expressive** and allows **easily new breakthroughs** simply by being 100% Julia and
# embedded in Julia ecosystem. As discussed in [1. **Why Julia?**](/pages/1_why_Julia/), having multiple dispatch with
Expand Down
4 changes: 2 additions & 2 deletions _literate/05_MCMC.jl
Original file line number Diff line number Diff line change
Expand Up @@ -116,7 +116,7 @@

# ### Simulations

# I will do some simulations to ilustrate MCMC algorithms and techniques. So, here's the initial setup:
# I will do some simulations to illustrate MCMC algorithms and techniques. So, here's the initial setup:

using Plots, StatsPlots, Distributions, LaTeXStrings, Random

Expand Down Expand Up @@ -183,7 +183,7 @@ const mvnormal = MvNormal(μ, Σ)

data = rand(mvnormal, N)';

# In the figure below it is possible to see a countour plot of the PDF of a multivariate normal distribution composed of two normal
# In the figure below it is possible to see a contour plot of the PDF of a multivariate normal distribution composed of two normal
# variables $X$ and $Y$, both with mean 0 and standard deviation 1.
# The correlation between $X$ and $Y$ is $\rho = 0.8$:

Expand Down
16 changes: 8 additions & 8 deletions _literate/08_ordinal_reg.jl
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,7 @@
# Most important, the distance between values is not the same.
# For example, imagine a pain score scale that goes from 1 to 10.
# The distance between 1 and 2 is different from the distance 9 to 10.
# Another example is opinion pools with their ubiquously disagree-agree range
# Another example is opinion pools with their ubiquitous disagree-agree range
# of plausible values.
# These are also known as Likert scale variables.
# The distance between "disagree" to "not agree or disagree" is different
Expand Down Expand Up @@ -68,7 +68,7 @@

# $$\ln \frac{p}{1-p} = \ln \frac{1}{1-1} = \ln 0 = \infty$$

# Thus, we only need $K-1$ intercepts for a $K$ possible depedent variables' response values.
# Thus, we only need $K-1$ intercepts for a $K$ possible dependent variables' response values.
# These are known as **cut points**.

# Each intercept implies a CDF for each value $K$.
Expand All @@ -86,9 +86,9 @@

# $$P(Y=k) = P(Y \leq k) - P(Y \leq k-1)$$

# where $Y$ is the depedent variable and $k \in K$ are the cut points for each intercept.
# where $Y$ is the dependent variable and $k \in K$ are the cut points for each intercept.

# Let me show you an example with some syntethic data.
# Let me show you an example with some synthetic data.

using DataFrames
using CairoMakie
Expand Down Expand Up @@ -292,13 +292,13 @@ end;

# First, let's deal with the new stuff in our model: the **`Bijectors.ordered`**.
# As I've said in the [4. **How to use Turing**](/pages/04_Turing/),
# Turing has a rich ecossystem of packages.
# Turing has a rich ecosystem of packages.
# Bijectors implements a set of functions for transforming constrained random variables
# (e.g. simplexes, intervals) to Euclidean space.
# Here we are defining `cutpoints` as a `ncateg - 1` vector of Student-$t$ distributions
# with mean 0, standard deviation 5 and degrees of freedom $\nu = 3$.
# Remember that we only need $K-1$ cutpoints for all of our $K$ intercepts.
# And we are also contraining it to be an ordered vector with `Bijectors.ordered`,
# And we are also constraining it to be an ordered vector with `Bijectors.ordered`,
# such that for all cutpoints $c_i$ we have $c_1 < c_2 < ... c_{k-1}$.

# As before, we are giving $\boldsymbol{\beta}$ a very weakly informative priors of a
Expand Down Expand Up @@ -422,11 +422,11 @@ end
# The `cutpoints` is the basal rate of the probability of our dependent variable
# having values less than a certain value.
# For example the cutpoint for having values less than `2` which its code represents
# the tobacco comsumption of 10-19 g/day has a median value of 20%.
# the tobacco consumption of 10-19 g/day has a median value of 20%.

# Now let's take a look at our coefficients
# All coefficients whose 95% credible intervals captures the value $\frac{1}{2} = 0.5$ tells
# that the effect on the propensity of tobacco comsumption is inconclusive.
# that the effect on the propensity of tobacco consumption is inconclusive.
# It is pretty much similar to a 95% credible interval that captures the 0 in
# the linear regression coefficients.

Expand Down