Marginal Imbalance:
A dataset is marginally imbalanced if one class is rare compared to other class.
$Pr(Y=1)\eqsim0$
Conditional Imbalance:
A dataset is conditionally imbalanced when it is easy to predict the correct labels for most cases.
$Pr(Y=1|X=0)\eqsim0 \; and Pr(Y=1|X=1)\eqsim1$
Sub-Sampling (Down-sampling):
Subsampling is a method for reducing data size by selecting a subset of the original data. The imbalance is meant to kept constant for both training and test set.
Up-Sampling:
In this approach cases from the minority classes are sampled with replacement until each class has the approximately same size.
SMOTE (synthetic minority over-sampling technique):
$$ \tilde\pi \approx \frac{e^{\hat\beta_0}}{1+e^{\hat\beta_0}} \newline \hat\beta_0\approx log \big(\frac{\tilde\pi}{1-\tilde\pi} ) $$
$$ \hat\beta_0^*\approx log \big(\frac{\pi}{1-\pi} ) $$
$$ \hat\beta_0^* = \hat\beta_0 - \big(\frac{\tilde\pi}{1-\tilde\pi} ) + \big(\frac{\pi}{1-\pi} ) $$
Here $\tilde\pi$ is estimated probability and $\pi$ is Prevalence of MI
$$ \pi = \frac{n_{cases}}{n_{controls}+n_{cases}} $$