1. What is the difference between ImageNet and Imagenette? When is it better to experiment on one versus the other?
    2. What is normalization?
    3. Why didn’t we have to care about normalization when using a pretrained model?
    4. What is progressive resizing?
    5. What is test time augmentation? How do you use it in fastai?
    6. Is using TTA at inference slower or faster than regular inference? Why?
    7. What is Mixup? How do you use it in fastai?
    8. Why does Mixup prevent the model from being too confident?
    9. Why does training with Mixup for five epochs end up worse than training without Mixup?
    10. What problems in your data can label smoothing help with?
    11. When using label smoothing with five categories, what is the target associated with the index 1?
    12. What is the first step to take when you want to prototype quick experiments on a new dataset?
    1. Use the fastai documentation to build a function that crops an image to a square in each of the four corners, then implement a TTA method that averages the predictions on a center crop and those four crops. Did it help? Is it better than the TTA method of fastai?
    2. Find the Mixup paper on arXiv and read it. Pick one or two more recent articles introducing variants of Mixup and read them, then try to implement them on your problem.
    3. Read the sidebar “Label Smoothing, the Paper”, look at the relevant section of the original paper and see if you can follow it. Don’t be afraid to ask for help!