Skip to content

Latest commit

 

History

History

Deep Learning Crash Course

by Giovanni Volpe, Benjamin Midtvedt, Jesús Pineda, Henrik Klein Moberg, Harshith Bachimanchi, Joana B. Pereira, Carlo Manzo
No Starch Press, San Francisco (CA), 2026
ISBN-13: 9781718503922
https://nostarch.com/deep-learning-crash-course


  1. Dense Neural Networks for Classification

  2. Dense Neural Networks for Regression

  3. Convolutional Neural Networks for Image Analysis

  4. Encoders–Decoders for Latent Space Manipulation

  5. U-Nets for Image Transformation

  6. Self-Supervised Learning to Exploit Symmetries

  7. Recurrent Neural Networks for Timeseries Analysis

  8. Attention and Transformers for Sequence Processing

  9. Generative Adversarial Networks for Image Synthesis

  10. Diffusion Models for Data Representation and Exploration
    Presents denoising diffusion models for generating and enhancing images, including text-to-image synthesis and image super-resolution.

  • Code 10-1: Generating Digits with a Diffusion Model
    Implements a Denoising Diffusion Probabilistic Model (DDPM) on MNIST digits. It explains the forward process (adding Gaussian noise at each time step) and the reverse process (a trained denoising U-Net), culminating in random but plausible digit images. It also demonstrates how forward and reverse diffusion steps can be visualized, as well as how different runs from the same noise yield different samples.

  • Code 10-A: Generating Bespoke Digits with a Conditional Diffusion Model
    Extends the DDPM to condition on class labels using classifier-free guidance. Allows specifying which MNIST digit to generate. After training, the network can produce custom digits on demand by blending conditional and unconditional outputs.

  • Code 10-B: Generating Images of Digits from Text Prompts
    Demonstrates a mini text-to-image pipeline by pairing a custom transformer encoder (or pretrained CLIP) with a diffusion model. It converts sentences like "There are three horses and two lions. How many lions?" into correct digits. It also showcases classifier-free guidance, adding textual context into an attention U-Net.

  • Code 10-C: Generating Super-Resolution Images
    Uses a conditional diffusion model to transform low-resolution microscopy images into detailed high-resolution counterparts, showcasing the power of diffusion-based upsampling. It adapts the forward and reverse diffusion to combine the noisy target image with the low-resolution input, effectively learning a mapping to super-resolve biological data.

  1. Graph Neural Networks for Relational Data Analysis

  2. Active Learning for Continuous Learning

  3. Reinforcement Learning for Strategy Optimization

  4. Reservoir Computing for Predicting Chaos