Lecture 1: OverviewSeptember 6, 2023 50 / 57 The Curse of Dimensionality Handling complexity Involve many variables, how can we handle this complexity without get- ting into trouble. Optimization and Integration Usually Overview September 6, 2023 51 / 57 How to Handle Complexity Properly dealing with complexity is a crucial issue for machine learning. Limiting complexity is one approach Use a model that is complex enough if can find out how to reduce the large number of variables to a small number. Averaging over complexity is the Bayesian approach. Use as complex a model might be needed, but don’t choose a single parameter0 码力 | 57 页 | 2.41 MB | 1 年前3
《Efficient Deep Learning Book》[EDL] Chapter 4 - Efficient Architecturesto the linear computation complexity of RNNs. However, attention is still faster in wall clock time because it processes entire sequences together. The quadratic complexity of attention is addressed through transformer variants with efficient self-attention mechanisms. These ideas tackle the quadratic complexity at various levels. The simplest idea is to chunk the input sequence of length n into blocks of the input sequence, thus compressing it in the process. Sparse attention attempts to reduce the complexity by picking a subset of parameters for attention computation. This subset can also be learned during0 码力 | 53 页 | 3.92 MB | 1 年前3
《Efficient Deep Learning Book》[EDL] Chapter 1 - Introductionhas been a race to create deeper networks with an ever larger number of parameters and increased complexity. In Computer Vision, several model architectures such as VGGNet, Inception, ResNet etc. (refer classification. Each new breakthrough in neural networks has led to an increase in the network complexity, number of parameters, the amount of training resources required to train the network, prediction0 码力 | 21 页 | 3.17 MB | 1 年前3
《Efficient Deep Learning Book》[EDL] Chapter 2 - Compression Techniquestechniques using the metrics relevant to our use case. In some cases, these techniques can help reduce complexity and improve generalization. Let us consider an arbitrary neural network layer. We can abstract B, C, and D are all matrices. The number of MACs in a model is another metric to measure its complexity. And since they are such a fundamental block of models, they have been optimized both in hardware0 码力 | 33 页 | 1.96 MB | 1 年前3
深度学习与PyTorch入门实战 - 33. regularizationthings should not be used than are necessary. Reduce Overfitting ▪ More data ▪ Constraint model complexity ▪ shallow ▪ regularization ▪ Dropout ▪ Data argumentation ▪ Early Stopping Regularization0 码力 | 10 页 | 952.77 KB | 1 年前3
《Efficient Deep Learning Book》[EDL] Chapter 6 - Advanced Learning Techniques - Technical Reviewlinearly or exponentially to reach a value of 1.0 in the final epoch. It controls when and how much complexity you want to introduce in the training. Figure 6-12 shows multiple examples of pacing functions0 码力 | 31 页 | 4.03 MB | 1 年前3
共 6 条
- 1













