|
7 | 7 | * The idea of high-dimensional space: the code is cut into high-dimensional space, and then a very detailed high-dimensional classification is done to separate it. Then the search is also high-dimensional, just like the code, it is entered into the treesitter to do training to obtain logical learning relationships. Most of NLP is a multi-classification problem in high-dimensional space.
|
8 | 8 | * Collect the input x and output y around you as training data, and mine their mapping relationship f(x) at any time. You can use GPT to generate certain data for your model training needs or write crawler to get you need data.
|
9 | 9 |
|
10 |
| -## Deep Learning Fundamentals |
11 |
| - |
12 |
| -Deep learning is a subset of machine learning that uses artificial neural networks with multiple layers (deep neural networks) to progressively extract higher-level features from raw input. Here are the key components and concepts: |
13 |
| - |
14 |
| -### Neural Network Architecture |
15 |
| -* Input Layer: Receives raw data and normalizes it for processing |
16 |
| -* Hidden Layers: Multiple layers that transform data through weighted connections |
17 |
| -* Output Layer: Produces the final prediction or output |
18 |
| -* Activation Functions: Non-linear functions (ReLU, sigmoid, tanh) that help networks learn complex patterns |
19 |
| - |
20 |
| -### Key Deep Learning Concepts |
21 |
| -1. Backpropagation |
22 |
| - * Algorithm for calculating gradients in neural networks |
23 |
| - * Efficiently updates weights by propagating error backwards through the network |
24 |
| - * Uses chain rule to compute partial derivatives |
25 |
| - |
26 |
| -2. Gradient Descent Optimization |
27 |
| - * Stochastic Gradient Descent (SGD) |
28 |
| - * Mini-batch Gradient Descent |
29 |
| - * Adaptive optimizers (Adam, RMSprop) |
30 |
| - |
31 |
| -3. Loss Functions |
32 |
| - * Mean Squared Error (MSE) for regression |
33 |
| - * Cross-Entropy Loss for classification |
34 |
| - * Custom loss functions for specific tasks |
35 |
| - |
36 |
| -4. Regularization Techniques |
37 |
| - * Dropout: Randomly deactivates neurons during training |
38 |
| - * L1/L2 Regularization: Adds penalty terms to prevent overfitting |
39 |
| - * Batch Normalization: Normalizes layer inputs for stable training |
40 |
| - |
41 |
| -### Deep Learning Architectures |
42 |
| - |
43 |
| -1. Convolutional Neural Networks (CNNs) |
44 |
| - * Specialized for processing grid-like data (images) |
45 |
| - * Key components: Convolutional layers, pooling layers, fully connected layers |
46 |
| - * Applications: Image classification, object detection, segmentation |
47 |
| - |
48 |
| -2. Recurrent Neural Networks (RNNs) |
49 |
| - * Process sequential data with memory of previous inputs |
50 |
| - * Variants: LSTM, GRU for handling long-term dependencies |
51 |
| - * Applications: Time series prediction, natural language processing |
52 |
| - |
53 |
| -3. Transformers |
54 |
| - * State-of-the-art architecture for sequence processing |
55 |
| - * Self-attention mechanism for capturing relationships |
56 |
| - * Applications: Language models, machine translation, text generation |
57 |
| - |
58 |
| -4. Autoencoders |
59 |
| - * Unsupervised learning for dimensionality reduction |
60 |
| - * Encoder-decoder architecture |
61 |
| - * Applications: Feature learning, denoising, anomaly detection |
62 |
| - |
63 | 10 | - [Python & R Machine Learning](#python--r-machine-learning)
|
64 | 11 | - [R Machine Learning](https://github.com/chanshunli/jim-emacs-machine-learning/tree/master/R-Lang-machine-learning)
|
65 | 12 | - [least squares method](#least-squares-method)
|
|
0 commit comments