How do neural networks work at a high level, and what are the risks of using them in investment decisions?
The CFA Level II curriculum now covers neural networks. I understand they're inspired by the brain, but the actual mechanics of layers, activation functions, and backpropagation are fuzzy. More importantly, what are the dangers of using them for financial forecasting?
Neural networks are powerful function approximators that can learn complex, non-linear relationships in data. For CFA Level II, you need to understand the architecture, training process, and key risks.
Architecture:
A neural network consists of layers of interconnected nodes (neurons):
- Input layer: Receives the features (e.g., P/E ratio, momentum score, volatility)
- Hidden layers: Transform inputs through weighted connections and activation functions. Each layer learns increasingly abstract representations.
- Output layer: Produces the prediction (e.g., expected return, buy/sell signal)
Each connection has a weight (parameter). A simple network might have hundreds of parameters; deep networks can have millions.
Activation Functions:
These introduce non-linearity — without them, a neural network would just be a linear regression regardless of depth:
- ReLU (Rectified Linear Unit): max(0, x) — most popular, computationally efficient
- Sigmoid: Maps to (0,1) — useful for probability outputs
- Tanh: Maps to (-1,1) — centers output around zero
Training (Backpropagation):
- Forward pass: input flows through the network to produce a prediction
- Calculate error (loss function): compare prediction to actual
- Backward pass: calculate how each weight contributed to the error
- Update weights to reduce error (gradient descent)
- Repeat thousands of times until convergence
Risks in Financial Applications:
| Risk | Description | Mitigation |
|---|---|---|
| Overfitting | Memorizes training data noise, fails on new data | Regularization, dropout, cross-validation |
| Black box | Cannot explain WHY a prediction was made | Use simpler models for regulatory/client transparency |
| Non-stationarity | Financial relationships change over time | Retrain frequently, use rolling windows |
| Data snooping | Testing many architectures until one 'works' | Strict out-of-sample testing protocol |
| Tail events | Training data may not include crashes/crises | Augment with stress scenarios |
Practical Example:
Atlas Capital builds a neural network to predict monthly stock returns using 50 fundamental and technical factors. The model achieves 65% accuracy on training data (2010-2022) but only 51% on test data (2023-2024). This gap signals severe overfitting — the model learned patterns specific to the training period that don't generalize.
Reducing the network from 3 hidden layers to 1, applying dropout (randomly disabling 20% of neurons during training), and adding L2 regularization brings test accuracy to 56% — modest but more reliable.
Exam Tip: The CFA exam focuses on conceptual understanding and risks, not mathematical details of backpropagation. Know the overfitting problem and mitigation techniques.
Explore ML in finance in our CFA Level II Quantitative Methods course.
Master Level II with our CFA Course
107 lessons · 200+ hours· Expert instruction
Related Questions
What exactly is the Capital Market Expectations (CME) framework and why does it matter for asset allocation?
How do business cycle phases affect asset class return expectations?
Can someone explain the Grinold–Kroner model step by step with numbers?
How do you forecast fixed-income returns using the building-blocks approach?
PPP vs Interest Rate Parity for forecasting exchange rates — when do I use which?
Join the Discussion
Ask questions and get expert answers.