Introduction

Welcome back to the third lesson of "Building and Applying Your Neural Network Library"! You've made tremendous progress in this course. In our first lesson, we successfully modularized our core neural network components — dense layers and activation functions. Then, in our previous lesson, we organized our training components by creating dedicated modules for loss functions and optimizers. We now have a well-structured foundation with a clean separation of concerns.

However, as you may have noticed from our previous training examples, we're still writing quite a bit of boilerplate code for each training session. We manually create layers, set up optimizers, write training loops, handle forward and backward passes, and coordinate all these components ourselves. While this gives us complete control, it also means we're repeating the same orchestration logic every time we want to train a model.

In this lesson, we're going to orchestrate all these components into a unified, high-level interface. We'll build a powerful Model abstract class that acts as the conductor of our neural network orchestra, coordinating layers, optimizers, and loss functions through clean, intuitive methodslike compile(), fit(), and predict(), as well as a SequentialModel concrete subclass that can be seen as a better and improved replacement for the MLP class we developed previously, providing a much more elegant and maintainable API for building and training neural networks. Let's get started!

The Need for Orchestration

Think of a symphony orchestra — while each musician is skilled at playing their individual instrument, the magic happens when a conductor coordinates all these talents toward a unified performance. Similarly, we've built excellent individual components (layers, optimizers, losses), but we need a conductor to orchestrate them into a seamless training experience.

Currently, our training process requires us to manually coordinate several moving parts:

  • Instantiate layers and build our network architecture
  • Create an optimizer with specific parameters
  • Define our loss function
  • Implement the training loop with forward passes, loss calculations, backward passes, and weight updates

This manual orchestration is error-prone and repetitive — exactly the kind of work that should be automated. What we need is a Model class that serves as this conductor, providing a high-level API that handles the complexities of training coordination while still giving us the flexibility to customize our network architecture, choose different optimizers and loss functions, and control training parameters.

Understanding Abstract Classes

Let's start by designing our orchestrator, which is a base Model class that defines the common interface and shared functionality for all types of neural networks. We'll use an abstract base class approach, which allows us to define the contract that all model types must follow while leaving room for specific implementations.

An abstract class is a special type of class that cannot be instantiated directly but serves as a blueprint for other classes. Think of it like an architectural blueprint for houses — you can't live in the blueprint itself, but it defines the essential structure that all houses built from it must have (foundation, walls, roof, etc.). In Python, we use the ABC (Abstract Base Class) and @abstractmethod decorator to create these blueprints. When we mark methods like _forward() and _backward() with @abstractmethod, we're saying "every model type must have these methods, but each one will implement them differently."

This approach is perfect for our neural network library because different model architectures (sequential, convolutional, recurrent) all need the same core functionality — they all need to perform forward passes, backward passes, and training — but each implements these operations differently. By using an abstract base class, we enforce a contract that guarantees every model type will have the methods our training system expects, while allowing each model to implement them in the way that makes sense for its architecture. This prevents bugs (you can't forget to implement a required method) and ensures our fit() method will work with any model type we create in the future.

Designing the Base Model Class

Now that we understand the role of abstract classes, let's implement our Model orchestrator:

This foundation establishes all the essential components our model will need to coordinate:

  • Layer stack management: The layers list holds our network architecture
  • Optimizer coordination: Centralized optimizer configuration and management
  • Loss function setup: Unified loss function and derivative handling
  • Compilation validation: The is_compiled flag prevents training misconfiguration
  • Future extensibility: The string-based configuration system makes adding new optimizers and loss functions straightforward
Implementing Training and Prediction Logic

Now let's add the core training and prediction functionality. The abstract method pattern we're using ensures scalability — any new model architecture can plug into our training orchestration system by simply implementing the required methods.

The fit() method orchestrates the entire training process with several key responsibilities:

  • Validation: Ensures the model is properly compiled before training
  • Batch processing: Handles flexible batch sizes for memory efficiency and training stability
  • Data management: Shuffles data each epoch to prevent overfitting to data order
  • Training coordination: Manages the forward-backward-update cycle automatically
  • Progress tracking: Provides informative training progress feedback
  • Architecture independence: Works with any model architecture through the abstract method pattern, with _forward and _backward methods

This design ensures our training system can scale to handle different model types, training strategies, and dataset sizes without requiring code changes to the core training logic.

Building the Sequential Model

Now we can create our concrete SequentialModel class that implements the abstract methods from our base Model. This class represents the familiar stack-of-layers architecture we've been working with, but now with a much cleaner interface and full integration into our extensible framework.

The SequentialModel demonstrates excellent separation of concerns:

  • Initialization flexibility: Supports both empty initialization and pre-built layer lists
  • Architecture building: The add() method provides the familiar interface for building networks layer by layer
  • Sequential processing: Forward pass flows data through layers in order, backward pass propagates gradients in reverse
  • Clean responsibility: Focuses solely on managing a linear sequence of layers while inheriting all training orchestration from the parent class

This design makes our library highly extensible — we can easily add other model types like convolutional networks or attention mechanisms by implementing the same abstract interface, while all the training logic remains reusable.

Using Our Orchestrated Model

Now let's see our orchestration in action! Here's how we can solve the XOR problem using our new high-level API, which is much cleaner and more maintainable than our previous manual approach. Notice how this same interface will work seamlessly when we extend our library with new layers, optimizers, or loss functions.

Notice how much cleaner this is compared to our previous training scripts! The beauty of this approach is that it maintains the flexibility to customize every aspect of our network while eliminating the repetitive boilerplate code. The extensible design means we can easily experiment with different architectures, optimizers, or hyperparameters without rewriting the training logic each time.

Discussing the Output

When we run our orchestrated model, we can see how it successfully learns the XOR problem while providing clear feedback about the training process. The consistent interface and automated orchestration ensure reliable, reproducible results across different model configurations.

The results demonstrate excellent performance — our orchestrated model achieves the same learning quality as our previous manual implementations, but with much cleaner, more maintainable code. The loss decreases smoothly from 0.25 to 0.0014, and the final predictions correctly solve the XOR problem with high confidence. More importantly, the training process is now fully automated, consistently implemented, and ready to scale to more complex problems and architectures.

Conclusion

Outstanding work! We've successfully built the orchestration layer for our neural network library, creating a powerful and elegant high-level API that coordinates all our modular components. Our Model and SequentialModel classes demonstrate how good software architecture can transform complex, error-prone manual processes into clean, automated workflows.

The orchestration pattern we've implemented provides simplified user interfaces, reduced code duplication, improved maintainability, enhanced extensibility for future development, and a solid foundation for scaling to production-level neural network applications.

We're now ready for the final step in our journey: putting our complete neural network library to work on a real-world dataset. In our next lesson, we'll apply everything we've built to the California Housing dataset, demonstrating how our library handles practical machine learning problems with multiple features and realistic data challenges.

Sign up
Join the 1M+ learners on CodeSignal
Be a part of our community of 1M+ users who develop and demonstrate their skills on CodeSignal