Best Practices for Lazy Loading Modules in Angular

Anton Ioffe - November 24th 2023 - 10 minutes read

As Angular applications grow in complexity and scale, performance optimization becomes crucial, with lazy loading serving as a key player in shaping a swift and efficient user experience. Navigating this landscape, however, poses its own set of challenges, often transforming lazy loading from a performance booster into a source of bottlenecks. In this deep dive, we're not just skimming the surface; we're dissecting performance pitfalls, architecting for scalable lazy loading, leveraging advanced routing, harnessing module federation, and syncing state management with RxJS to ensure that every bit of your application performs precisely as intended. Prepare to fortify your Angular applications, as we unravel the fabrics of optimally lazy-loaded modules, and perfect your mastery over Angular's dynamic capabilities.

Diagnosing and Resolving Performance Bottlenecks in Angular Lazy Loading

Lazy loading in Angular aims to enhance application performance by loading feature modules on demand. However, when not implemented properly, it can lead to performance bottlenecks that negate its benefits. One common pitfall is the size of the chunks being lazy-loaded. If a feature module is too large or includes unnecessary dependencies, it can significantly delay the loading time. To diagnose such issues, use the webpack Bundle Analyzer to visualize the size of your bundles. It helps identify oversized chunks and potential redundancies.

Profiling app load times is another essential step in diagnosing performance bottlenecks. Tools like the Chrome DevTools' Performance tab allow developers to measure metrics like Time to Interactive and First Contentful Paint, highlighting the impact lazy loading has on load times. Inspecting the timeline and flame charts can pinpoint exact moments when lazy-loaded modules affect performance, guiding optimization strategies.

Once you've identified sizeable chunks or slowdowns during load times, revisiting the organization of your code and dependencies is key. Ensure that your modules are not only logically separated but also trimmed of unnecessary imports. For instance, if multiple feature modules share a common set of utilities, consider abstracting these into a shared module that can be loaded once instead of being duplicated across lazy-loaded chunks.

Another way to resolve performance bottlenecks is to optimize the loading strategy of your feature modules. Using Angular’s route preloading strategies can strike a balance between immediate loading and lazy loading. By predicting user behavior, you can preload certain modules that the user is likely to access, thus smoothing out load times and improving user experience without compromising initial performance.

Lastly, when error tracing indicates a bottleneck, it’s often useful to reevaluate the placement and size of feature modules. Splitting a large module into smaller, more manageable chunks can improve load times and create a more responsive application. Fine-tuning the granularity of your modules not only aids in achieving optimal lazy loading but also increases the maintainability of the codebase. Regularly reviewing the structure and loading behavior as the application evolves is crucial to maintaining performance gains from lazy loading.

Architectural Patterns for Scalable Lazy Loading

Adopting feature-based module splitting is a fundamental pattern for scalable lazy loading in Angular applications. By dividing your code into feature modules, each encapsulating a coherent set of functionalities, you can load only the necessary code per user interaction. This practice aligns with domain-driven design, enforcing separation of concerns and enhancing readability. When designing your architecture, carefully consider which components, services, and pipes belong in the feature module. It's critical that each module is self-contained, with dependencies minimized to prevent tightly coupled code that could derail the lazy loading benefits.

For functionalities that transcend a single feature, shared modules play a vital role. A shared module typically contains UI components, pipes, and directives that are reused across multiple feature modules. However, it's imperative to avoid bloating the shared module with unnecessary code, as this could lead to increased bundle sizes. Strategically leveraging shared modules wherein only truly reusable elements are included promotes a lighter and more efficient application structure. When a feature-specific component starts being used across other modules, that's when it should be considered for migration to the shared module.

Conversely, the core module pattern is used for single-use components and application-wide services. These are generally loaded eagerly, as they form the foundational elements that your application relies on from the start. Defining a clear boundary between core and feature modules is crucial for preserving the initial load performance. Core modules should not be bloated with features but should contain only the necessary services that require a single instance throughout the app's lifecycle.

Navigating the decision of what to include in core, shared, and feature modules requires careful deliberation. Keep performance at the forefront of these decisions by constantly asking if the inclusion of specific code justifies the cost in terms of application loading time and runtime efficiency. While core functionalities are essential from the beginning, features should only be loaded when necessary, and shared utilities must be judiciously evaluated for their cross-cutting nature.

Lastly, fostering tight cohesion within feature modules while maintaining loose coupling between them enables high modularity and reusability. Structuring your Angular application in this manner can significantly ease maintainability and scale. It's not merely a matter of technical structuring; this pattern also supports team autonomy, allowing different groups to work on distinct features with minimal overlap or interference. This separation enables more straightforward parallel development and can help prevent merge conflicts in larger, more complex projects.

Advanced Routing Techniques for Optimized Lazy Loading

Angular's routing mechanism offers a sophisticated platform for implementing lazy loading, but its efficiency hinges on properly configured routes. When setting up lazy-loaded feature modules, it is vital to delineate routes using the loadChildren property. This directs Angular to only load the module when the user navigates to the path associated with that module. For large applications, breaking down the app into well-defined feature modules can significantly streamline the loading process and manage dependencies more effectively.

Preload strategies complement lazy loading by pre-fetching modules after the initial load, which can greatly enhance user experience during later navigation. Angular provides a PreloadAllModules strategy out of the box, which is a good starting point but can be suboptimal for large applications with numerous modules. To mitigate this, developers can construct custom preloading strategies. Custom strategies allow for greater control over which modules are preloaded and when, taking into account user behavior or application-specific conditions.

For example, a custom preloading strategy can be designed to only preload modules when the network is idle or when the user is engaging with content related to the feature module. This approach leverages Angular's Route configuration, where each route's data property can include custom metadata that a preloading strategy can use to make decisions. Developers can thus fine-tune module loading, prioritizing critical features while deferring others until they are likely to be needed.

Crafting custom preloading strategies requires an understanding of the Angular Router's PreloadingStrategy interface. Below is a real-world code example demonstrating how to implement a custom strategy:

import { PreloadingStrategy, Route } from '@angular/router';
import { Observable, of } from 'rxjs';

export class CustomPreloadingStrategy implements PreloadingStrategy {
    preload(route: Route, load: Function): Observable<any> {
        // Check if the route has data and a preload key set to true
        if ( &&['preload']) {
            return load();
        } else {
            // Return an Observable of null for routes that shouldn't be preloaded
            return of(null);

With this custom strategy in place, you can modify the routes configuration to specify which routes should be preloaded:

const routes: Routes = [
        path: 'feature-module',
        loadChildren: () => import('./feature-module/feature.module').then(m => m.FeatureModule),
        data: { preload: true } // This route will be preloaded based on the custom strategy
    // ...other routes

Route guards are an additional layer of routing control that can protect lazy-loaded routes. While not directly part of the preloading process, they ensure that modules are not loaded until certain criteria—such as user authentication—are met. This reinforces the security of the lazy loading approach, as it not only defers loading resources until necessary but also verifies that the user has permission to view the content provided by the module. Through the careful orchestration of route configurations, custom preloading strategies, and protective route guards, developers can effectively manage resource loading for an optimal user navigational flow.

Streamlining Code with Angular Module Federation

Module Federation in Angular leverages a microfrontend architecture to further modularize applications and enable dynamic loading of features. By using Module Federation, different teams can build and deploy features independently of one another, enhancing scalability and update flexibility. However, the approach raises considerations around increased complexity for configuration and management, as well as potential overhead for inter-module communication.

Implementing Module Federation necessitates a deeper understanding of Angular's build system, particularly with regards to the optimization of the federated modules. Care must be taken to ensure that shared dependencies are accessed in the most efficient manner, avoiding redundancy across different modules. Moreover, Module Federation can lead to increased initial load times if not managed properly, due to the overhead of loading additional code for setup.

A common pitfall in Module Federation is version mismatch between shared libraries. When federated modules rely on different versions of a shared dependency, this can lead to runtime errors or bloated bundle sizes as multiple versions are fetched. Solutions to this involve aligning versions across teams or implementing strategies to fallback to a single version at runtime. Ensuring that the Module Federation build process accounts for these discrepancies is critical for maintaining robustness.

A real-world example of Angular Module Federation integration can be seen in a dashboard application where each team is responsible for a particular dashboard widget. Each widget can be independently developed and deployed, with the main application shell dynamically loading these widgets as federated modules. This allows for updates to individual widgets that can be rolled out without redeploying the entire dashboard, resulting in a more agile development process.

Despite the potential pitfalls, Module Federation unlocks a new level of modularity that enables large-scale projects to be broken down into more manageable and independently deliverable pieces. It encourages a culture of autonomous development teams, while also posing thoughtful questions about the trade-off between autonomy and the overhead of coordination. In architecting applications with Module Federation, developers must balance these considerations to achieve both the scalable modularization of features and a cohesive, performant user experience.

Optimizing Lazy Loading with State Management and RxJS Patterns

Optimizing state management within lazy-loaded Angular modules requires a strategic approach that leverages RxJS and modular feature states. To foster a stable application state, it’s crucial to initialize feature states as early as possible. Utilize the ngrx StoreModule.forFeature() method, registering anticipated states in the application's root, even before the corresponding lazy-loaded feature is navigated to. This preemptively avoids undefined state errors when cross-feature communication occurs. In conjunction with RxJS, use BehaviorSubject or ReplaySubject to cache the last known value from a feature's state, ensuring that even if the feature module has not been loaded yet, a valid default state is always available for other features to consume.

Maintaining efficient data streams while employing lazy loading hinges on well-orchestrated RxJS patterns. Threading data across features can be achieved through a shared Angular service with an integrated RxJS Subject. When a feature module is finally loaded, its components subscribe to the shared Subject, which guarantees the reception of the latest emitted values representing the shared state. This service-based state sharing model is favored in scenarios where eager state initialization is impractical. It allows features to be truly siloed, only interacting via well-defined data streams.

Furthermore, leverage the selector functions provided by ngrx, which utilize memoization to optimize state querying. Carefully design selectors to extract only necessary slices of the state, diminishing the likelihood of performance hitches due to excessive data processing or redundant computations. Employ RxJS operators like distinctUntilChanged to prevent unnecessary re-rendering in components by filtering out unmodified state portions. In this way, selectors become a focal point for optimizing both data access patterns and application responsiveness.

For sustainable long-term state management, establish a feature state hygiene protocol. Ensure that when a lazy-loaded module is disposed of, its state does not persist redundantly in the main store, potentially leading to memory leaks. Utilize ngrx effects with RxJS teardown logic to clean up state slices when they are no longer necessary. This typically involves reacting to the Angular Router's navigation events to detect when a user exits a feature module, prompting a state clearance action to reset the relevant state slice.

To encapsulate state management concerns within feature modules, adopt a pattern of reflected state changes whereby actions dispatched from one module trigger effects in another. This decoupling enhances modularity, but it also necessitates stringent type safety and adherence to predefined interfaces to prevent run-time errors. A clear contract between modules about the nature and structure of the state to be shared or modified ensures smooth, predictable state transitions across lazy boundaries. Moreover, in an RxJS context, utilize higher-order mapping operators to manage complex data transformation workflows, allowing lazy modules to inject state changes into global streams without direct state access.

These techniques streamline the integration of lazy loading with Angular’s state management in a manner that safeguards application integrity and promotes systematic scalability. The combination of ngrx paradigms for feature state initialization, RxJS-powered services for state sharing, and strict state lifecycle management forms a robust foundation for efficiently constructed lazy-loaded modules within a complex Angular application.


In this article on "Best Practices for Lazy Loading Modules in Angular," the author explores the importance of performance optimization in Angular applications and how lazy loading can be a powerful tool. They provide insights on diagnosing and resolving performance bottlenecks, architectural patterns for scalable lazy loading, advanced routing techniques, and optimizing lazy loading with state management and RxJS patterns. The key takeaways from this article are the need to carefully manage chunk sizes and load times, adopt feature-based module splitting for scalability, utilize advanced routing strategies, and implement efficient state management. The challenging task for readers is to analyze their own Angular applications, identify potential performance bottlenecks, and implement lazy loading strategies to improve efficiency.

Don't Get Left Behind:
The Top 5 Career-Ending Mistakes Software Developers Make
FREE Cheat Sheet for Software Developers