Skip to content
Merged
Show file tree
Hide file tree
Changes from 4 commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
5 changes: 5 additions & 0 deletions apps/site/authors.json
Original file line number Diff line number Diff line change
Expand Up @@ -14,6 +14,11 @@
"name": "Augustin Mauroy",
"website": "https://github.com/AugustinMauroy"
},
"Aviv Keller": {
Comment thread
avivkeller marked this conversation as resolved.
"id": "avivkeller",
"name": "Aviv Keller",
"website": "https://github.com/avivkeller"
},
"Ben Noordhuis": {
"id": "bnoordhuis",
"name": "Ben Noordhuis",
Expand Down
3 changes: 3 additions & 0 deletions apps/site/pages/en/learn/diagnostics/memory/index.md
Original file line number Diff line number Diff line change
Expand Up @@ -14,6 +14,9 @@ leaks is possible through retainers. As Node.js applications are usually
multi-tenant, business critical, and long-running, providing an accessible and
efficient way of finding a memory leak is essential.

You can also fine-tune memory to get specific results. Check out
[Understanding and Tuning Memory](/learn/diagnostics/memory/understanding-and-tuning-memory) for more details.

### Symptoms

The user observes continuously increasing memory usage _(can be fast or slow,
Expand Down
Original file line number Diff line number Diff line change
@@ -0,0 +1,137 @@
---
title: Understanding and Tuning Memory
layout: learn
authors: avivkeller
---

# Understanding and Tuning Memory

Node.js, built on Google’s V8 JavaScript engine, offers a powerful runtime for running JavaScript on the server side. However, as your applications grow, managing memory becomes a critical task for maintaining optimal performance and managing problems like memory leaks or crashes. In this article, we’ll explore how to monitor, manage, and optimize memory usage within Node.js. We’ll also cover important V8 concepts like the heap and garbage collection and discuss how to use command-line flags to fine-tune memory behavior.

## How V8 Manages Memory

Comment thread
avivkeller marked this conversation as resolved.
At its core, V8 divides memory into several parts, with two primary areas being the **heap** and the **stack**. Understanding these spaces, especially how the heap is managed, is key to improving memory usage in your app.

### The Heap

V8 manages memory using two main areas in the heap:

1. **New Space**: This is where new, short-lived objects are allocated. Because objects here are expected to be temporary, garbage collection occurs frequently, allowing memory to be reclaimed quickly.

2. **Old Space**: Objects that survive several garbage collection cycles in the New Space are moved to the Old Space. Since these objects are more persistent, garbage collection in this space occurs less often but is more resource-intensive when it happens.

In V8, memory for JavaScript objects, arrays, and functions is allocated in the **heap**. The size of the heap is not fixed, and exceeding the available memory can result in an "out-of-memory" error, causing your application to crash.

To check the current heap size limit, you can use the `v8` module.

```cjs
const v8 = require('node:v8');
const heapSizeLimit = v8.getHeapStatistics().heap_size_limit;
const heapSizeInGB = heapSizeLimit / (1024 * 1024 * 1024);

console.log(`${heapSizeInGB} GB`);
```

This will output the maximum heap size in gigabytes, which is based on your system's available memory.

### The Stack

In addition to the heap, V8 also uses the **stack** for memory management. The stack is a region of memory used to store local variables and function call information. Unlike the heap, which is managed by V8's garbage collector, the stack operates on a Last In, First Out (LIFO) principle.

Whenever a function is called, a new frame is pushed onto the stack. When the function returns, its frame is popped off. The stack is much smaller in size compared to the heap, but it is faster for memory allocation and deallocation. However, the stack has a limited size, and excessive use of memory (such as with deep recursion) can result in a **stack overflow**.

## Monitoring Memory Usage

Before tuning memory usage, it’s important to understand how much memory your application is consuming. Node.js and V8 provide several tools for monitoring memory usage.

### Using `process.memoryUsage()`

The `process.memoryUsage()` method provides insights into how much memory your Node.js process is using. It returns an object with details like:

- **`rss`** (Resident Set Size): The total memory allocated to your process, including heap and other areas.
- **`heapTotal`**: The total memory allocated for the heap.
- **`heapUsed`**: The memory currently in use within the heap.
- **`external`**: Memory used by external resources like bindings to C++ libraries.
- **`arrayBuffers`**: Memory allocated to various Buffer-like objects.

Here’s an example:

```javascript
console.log(process.memoryUsage());
```

The output will look like:

```json
{
"rss": 25837568,
"heapTotal": 5238784,
"heapUsed": 3666120,
"external": 1274076,
"arrayBuffers": 10515
}
```

By monitoring these values over time, you can identify if memory usage continually increases over time, a common sign of memory leaks.

## Command-Line Flags for Memory Tuning

Node.js offers several command-line flags that help you fine-tune memory-related settings. Below are some key options you can use to optimize memory usage.

### `--max-old-space-size`

This flag sets a limit on the size of the **Old Space** in the V8 heap, where long-lived objects are stored. If your application uses a significant amount of memory, you might need to adjust this limit.

To increase the Old Space limit to 4 GB, for example, use:

```bash
node --max-old-space-size=4096 app.js
```

This sets the Old Space size to 4096 MB (4 GB), but you should adjust this based on your system's available memory.

### `--max-semi-space-size`

The `--max-semi-space-size` flag controls the size of the **New Space** in the V8 heap. New Space stores newly created objects, which are garbage collected frequently. Increasing this size can help reduce the frequency of minor garbage collections, which can in turn improve performance if your application allocates a lot of short-lived objects.

For example:

```bash
node --max-semi-space-size=64 app.js
```

This increases the New Space to 64 MB (compared to the default, which is typically much smaller). This adjustment can help in high-throughput applications where frequent garbage collection of short-lived objects is adding noticeable overhead.

### `--gc-interval`

This flag adjusts how frequently garbage collection cycles occur. V8 normally determines this interval automatically based on heuristics, but in special cases, you might want to override it.

For example:

```bash
node --gc-interval=100 app.js
```

Setting this to 100 forces V8 to attempt garbage collection every 100 ms. This might be useful in low-latency systems where predictable memory cleanup is needed, such as a long-running service that must consistently stay within a narrow memory footprint. However, changing this without careful testing can degrade performance due to excessive garbage collection.

### `--expose-gc`

With the `--expose-gc` flag, you can manually trigger garbage collection from your application code. This can be helpful in specific cases where you know a large chunk of memory is no longer needed and want to release it immediately—for example, after processing a large data set or completing a batch job.

To enable it, start your app with:

```bash
node --expose-gc app.js
```

Then, within your code, you can call:

```javascript
global.gc();
```

Keep in mind that manually triggering garbage collection **does not disable** the normal GC algorithm. V8 will still perform automatic garbage collection as needed. Manual calls are supplemental and should be used with caution, as overuse can negatively impact performance.

## Putting It All Together

By adjusting settings for the Old Space and New Space sizes, selectively triggering garbage collection, and configuring heap limits, you can optimize your application’s memory usage and improve its overall performance. These tools give you the power to better manage memory in high-demand scenarios and maintain stability as your applications scale.
Comment thread
avivkeller marked this conversation as resolved.
7 changes: 4 additions & 3 deletions apps/site/pages/en/learn/diagnostics/user-journey.md
Original file line number Diff line number Diff line change
Expand Up @@ -5,12 +5,13 @@ layout: learn

# User Journey

These diagnostics guides were created by the [Diagnostics Working Group][] with the
objective of providing guidance when diagnosing an issue in a user's
application.
These diagnostics guides were created by the [Diagnostics Working Group][]
and the [Node.js Website Team][] with the objective of providing guidance
when diagnosing an issue in a user's application.

The documentation project is organized based on user journey. Those journeys
are a coherent set of step-by-step procedures that a user can follow to
root-cause their issues.

[Diagnostics Working Group]: https://github.com/nodejs/diagnostics
[Node.js Website Team]: https://github.com/nodejs/nodejs.org
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
---
title: Collecting code coverage in Node.js
layout: learn
authors: RedYetiDev
authors: avivkeller
---

# Collecting code coverage in Node.js
Expand Down
1 change: 1 addition & 0 deletions packages/i18n/locales/en.json
Original file line number Diff line number Diff line change
Expand Up @@ -108,6 +108,7 @@
"links": {
"diagnostics": "Diagnostics",
"userJourney": "User Journey",
"understandingAndFineTuningMemory": "Understanding and Fine Tuning Memory",
"memory": "Memory",
"liveDebugging": "Live Debugging",
"poorPerformance": "Poor Performance",
Expand Down
Loading