By combining multiple technologies such as network compression, multi-threading, dynamic TCP window size, advanced deduplication and global caching, WAN acceleration provides increased transport speed for Backup Copy and Replication jobs on low bandwidth or limited network links; or in any situation where we want to optimize the network usage, at the expense of other components (cpu/memory/disk of the accelerators). This technology is specifically designed to accelerate Veeam jobs. Any other WAN acceleration technology should be disabled for Veeam traffic.
Starting with v10 Veeam Backup and Replication supports two operational modes of WAN accelerators: Low bandwidth mode for the links with less than 100 Mbps throughput and High bandwidth mode, supporting links with up to 1 Gbps throughput. The operational mode can be controlled in WAN accelerator properties.
|1-100 Mb/s||Low Bandwidth|
|100 Mb/s -1Gb/s||High Bandwidth|
Note that these values are the suggested operational limits of the WAN accelerator, or said differently, where the selected option performs at its best. Customers with no specific needs for bandwidth usage, or backup execution time, can also use direct connection with low bandwidth.
The main difference between the two modes is the use of the “Global cache”. High bandwidth mode does not use it, instead it uses advanced compression and deduplication mechanisms, including digest files, used also by low bandwidth mode, thus the disk space and compute requirements are lower for High-bandwidth mode in many cases. High-bandwidth mode puts more load on source than target.
WAN accelerators work in pairs: one WAN accelerator is deployed at the source site and another at the target site. Many to one scenario (e.g. ROBO-sites to a central site) is also possible but needs accurate sizing and has higher system requirements.
How many source WAN accelerators do I need?
Since WAN accelerator processes VM disks sequentially, you may consider using more than one source WAN accelerator on a single site to distribute tasks between them. Many-to-one scenario is supported; in this case the target WAN accelerator should be sized carefully to be able to accept multiple incoming connection, and in general we recommend a maximum ratio of 4:1. Higher ratios are allowed in some specific situations, like many small links connecting to one large target WAN Accelerator; in this case we have observed ratio up to 20:1, but we suggest in this case to evaluate the design with a Veeam expert.
If you are not sure about the mode you are going to use, or you plan for many-to-one scenario with various source modes, make sure you choose High-bandwidth mode during initial configuration of the target WAN accelerator settings. In such a case you will be able to work in both low and high bandwidth modes.
If you choose low bandwidth mode during the target configuration, WAN acceleration will process everything in low bandwidth mode only, despite of the source WAN accelerator settings.
The source WAN accelerator consumes a high amount of CPU and memory whilst re-applying the WAN optimized compression algorithm. Recommended system configuration is 4 CPU cores and 8 GB RAM.
When you are using the WAN accelerator server also for other Veeam roles like Proxy or Repository make sure not to overcommit CPU and memory resources and there are enough resources for each role. In general, we recommend to build machines where the WAN Accelerator is the only Veeam role.
The I/O requirements for the source WAN accelerator spikes every time a new VM disk starts processing, so the typical I/O pattern is made of many small blocks. Thus, it is recommended to deploy WAN accelerators on disk with decent I/O performance, avoiding high latency spinning disks.
Source WAN accelerator IOPS
Source WAN accelerator stores only the digests of the processed VM OS data blocks in the VeeamWAN folder on the disk that you select when you configure the WAN accelerator. Disk digest file is created and placed in
In low bandwidth (LB) mode each digest file consumes up to 2% of its source VM disk size since it uses blocks of variable size, in high bandwidth (HB) disk consumption is roughly 1% and fixed 4 KB block size is used.
Each digest file consumes up to 2% of its source VM disk size. VM disk size is calculated based on the allocated space, not used space. This means, for example, that a 2 TB VM disk file can produce a single set of digest files up to 40 GB in size.
WAN accelerator keeps 2 copies of digest files for each processed VM disk resulting from the previous session and current session. At the end of the session these 2 files are merged into 1. As we need to count for this, we will calculate 5% of source VM Disk size for digest files. Additionally, plan for 10 GB of working space for payloads and other temporary files.
o Formula: Digests low bandwidth = (Source data size in GB) * 2% • Example LB: with 2 TB source data: (2,000 GB * 2 %)= 40 GB o Formula: Digests high bandwidth = (Source data size in GB) * 1% • Example HB: with 2 TB source data: (2,000 GB * 1 %)= 20 GB
Note: The cache size setting in Low bandwidth on the source WAN accelerator will always be ignored, the digest files will be produced regardless of cache size setting configured. They may consume considerable disk space. Even if configuring the cache size on a source WAN accelerator is not as important, it still must exist as a number.
Another folder created on the source WAN accelerator working in Low bandwidth mode is the VeeamWAN\GlobalCache\src. The only file created in this directory is data.veeamdrf file. This file will be synchronized from the target WAN accelerator in following cases:
- Source WAN Accelerator cache was manually cleared, or digests deleted
- There was not enough space on the source WAN accelerator
- It is the first session after enabling the WAN accelerator setting
- If the job was seeded/mapped
The size of this file is typically up to 2% of the configured target cache size (see sizing target WAN); thus, it may take some time for the initial data transfer to begin
o Formula: data.veeamdrf = TargetCacheSize * 2%
The following recommendations apply to configuring a target WAN accelerator.
- The cache size setting configured on the target WAN accelerator will be applied per connected source WAN accelerator. This should be taken into account when sizing for many-to-one scenarios, as configuring 100 GB cache size will result in 100 GB multiplied by the number of source connections configured for each target WAN accelerator.
- It is recommended to configure the cache size at 10 GB for each unique operating system processed by the WAN accelerator.
Note: Each version of Microsoft Windows OS is considered to be unique, while all Linux distributions are treated as one OS.
Although a target WAN accelerator consumes less CPU resources than the source, the I/O requirements for the target side are higher. For each processed data block, the WAN accelerator will update the cache file (if required), or it may retrieve the data block from the target repository (if possible). The cache is used only for blocks belonging to Operating Systems and Microsoft Exchange, all the other blocks are processed only with the WAN optimized data reduction algorithm.
Target WAN accelerator IOPS
Tests show that there are no significant performance differences in using spinning disk drives as storage for the target WAN accelerator cache rather than flash storage. However, when multiple source WAN accelerators are connected to a single target WAN accelerator (many-to-one deployment), it is recommended to use SSD or equivalent storage for the target cache, as the I/O is now the sum of all the different sources.
Since a target WAN accelerator utilizing only high bandwidth mode does not use global cache, neither it runs resource–intensive operations, it can be configured with at least 4 CPU cores and 8GB RAM as per the minimal system requirements.
By default, there is no permanent data stored on the target WAN accelerator working in High-bandwidth mode. However, in case of many-to-one scenario and mixed mode, please ensure that you have made calculations for all the low-bandwidth sources connected to the same target WAN accelerator.
It is possible that target WAN accelerator will require space for temporary storage of digest files. These cases include a manual clear cache operation on the source WAN accelerator, a corrupted digest file or other operations which require a re-calculation of the digest files. In these cases, the digest will be re-calculated from the target side data and the calculated digest file - which is 1- 2% of the source data - will be transferred to the source WAN accelerator and removed from target WAN accelerator afterwards.
o Formula: Digests max size = (Source data size in GB) * 2% • Example: with 1 source 2 TB source data: (2,000 GB * 2 %)= 40 GB • Example: with 5 sources 2 TB source data: (5*2,000 GB * 2 %)= 200 GB
Digest re-calculation on target will always be 2% of source data, if source WAN mode is low.
If working in a many to one scenario you should increase the CPU/RAM on target depending on the number of sources and source modes.
Ensure that sufficient space has been allocated for global cache on the target WAN accelerator.
- Count at least 10 GB per each different OS that is backed up.
- Plan for additional 20 GB of working space for cache population, payload and other temporary files.
- If the cache is pre-populated, an additional temporary cache is created. The temporary cache will be converted into being the cache used for the first connected source. Subsequently connected sources will duplicate the cache of the first pair. As caches are duplicated the configured cache size is considered per pair of WAN accelerators.
• Formula for configured cache size (insert this number in configuration wizard):
o (Number of operating systems * 10 GB) + 20 GB
• Formula for used disk space (Many-to-one scenario)
o (Number of sources * <formula for configured cache size>)
• Example with one source and two operating systems:
o Configured cache size: (2 operating systems * 10 GB) + 20 GB = 40 GB o Used disk space: (1 source * 40 GB) = 40 GB
• Example with five sources and four operating systems:
o Configured cache size: (4 operating systems * 10 GB) + 20 GB = 60 GB o Used disk space: (5 sources * 60 GB) = 300 GB
In some cases space for digest data is also required on the target WAN accelerator. These cases include a manually clear cache operation on the source WAN accelerator, a corrupted digest file or other operations which require a re-calculation of the digest cache.
In these cases the digest will be re-calculated from the target side data and the calculated digest file - which is 2% of the source data - will be transferred to the source WAN accelerator.
If the required 2% of source data is not available on the target side WAN accelerator the transport will continue in the limited mode which will not do any more deduplication.
• Example with 2TB of source data
o 2TB source * 2% = 40GB digest data
• Example with 5 sources each 2TB of source data
o 5 * 2TB source * 2% = 5 * 40GB = 200 GB digest space
For understanding how the disk space is consumed, please see the following sections.
For each pair there will be a subfolder in the trg directory, with a UUID describing which source WAN accelerator the cache is attached to. In each of those subfolders, the
blob.bin file containing the cache will be located. That file size corresponds to the setting configured in the management console. Note: The
blob.bin file will exist for all connected source WAN accelerators.
When connecting a new source WAN accelerator, the temp folder will temporarily contain the
data.veeamdrf file that is later transferred to the source containing the cache manifest.