4

It's for a drive 200GB in size with only virtual machine images. Additional files may include a couple of xml files. The images themselves range from 5GB to 40GB in size.

Is 2MB a good idea or a terrible one? Logic dictates that large units mean less units to keep track of, but are there any performance hits I should be aware of?

I'm referring particularly to performance hits inside VMs, due to fragmentation when using smaller AUS. I was kind of hoping that a larger AUS would mean less seek time and a slightly increased performance.

I'm using Win10 and 2M is the max available AUS option.

Sorry if this question already exists. I couldn't find many that talk of units more than 64K, mine being in megabytes, I really needed to confirm before proceeding.

Glitch
  • 414

2 Answers2

1

Higher cluster sizes mean that the $Cluster file in the MFT is smaller, and that less indexes are needed to track data across the volume.

This translates to an increase in disk space but since this isn't the mid-90's anymore, it's probably not worth it--at least not for a 200GB drive.

Regarding performance, it may be improved slightly or at least not affected if data tends to be accessed sequentially (like playing video/music) or in chunks around the size of the cluster.

This may not apply with VMs. If the vmdks are not also both at a 2M cluster size and aligned with the disks clusters, I would think random access might be hurt by large cluster sizes as you're asking NTFS to load 2M of data when just one 4096 or 512-byte block in that sector may be needed.

You're in a similar situation with newer "4K" format hard drives - they internally read/write in 4096-byte chunks, but still allow the OS to request 512-byte chunks. If your data is not 4096-byte aligned it will be double-reading on many requests. OSes align the data now so probably not something you need to think about with--but I am saying your situation above with VMs and changing cluster sizes could be creating a similar problem.

LawrenceC
  • 75,182
0

Running benchmarks with different allocation size showed me empirically that 4k will give you the best performance. This is because the actual sector size on modern HDDs is 4k. I did the test on a 5TB drive using files that are 1GB min in size.

In general, if the cluster size exceeds the size of the IO, certain workflows can trigger unintended IOs to occur. Even if no extra IO occurs, you don't benefit from any speedup, since in the end it's translated to 4k final writes or read on the actual disk.

You might get less fragmentation over time, but I'm assuming your drive is an SSD given that it's only 200 GB, so fragmentation isn't really an issue.