Microsoft has quietly rewired a core part of Windows Server so modern NVMe SSDs can finally speak to the OS without being forced through a decades‑old SCSI translation layer — and the results, in Microsoft’s lab numbers, are eye‑watering: substantially higher IOPS and major CPU savings on NVMe‑attached storage.
For decades Windows exposed block storage through SCSI‑style semantics, an abstraction that made sense when spinning disks and SATA controllers dominated the market. NVMe (Non‑Volatile Memory Express), introduced to exploit flash media’s parallelism, arrived as a protocol that maps tightly to PCI Express and supports many submission/completion queues and very deep queue depths. Linux adopted native NVMe years ago; Windows, until now, largely continued to funnel NVMe I/O through a SCSI‑oriented path in the kernel. That mismatch created unnecessary serialization, locking, and CPU overhead as SSD throughput and parallelism outpaced the assumptions baked into the older stack. Microsoft’s response is a platform‑level modernization delivered in Windows Server 2025: a native NVMe storage path that bypasses the SCSI translation layer, exposes multi‑queue semantics to the OS, and rethinks how I/O work is scheduled across cores. The feature is shipped in servicing and is opt‑in by design; administrators must apply the servicing bundle and enable the policy to switch onto the new path. Independent coverage and community testing have largely confirmed the direction and headline numbers, although results vary by hardware, firmware, workloads and driver choices.
Native NVMe fixes three core mismatches:
That said, realized benefits will vary. Smaller shops and consumer‑grade systems are unlikely to see the same uplift as enterprise server deployments equipped with PCIe Gen‑5 drives and well‑tuned platforms. The change also elevates the importance of firmware and driver support from vendors — operating teams must maintain firmware governance and test matrices as part of normal change control.
Engineers who manage NVMe‑backed servers should plan a measured test-and‑rollout program: baseline, firmware and driver updates, reproduce Microsoft’s DiskSpd test to measure raw delta, and then validate representative application workloads under a staged deployment plan. When done correctly, the new native NVMe path can materially increase storage performance and efficiency — but only if you respect the complexity and test thoroughly before switching production fleets to the new stack.
Source: heise online Windows: Microsoft wants to massively improve SSD performance
Background
For decades Windows exposed block storage through SCSI‑style semantics, an abstraction that made sense when spinning disks and SATA controllers dominated the market. NVMe (Non‑Volatile Memory Express), introduced to exploit flash media’s parallelism, arrived as a protocol that maps tightly to PCI Express and supports many submission/completion queues and very deep queue depths. Linux adopted native NVMe years ago; Windows, until now, largely continued to funnel NVMe I/O through a SCSI‑oriented path in the kernel. That mismatch created unnecessary serialization, locking, and CPU overhead as SSD throughput and parallelism outpaced the assumptions baked into the older stack. Microsoft’s response is a platform‑level modernization delivered in Windows Server 2025: a native NVMe storage path that bypasses the SCSI translation layer, exposes multi‑queue semantics to the OS, and rethinks how I/O work is scheduled across cores. The feature is shipped in servicing and is opt‑in by design; administrators must apply the servicing bundle and enable the policy to switch onto the new path. Independent coverage and community testing have largely confirmed the direction and headline numbers, although results vary by hardware, firmware, workloads and driver choices. What Microsoft shipped — the essentials
Native NVMe: the high‑level change
- What changed: Windows Server 2025 now includes a native NVMe I/O path that removes the per‑I/O translation into SCSI and makes the kernel NVMe‑aware. This alignment reduces serialization and lets the host exploit NVMe’s per‑core queueing model.
- Delivery model: The capability landed via the servicing wave (an October cumulative update for WS2025) and is disabled by default. Administrators must enable it explicitly after installing the update.
- How to enable (official example): Microsoft published a registry/PowerShell example that toggles the feature after installing the LCU:
reg add HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Policies\Microsoft\FeatureManagement\Overrides /v 1176759950 /t REG_DWORD /d 1 /f
Microsoft also documented Group Policy paths for controlled deployment. Treat that registry command as the vendor‑documented opt‑in toggle and follow change‑control best practices.
Performance claims Microsoft published
Microsoft published microbenchmark artifacts and results to make the change reproducible:- IOPS uplift: In DiskSpd 4K random read microbenchmarks on NTFS, Microsoft reports up to ~80% higher IOPS for WS2025 with native NVMe versus WS2022 (specific test points).
- CPU efficiency: The same tests show roughly ~45% fewer CPU cycles per I/O on the cited workload/configuration.
- Absolute headroom examples: Microsoft illustrated PCIe Gen‑5 enterprise SSDs reaching multi‑million IOPS figures (for instance, ~3.3 million IOPS in one chart) when the software stack is not the limiting factor. These are useful demonstrations of headroom; actual numbers will depend on hardware, firmware, driver and topology.
Why this matters technically
NVMe was architected for flash: many independent queue pairs, per‑core queue affinity, and minimal submission/completion cost. The legacy SCSI‑centric path in Windows historically converted NVMe commands into SCSI semantics, which was fine for older devices but became a bottleneck as PCIe Gen‑4/Gen‑5 SSDs and NVMe‑capable HBAs increased parallelism and raw throughput.Native NVMe fixes three core mismatches:
- Queue model alignment: Exposes multi‑queue semantics instead of collapsing parallelism into a serialized path.
- Lower per‑I/O CPU cost: Fewer translations and lock‑heavy code paths free CPU cycles for application work instead of I/O housekeeping.
- Improved tail latency: Reduced kernel synchronization and fewer context switches improve both average latency and the high percentile (p99/p999) behavior that matters to databases and virtualization hosts.
Independent verification and third‑party reporting
Multiple outlets and community threads have reproduced or analyzed Microsoft’s announcement and microbenchmark claims. Heise noted the registry toggle and reproduced Microsoft’s key numbers (IOPS jump and CPU savings) in their coverage. Tom’s Hardware, among other respected trade outlets, framed the change as a major modernization and reiterated that gains are most pronounced on high‑parallelism hardware such as PCIe Gen‑5 NVMe drives. Community and early independent tests show a consistent trend: native NVMe reduces OS overhead and can unlock far higher IOPS on well‑matched enterprise drives, but the magnitude of uplift varies widely depending on:- Drive model and firmware
- Whether the in‑box StorNVMe.sys driver is used versus a vendor‑supplied driver
- Platform topology (PCIe lanes, CPU sockets, NUMA layout)
- Real workload characteristics (IO size, read/write mix, queue depth)
Where you’ll see the biggest gains
- PCIe Gen‑5 enterprise SSDs: These drives have the parallelism to show the largest deltas; Microsoft’s charts highlight moves from ~1.8M to ~3.3M IOPS in sample tests. Expect the biggest relative improvements on workloads that stress small random I/O and create high concurrency across cores.
- High‑concurrency server workloads: OLTP, high‑density virtualization (Hyper‑V boot storms, VDI), and large scale file‑serving where metadata and small‑block operations dominate will benefit most.
- Compute‑heavy nodes using NVMe scratch: Reduced CPU per‑I/O leaves more cycles for compute work in AI/ML training or data preprocessing scenarios.
Risks, caveats and practical compatibility concerns
This is a deep kernel change and must be treated like any other platform modernization: the upside is real, but so are the operational risks.- Opt‑in, for a reason: Microsoft delivered the new path disabled by default. That’s deliberate: registry toggles change kernel behavior and require validation. Use a staged rollout with clear rollback plans.
- Driver differences matter: If a system uses vendor‑supplied NVMe drivers (instead of the in‑box StorNVMe.sys family), the observed delta may be small or different. Some vendor drivers already implement NVMe optimizations tailored for their controllers. Compare vendor driver performance to the Microsoft in‑box driver during testing.
- Firmware and microcode interplay: NVMe drive firmware and platform BIOS/UEFI settings can dramatically affect behavior under heavy load. Test across firmware revisions.
- Cluster and replication interactions: If you use Storage Spaces Direct (S2D), NVMe‑oF, Storage Replica, or hyperconverged topologies, validate failover/repair/resync and replication behavior under the native NVMe path. Kernel I/O semantics affect recovery behavior as well as steady‑state performance.
- Microbenchmark vs. real workloads: DiskSpd 4K random read tests are excellent for measuring OS per‑I/O cost, but they’re synthetic. Application‑level gains require running representative workloads and measuring p50/p99/p999 latencies, CPU per‑IO, and end‑to‑end throughput.
- Potential for regressions: Large cumulative updates can include unrelated fixes and regressions. Organizations should be mindful that a single LCU can touch many subsystems; perform broad functional and regression testing alongside performance validation.
How to validate — practical checklist for engineers and admins
Follow a disciplined validation plan before enabling native NVMe in production.- Baseline everything:
- Collect pre‑change metrics: IOPS, throughput, average and high‑percentile latency, CPU utilization, queue depths, and application response times.
- Capture platform topology: CPU sockets, PCIe lane mapping, NUMA nodes, BIOS/UEFI settings.
- Update firmware and drivers:
- Apply vendor‑recommended NVMe firmware and platform BIOS/UEFI updates.
- Decide whether to use the Microsoft in‑box NVMe driver (StorNVMe.sys) or vendor drivers. Test both.
- Install servicing (the LCU that contains the native NVMe components) in a lab or canary ring.
- Enable native NVMe using the supported toggle after installing the update (example registry command provided by Microsoft): reg add HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Policies\Microsoft\FeatureManagement\Overrides /v 1176759950 /t REG_DWORD /d 1 /f. Document and automate the change via Group Policy or deployment tooling to standardize across hosts.
- Recreate Microsoft’s DiskSpd microbenchmark to observe the raw delta:
- Example DiskSpd invocation used by Microsoft: diskspd.exe -b4k -r -Su -t8 -L -o32 -W10 -d30. This is a useful reproducible test for small random I/O.
- Run representative real‑world workloads (databases, VM boot storms, file server traffic) and capture p50/p99/p999 latency, CPU per‑IO and application throughput.
- Validate edge cases: failover, resync, replication and NVMe‑oF behavior if applicable.
- Rollout plan: canary → pilot → phased production rings with rollback windows and runbooks.
Deployment best practices and monitoring
- Automate telemetry collection and use baseline comparison dashboards to track regressions.
- Monitor NVMe SMART attributes and OS counters for queue depths and I/O stalls as early warning signals.
- Keep close coordination with hardware vendors; vendor drivers or firmware updates may deliver complementary or competing optimizations.
- Maintain a rollback strategy: either remove the LCU or revert the policy toggle and verify application behavior and filesystem integrity before mass rollouts.
- Prefer staged rollouts in heterogeneous fleets: start with hosts that have enterprise‑grade NVMe drives and known firmware/driver compatibility, then expand.
Business and operational implications
For organizations that operate storage‑intensive workloads, freed CPU cycles and higher IOPS headroom can translate into measurable cost savings and higher VM or application density. Reduced CPU overhead per I/O means more cores available for business logic, potentially lowering infrastructure spend or increasing throughput without hardware upgrades.That said, realized benefits will vary. Smaller shops and consumer‑grade systems are unlikely to see the same uplift as enterprise server deployments equipped with PCIe Gen‑5 drives and well‑tuned platforms. The change also elevates the importance of firmware and driver support from vendors — operating teams must maintain firmware governance and test matrices as part of normal change control.
Final assessment — notable strengths and the risks you must manage
Strengths- Platform modernization at the right layer: Aligns Windows’ I/O model with NVMe’s design rather than bolting optimizations on top of SCSI emulation. That architectural fix scales across workloads.
- Tangible efficiency: Microsoft’s published results — significant IOPS uplifts and CPU savings — are reproducible in controlled lab environments and echoed by industry reporting.
- Future readiness: Exposes advanced NVMe features to the kernel and positions Windows to exploit next‑generation storage innovations.
- Heterogeneous results in the real world: Vendor drivers, firmware, and topology will determine whether you see the full gains. Some consumer drives and configurations show mixed results in early community reports.
- Operational risk of kernel changes: Enabling a new kernel I/O path should not be done casually in production; staged validation and rollback planning are essential.
- Vendor coordination required: Drive and platform vendors must be engaged to ensure supportability, especially for clustered and NVMe‑oF deployments.
Bottom line
Microsoft’s native NVMe in Windows Server 2025 is a long‑overdue, structural modernization that removes a software bottleneck for modern NVMe hardware. In Microsoft’s published lab tests and in independent reporting the result is clearer: on appropriately matched hardware you can expect substantially higher IOPS and meaningful CPU savings. The change is not a magic bullet — it’s an enabling technology that requires careful validation, coordination with vendors, and disciplined rollout practices.Engineers who manage NVMe‑backed servers should plan a measured test-and‑rollout program: baseline, firmware and driver updates, reproduce Microsoft’s DiskSpd test to measure raw delta, and then validate representative application workloads under a staged deployment plan. When done correctly, the new native NVMe path can materially increase storage performance and efficiency — but only if you respect the complexity and test thoroughly before switching production fleets to the new stack.
Source: heise online Windows: Microsoft wants to massively improve SSD performance