Phison has publicly disowned a circulated advisory that claimed Windows 11’s August cumulative update was uniquely “killing” Phison‑based SSDs, while the vendor — and several independent labs — simultaneously confirm an industry‑wide storage regression tied to the August 12, 2025 cumulative (KB5063878) that can make some NVMe drives disappear under sustained, large writes. (support.microsoft.com) (tomshardware.com)
Microsoft released the August 12, 2025 Windows 11 cumulative update (KB5063878, OS Build 26100.4946) as part of Patch Tuesday. The official update notes list security and quality fixes and initially reported no known storage regressions; within days, however, community test benches produced a reproducible failure fingerprint: under sustained sequential writes (commonly reported near the ~50 GB mark, often when the drive is ~50–60% full), certain NVMe SSDs stop responding, disappear from Windows, and in some instances return corrupted or remain inaccessible. (support.microsoft.com) (tomshardware.com)
Independent reproductions and specialist outlets converged on this fingerprint and raised a critical observation: while early samples over‑represented drives using Phison controllers, the phenomenon was not strictly limited to one vendor or one controller family — multiple vendors’ drives have appeared in community collations. This created an investigative posture: forensic correlation between Microsoft’s host telemetry and controller/drive vendor telemetry would be required to establish root cause and scope. (tomshardware.com)
Cross‑referencing the lab findings with Microsoft’s telemetry is the next crucial step. Microsoft’s KB for KB5063878 confirms the update’s release date and content but did not initially list storage regressions; Microsoft has engaged partners and is investigating reports surfaced via Feedback Hub and vendor channels. Historically, Microsoft has used Known Issue Rollback (KIR) or targeted mitigations where host behavior is implicated; those remain plausible remediation paths if telemetry points to the host stack. (support.microsoft.com, bleepingcomputer.com)
Key points for accountability:
Community posts reflect the anxiety and the pragmatic instincts of power users: advice ranges from “back up everything now” to “delay the update” to detailed test requests for Phison or other vendors to publish validated test results. Those requests are reasonable; however, publishing broad claims without vendor telemetry risks compounding harm.
Immediate steps (priority):
For users and admins the defensible posture is simple and urgent: back up critical data, avoid heavy sequential writes on patched systems until vendors publish mitigations, and stage updates in representative pilot rings before full deployment. For vendors and platform owners the durable lessons are procedural: better pre‑release stress matrices, faster authoritative advisories that include confirmed firmware IDs, and a standardized telemetry exchange that speeds root‑cause correlation without compromising user privacy.
This remains an active, evolving situation. Public posts and community reproductions provide crucial, early forensic leads; final root cause and durable fixes will require telemetry correlation across Microsoft, controller vendors, and SSD OEMs. Until firmware upgrades or OS mitigations are published and validated, prioritize backups and conservative update staging to minimize the risk of data loss. (tomshardware.com, bleepingcomputer.com)
Source: TechPowerUp Phison Responds to Falsified SSD Controller Issue Document Circulating Online
Background / Overview
Microsoft released the August 12, 2025 Windows 11 cumulative update (KB5063878, OS Build 26100.4946) as part of Patch Tuesday. The official update notes list security and quality fixes and initially reported no known storage regressions; within days, however, community test benches produced a reproducible failure fingerprint: under sustained sequential writes (commonly reported near the ~50 GB mark, often when the drive is ~50–60% full), certain NVMe SSDs stop responding, disappear from Windows, and in some instances return corrupted or remain inaccessible. (support.microsoft.com) (tomshardware.com)Independent reproductions and specialist outlets converged on this fingerprint and raised a critical observation: while early samples over‑represented drives using Phison controllers, the phenomenon was not strictly limited to one vendor or one controller family — multiple vendors’ drives have appeared in community collations. This created an investigative posture: forensic correlation between Microsoft’s host telemetry and controller/drive vendor telemetry would be required to establish root cause and scope. (tomshardware.com)
What circulated — the falsified Phison advisory and the vendor response
The forged advisory: form and impact
A document purporting to be an internal Phison advisory began circulating in partner channels and enthusiast forums. The document named specific controller families and used alarmist language about “permanent data loss,” implying that the Windows update uniquely and catastrophically affected Phison controllers. That message, once it spread, risked producing immediate and expensive consequences: panic returns or RMAs, flooded support channels, and reputational damage to an entire supplier ecosystem.Phison’s official posture
Phison publicly denounced the circulated advisory as falsified — stating it was neither an official nor unofficial company communication — and announced it would pursue “appropriate legal action” against those distributing the forged material. At the same time, Phison acknowledged it was investigating reports of “industry‑wide effects” associated with KB5063878 (and related preview KBs) and coordinating with partners and Microsoft to validate affected controller families and roll out remedial firmware or partner advisories as necessary. Treat statements of “legal action” as an intent communicated by Phison until public court filings or other legal notices are produced. (tomshardware.com)Why the fake advisory mattered
- It conflated community test lists and rumors with confirmed engineering telemetry.
- It directed blame to a single vendor before cross‑stack forensic correlation could occur.
- It created operational risks for vendors and integrators who might prematurely react or issue blanket recalls.
- It complicated triage by incentivizing users and partners to trust an unauthenticated internal memo rather than vendor‑verified advisories.
Technical anatomy: what the reproducible failure fingerprint suggests
Symptom cluster (what testers have reproducibly seen)
Multiple independent test benches reproduced a specific and consistent set of symptoms:- A sustained, large sequential write (practical heuristics point to ~50 GB continuous writes) against a moderately to heavily used NVMe SSD.
- At or near the transfer volume threshold, the target device stops responding and disappears from File Explorer, Device Manager, and vendor utilities; SMART/NVMe telemetry may become unreadable.
- In many cases a reboot restores the device; in a smaller subset the drive remains inaccessible or returns with corrupted file system metadata. (tomshardware.com)
Two leading technical hypotheses
- Host Memory Buffer (HMB) semantics and DRAM‑less designs
Many cost‑optimized NVMe SSDs omit dedicated DRAM and rely on Host Memory Buffer (HMB) to borrow system RAM for mapping tables. If the Windows update altered HMB allocation timing, initialization, or teardown semantics — or changed NVMe command pacing — it could expose race conditions in firmware that assume different lifetimes or ordering. That mismatch can produce controller hangs under heavy metadata churn. (tomshardware.com) - Sustained sequential‑write stress and mapping/Garbage Collection timing
Long, continuous writes stress SLC cache behavior, mapping updates and garbage collection. If the OS changed how it issues flushes, orders NVMe commands, or manages DMA buffers, it could create a command cadence outside of firmware’s tested envelope and surface latent timing bugs that result in a controller-level hang requiring a power cycle to recover. The observed unreadable SMART telemetry during incidents supports the controller‑hang hypothesis. (bleepingcomputer.com, tomshardware.com)
Why attribution is hard
- The same model and controller often behave differently across motherboards, UEFI versions, and driver stacks.
- Firmware revision differences and vendor SKU validation (BOM-level differences) change behavior.
- Public community lists are noisy and evolve as testers add or remove models after further validation.
- Definitive attribution requires secure controller logs and telemetry correlation between Microsoft and SSD vendors — data that only vendors and platform owners can reasonably aggregate at scale.
What the labs and community tests show — cross‑checking the data
A high‑profile community test (21 drives) reported many devices losing enumeration mid‑write and several unrecoverable failures; that independent benching, together with multiple other reproductions, moved the issue from scattered forum noise to an industry investigation. Tom’s Hardware, BleepingComputer, Windows Central and other outlets replicated or collated these results and documented practical heuristics — the ~50 GB continuous write threshold and the ~50–60% used capacity heuristics surfaced repeatedly across test benches. These numbers are operational heuristics, not immutable thresholds, but they have proved valuable for triage. (tomshardware.com, bleepingcomputer.com)Cross‑referencing the lab findings with Microsoft’s telemetry is the next crucial step. Microsoft’s KB for KB5063878 confirms the update’s release date and content but did not initially list storage regressions; Microsoft has engaged partners and is investigating reports surfaced via Feedback Hub and vendor channels. Historically, Microsoft has used Known Issue Rollback (KIR) or targeted mitigations where host behavior is implicated; those remain plausible remediation paths if telemetry points to the host stack. (support.microsoft.com, bleepingcomputer.com)
Responsibility: Windows, controller firmware, or both?
The evidence points to a cross‑stack interaction: the Windows update appears to have altered host behavior in ways that exercise latent firmware assumptions in some controllers. That does not absolve SSD controllers or their firmware of responsibility — controller firmware must be robust to normative host behavior — but it does mean pinpointing the primary cause requires telemetry from both sides.Key points for accountability:
- If a firmware race condition is triggered by perfectly valid NVMe commands issued by a modern OS, that is a firmware robustness issue requiring controller/firmware remediation.
- If Microsoft’s update introduced a non‑standard NVMe behavior or a regression in NVMe/HMB handling, Microsoft could mitigate via KIR or micro‑patch while vendors prepare firmware updates.
- In real environments the truth may be hybrid: an OS change exposes a firmware timing assumption, so durable correction requires both OS hardening and firmware fixes. (tomshardware.com, bleepingcomputer.com)
The practical fallout: why misinformation makes a bad situation worse
The forged Phison advisory distorted triage. When users, system builders, or smaller partners treat unauthenticated documents as authoritative, response actions can be misdirected:- Customers may rush warranty claims or returns for unaffected drives.
- Vendors may waste support cycles chasing spurious leads.
- Analysts and media risk amplifying incorrect attribution and increasing panic.
Community posts reflect the anxiety and the pragmatic instincts of power users: advice ranges from “back up everything now” to “delay the update” to detailed test requests for Phison or other vendors to publish validated test results. Those requests are reasonable; however, publishing broad claims without vendor telemetry risks compounding harm.
Actionable guidance — what users and IT teams should do now
Follow a conservative, risk‑first approach while vendors and Microsoft coordinate remediation.Immediate steps (priority):
- Back up irreplaceable data now. Use the 3‑2‑1 rule: three copies, on two different media, one offsite. Backups are the single most reliable defense against mid‑write corruption or permanent loss.
- If you have not installed KB5063878 (or the related KB preview) and you perform large sequential writes, delay deployment until vendors or Microsoft publish validated guidance. (tomshardware.com)
- If KB5063878 is installed, avoid sustained large single‑shot writes (> ~50 GB) where feasible; split large transfers into smaller batches. This is a practical mitigation while fixes are validated. (tomshardware.com)
- Inventory SSD models and firmware across your fleet using tools like CrystalDiskInfo or vendor utilities. Prioritize DRAM‑less or HMB‑dependent devices for testing and staging.
- Stage the update in a pilot ring that mirrors production storage hardware and workloads before pushing to broad deployment.
- If a drive disappears mid‑write: stop writing, collect Event Viewer logs and vendor diagnostics, and create a block‑level forensic image before attempting recovery or firmware flashes. Contact vendor support and provide logs. Imaging preserves recovery options.
- Some advanced workarounds (e.g., registry changes that alter HMB behavior) have been discussed in specialist channels. These carry performance penalties and risk; only experienced administrators should consider them and only with full backups in place.
What vendors and Microsoft should do — engineering and process recommendations
This incident is a reminder that modern storage reliability depends on co‑engineering across OS, drivers, and firmware. Practical steps that should become standard:- Expand pre‑release stress testing to include long sustained sequential writes and HMB allocation patterns across representative BIOS and driver permutations.
- Formalize a minimal cross‑vendor telemetry schema (privacy‑respecting) that enables rapid correlation between platform and controller events without exposing user data.
- Publish authoritative vendor advisories that list confirmed affected firmware IDs and SKUs rather than leaving partners to rely on noisy, crowd‑sourced lists.
- Use Microsoft’s Known Issue Rollback and targeted mitigations as a short‑term blast‑radius control while firmware fixes are validated and staged.
Critical analysis: strengths and risks in the current response
Strengths- The vendor posture (Phison and others) has generally been measured: acknowledge investigation, coordinate with partners, and avoid premature attribution. That measured approach reduces the risk of misguided mass recalls.
- Community reproducibility across independent benches provided a concrete failure fingerprint, giving vendors actionable data for cross‑stack correlation. (tomshardware.com)
- Communication gaps created a vacuum quickly filled by a falsified advisory. When vendor messaging is sparse on details like firmware IDs, community rumor fills the gap and can cause significant collateral damage.
- SKU‑specific firmware validation is slow and can create staggered remediation timelines across branding and retail SKUs. That leaves an operational exposure window for enterprise fleets.
- Public claims of “legal action” are useful to deter forgery, but they do not address the more immediate operational issues: telemetry sharing, validated advisories, and fast rollouts of firmware or OS mitigations.
Closing assessment and takeaways
The event is a textbook cross‑stack incident: a platform update coincided with a host‑behavior change that exposed latency and timing assumptions in some SSD controller firmware. Community labs reproduced a clear and actionable failure fingerprint (sustained writes near ~50 GB on partially filled drives), and that reproducibility moved vendors into coordinated investigation. Microsoft has acknowledged the update and engaged partners; Phison has publicly disowned a falsified advisory while pursuing remedies and investigating possible affected controllers with partners. (support.microsoft.com, tomshardware.com)For users and admins the defensible posture is simple and urgent: back up critical data, avoid heavy sequential writes on patched systems until vendors publish mitigations, and stage updates in representative pilot rings before full deployment. For vendors and platform owners the durable lessons are procedural: better pre‑release stress matrices, faster authoritative advisories that include confirmed firmware IDs, and a standardized telemetry exchange that speeds root‑cause correlation without compromising user privacy.
This remains an active, evolving situation. Public posts and community reproductions provide crucial, early forensic leads; final root cause and durable fixes will require telemetry correlation across Microsoft, controller vendors, and SSD OEMs. Until firmware upgrades or OS mitigations are published and validated, prioritize backups and conservative update staging to minimize the risk of data loss. (tomshardware.com, bleepingcomputer.com)
Source: TechPowerUp Phison Responds to Falsified SSD Controller Issue Document Circulating Online