I am under distress, here's my scenario. iIve been using a Startech 8 disk usb enclousure in jbod setting and encrypted every disk with truecrypt (full disk encryption).
Yesterday i initiated a large data copy via explorer from one usb drive to another usb drive (inside the same enclosure) the copy ran for hours.
Later under that day eventviewer suddenly report
Code: Select all
The IO operation at logical block address 0x252c2bb30 for Disk 12 (PDO name: \Device\0000006e) was retried.
After this event, it was followed by a large number of:
Code: Select all
A corruption was discovered in the file system structure on volume S:
I didn't notice the issue until 3-4 hours after this message in the eventlog when i attempted to copy a file from one of the drives only to have it return "file not found" via network. Once i went into the server i noticed all the drives were mounted in truecrypt but none of the files were accessible. I figured it was a windows bug so i dismounted all the encrypted drives, and then rebooted the entire system.
Once back online 5 out of 8 drives are inaccessible. They all mount fine in truecrypt so the header appears valid, but i guess the MFT is broken on all 5 drives. What's even stranger is that the S: drive that windows reported in eventviewer was troublesome works perfectly fine.
Now to the extremly interesting part...
There are 8 drives in the enclosure, but only 2 different type of drives.
- 5 drives are of model ST800AS0002 also known as "Seagate Archive V2" installed in enclosure slots 1, 4, 5, 6, 7
- 3 drives are of model ST8000VN0022 also known as "Seagate Ironwolf" installed in enclosure slots 2, 3, 8
All the Seagate Archive V2 are inaccessible
, all of them appears to have broken MFT.
All of the Seagate Ironwolf are accessible and works
fine like nothing ever happened.
So what exactly happened here?
All of the usb drives were running in "Quick Removal" state, which means it should be safe for the drives to be removed without notice. I also have truecrypt backup headers to every drive but like i said before, the headers are fine, the drives mount fine.
Since all of these drives are 8TB size i've yet to manage to analyze one of them fully, i ran testdisk and it was searching MFT for about 15 hours until i manually cancelled it due to it not making any sense what so ever, the advertised size of 0/270000000000 made sense at the beginning making me think once 0 reaches the number advertised at the othe side of the "/" it would be done, but it surpassed this, and then it looped back to 0 and kept on going. (See screenshot for info on that: https://i.imgur.com/EJcfWF7.png
I am really out of luck here, i cannot believe this happened. I have taken every damn procaution against dataloss, i opted for a JBOD solution because of worry of bad raid card writing corrupt data to the array and making me loose everything, i felt like losing 1 drive is better then an entire array, i ran the enclosure in quick removal mode to ensure no corruption would occur if power outtage happens, it is also connected to a UPS and monitor smart on daily via email reports.
Even though i did all of this... this somehow happened? I have no idea how it happened, even if i can somehow salvage the problem and rebuild the MFT index i still am very unsure if i ever want to use the enclosure again. It has been running for over 1 year with zero problems and suddenly baam. 60% of the entire data index is gone.