Silicon Motion SM2256 Preview
A Closer Look At Silicon Motion's SM2256
SSD vendors who partner with Silicon Motion have the option to use its reference design or build their own configuration.
At this year's CES, we saw three SM2256-based R&D boards first-hand. At this time, Silicon Motion is preparing three TLC flash types for use with its controller. Again, today we have one with Samsung's 19nm TLC. The company has also demonstrated the SM2256 with Toshiba A19 TLC and SK Hynix's second-gen 1xnm flash.
Today marks the second time I've tested an early sample with the SM2256. At Flash Memory Summit last year Paul Alcorn, Jon Coulter and I had 24 hours with this controller paired with Toshiba A19 TLC. Samsung's 19nm TLC is a little faster, and the controller is more mature this time around.
I was a bit surprised when Silicon Motion displayed the first early sample at CES with Samsung TLC flash. With its own successful line of SSDs, Samsung doesn't just hand flash out to competitors. I’ve tested many client SSDs over
Stay On the Cutting Edge: Get the Tom's Hardware Newsletter
Get Tom's Hardware's best news and in-depth reviews, straight to your inbox.
Current page: A Closer Look At Silicon Motion's SM2256
Prev Page Introduction And Specifications Next Page Four-Corner Performance Testing-
Maxx_Power I hope this doesn't encourage manufacturers to use even lower endurance NAND in their SSDs.Reply -
InvalidError
It most likely will. Simply going from 20-22nm TLC to 14-16nm will likely ensure that with smaller trapped charge, increased leakage, more exotic dielectric materials to keep the first two in check.15495485 said:I hope this doesn't encourage manufacturers to use even lower endurance NAND in their SSDs. -
unityole this is the way where we are heading, only thing to get good grade SSD is probably to spend big. good grade as in HET MLC flash and pcie 3.0 performance.Reply -
alextheblue As long as they don't corrupt data, I'm not really freaking out over endurance. When a HDD died, that was a major concern. If an SSD that's a few years old starts to really reach the end of its useful life, as long as I can recover everything and dump it onto a new drive, I'm happy.Reply
If you have a really heavy workload though, by all means get a high-end unit. -
JoeMomma I am using a small ssd for Windows that is backup weekly. Saved my butt last week. Speed is key. I can provide my own reliability.Reply -
jasonkaler Here's an idea:Reply
While they're making 3d nand, why don't they add an extra layer for parity and then use the raid-5 algorithm?
e.g. 8 layers for data, adding 1 extra for parity. Not that much extra overhead, but data will be much more reliable. -
InvalidError
SSD/MMC controllers already use FAR more complex and rugged algorithms than plain parity. Parity only lets you detect single-bit errors. It only allows you to "correct" errors if you know where the error was, such as a drive failure in a RAID3/5 array. If you want to correct arbitrary errors without knowing their location beforehand, you need block codes and those require about twice as many extra bits as the number of correctable bit-errors you want to implement. (I say "about twice as many" because twice is the general requirement for uncorrelated, non-deterministic errors. If typical failures on a given media tend to be correlated or deterministic, then it becomes possible to use less than two coding bits per correctable error.)15517825 said:Here's an idea:
While they're making 3d nand, why don't they add an extra layer for parity and then use the raid-5 algorithm?
e.g. 8 layers for data, adding 1 extra for parity. Not that much extra overhead, but data will be much more reliable.
-
jasonkaler 15519330 said:
SSD/MMC controllers already use FAR more complex and rugged algorithms than plain parity. Parity only lets you detect single-bit errors. It only allows you to "correct" errors if you know where the error was, such as a drive failure in a RAID3/5 array. If you want to correct arbitrary errors without knowing their location beforehand, you need block codes and those require about twice as many extra bits as the number of correctable bit-errors you want to implement. (I say "about twice as many" because twice is the general requirement for uncorrelated, non-deterministic errors. If typical failures on a given media tend to be correlated or deterministic, then it becomes possible to use less than two coding bits per correctable error.)15517825 said:Here's an idea:
While they're making 3d nand, why don't they add an extra layer for parity and then use the raid-5 algorithm?
e.g. 8 layers for data, adding 1 extra for parity. Not that much extra overhead, but data will be much more reliable.
No you don't. Each sector has CRC right?
So if the sector read fails CRC, simply calculate CRC replacing each layer in turn with the raid parity bit.
All of them will be off except for one with the faulty bit.
And these CRC's can all be calculated in parallel so there would be 0 overhead with regard to time.
Easy huh?
-
InvalidError
If you know which drive/sector is bad thanks to a read error, your scheme is needlessly complicated: you can simply ignore ("erase") the known-bad data and calculate it by simply XORing all remaining volumes. But you needed the extra bits from the HDD's "CRC" to know that the sector was bad in the first place.15536005 said:No you don't. Each sector has CRC right?
So if the sector read fails CRC, simply calculate CRC replacing each layer in turn with the raid parity bit.
All of them will be off except for one with the faulty bit.
And these CRC's can all be calculated in parallel so there would be 0 overhead with regard to time.
Easy huh?
In the case of a silent error though, which is what you get if you have an even-count bit error when using parity alone, you have no idea where the error is or even that there ever was an error in the first place. That's why more complex error detection and correction block codes exist and are used wherever read/receive errors carry a high cost, such as performance, reliability, monetary cost or loss of data. -
Eggz I know this article was about the SM2256, but the graphs really made the SanDisk PRO shine bright! In the latency tests, which content creators care about, nothing seemed to phase it, doing better than even the 850 Pro - consistently!Reply
BUT, one significant critique I have was the density limitation. Everything here was based on the ~250 GB drives. Comparing a drive with the exact name, but in a different density, is akin to comparing two entirely different drives.
I realize producing the data can be time consuming, but having the same information at three density points would be extremely helpful for purchasing decisions - lowest, highest, and middle densities.