“the medium is silica crystal, similar to optical cable, it’s highly durable. It’s also capacious: The technology can store up to 360 TB of data on a 5-inch glass platter.”
How hf can you have 5D space within 3D space? This sounds like marketing bullshit.
The 5D Memory Crystal stores data by using tiny voxels – 3D pixels – in fused silica glass, etched by femtosecond laser pulses. These voxels possess “birefringence,” meaning that their light refraction characteristics vary depending upon the polarization and direction of incoming light.
That difference in light orientation and strength can be read in conjunction with the voxel’s location (x, y, z coordinates), allowing data to be encoded in five dimensional space.
Oh, I get it now. It’s a five-dimensional mathematical space which is given by the three physical space dimensions plus the difference in light orientation and the difference the light strength.
@[email protected] 14,000,000,000 years
Removed by mod
And then again 13,000,000,000 years later.
I wonder what the read write speed is. Imagine storing your entire movie collection in a crystal the size of a coaster.
Might not be for home consumers anytime soon, article says: “In the next 18 months, the company hopes to have a field-deployable read device that customers can use to read archived data. But SPhotonix isn’t presently targeting the consumer market. Kazansky estimates that the initial cost of the read device will be about $6,000 and the initial cost of the write device will be about $30,000.”
Then goes on to mention they need about 3-4 years of R&D so they can be ready to license the tech
In case you missed it in the article, the transfer speeds are mentioned just two paragraphs prior to the one you cited:
Over the next three to four years, Kazansky said, SPhotonix aims to improve the data transfer speed of its technology from a write time of 4 megabytes per second (MBps) and read time of 30 MBps to a read/write speed of 500 MBps, which would be competitive with archival tape backup systems.
Writing 360 TB at 4 MB/s will take over 1000 days, almost 3 years. Retrieving 360 TB at a rate of 30 MB/s is about 138 days. That capacity to bitrate ratio that is going to be really hard to use in a practical way, and it’ll be critical to get that speed up. Their target of 500 MB/s is still more than 8 days to read or write the data from one storage platter.
One counterpoint - even with a weak speed to capacity ratio it could be very useful to have a lot of storage for incremental backup solutions, where you have a small index to check what needs to be backed up, only need to write new/modified data, and when restoring you only need to read the indexes and the amount you’re actually restoring. This saves time writing the data and lets you keep access to historical versions.
There’s two caveats here, of course, assuming those are not rewritable. One, you need to be able to quickly seek to the latest index, which can’t reliably be at the start, and two, you need a format that works without rewriting any data, possibly with a footer (like tar or zip, forgot which one), which introduces extra complexity (though I foresee a potential trick where the previous index can leave an unallocated block of data to write the address of the next index, to be written later)
If it’s slow, then it’s the central backup and you use anything else for regular use. Just having it as a fallback for recovery would be huge.
I’ll have a crystal collection that’s actually useful
“This one’s for memory.”
“You actually believe in that garbage?”
“No, you don’t understand…”
A friendly request - please de-clickbait your headlines and say what the material is (although you do mention it in your summary).
When a post is a link to an article, I would prefer that the post title match the article. Many news communities actually require that.
Remember that CDs, CDRs, and so on were originally pitched as surviving 100 years. Turns out they last a highly variable amount of time but potentially as little as 2-3 years before they degrade, depending on the construction.
So I’ll just say, this is clearly a theoretical value.
Edit: Words.
How can you be so sure they haven’t already done durability testing??
Because they weren’t invented in 1925? Any durability testing you do today is about assumptions where you accelerate the process for a year by heating it or exposing it to water or whatever will degrade it most to some factor above normal and then extrapolate. That extrapolation was wildly wrong with CDs and it could be with this medium too. Or it might last a lot longer. What they have not done is written to a bunch of them and stored them in a variety of ways for 100 years and concluded they last that long.
Oh, I see the confusion.
I was talking about the new media in the article, not CDs.
So… so are they. The new media has not yet been tested for 100 years because they were not invented 100 years ago





