Skip to main content

The roadmap for LTO tape has been extended up to generation 14 which is projected to store up to 576TB of uncompressed data and a unbelievable amount of 1440TB of compressed data on one tape.

It will be interesting to see if this can be realized and the capacity be doubled with each of the coming generations...

See more here:

https://www.lto.org/2022/09/lto-program-announces-extension-to-the-lto-tape-technology-roadmap-to-generation-14/

 

My libraries have 8 LTO8 drives in each currently.  I’ll most likely be at 12 per in the next short while.

They are beasts but the price tag wasn’t too bad for the pair.  It’s all relative when you need PB’s of backup space.

I’ve been a fan of the IBM TS4500 libraries.  you can have 12 Drives in frame one and 16 in each additional frame (up to 128 per library)

I think if you went 18 frames you are a max of 128 lto8 drives and 23k tapes or something crazy. It’s like 700PB compressed.  Even 660 LTO8 Tapes in a single frame with 12 drives is a monster backup system though. I’m happy my SAN’s are great and my FC switches allow me the throughput to really push this all as those tapes will hit their max data rates. 

 

In a previous life I was an IBM SSR. I replaced and repaired these things all the time. They seem complex and that robot gets a workout, but it’s actually all pretty straight forward, and knock on wood other than the odd gripper replacement here and there, they are super durable. 

 

 

 

 


Yes, I like the TS4500 and TS3500, too. 😎

Had 7 of them with 6 to 12 Frames and in sum nearly 300 Jaguar drives (no LTO), 2 roboter arms each and something around 15000 tapes at one customer. There were no HD frames at this time… Today there would be several frames less.


While at an IBM facility I saw a product only 3 customers in the world had. It was a chute that would allow tapes to go from one library to another over top of the isle.   I guess not too many people pushed for 40 frame systems to make it become more popular.   lol

 

I just got rid of a 3500 was a very solid machine. 


Nice, how much tapes can it manage?

 

 

For my Qualstar, reading the current specs, it looks like it would hold about 1700 tapes between the main unit and the MEM add-on unit.  I don’t think we every had it completely full as some tapes were always stored offsite in a secured location, but to say it was a lot was an understatement to me.  When I decommissioned it, we had issues with the robot being out of alignment again so we just had it unlock the door and manually removed the tapes.  We took one of those plastic rubbermaid carts and stacked them all up on top of it.  I think we removed 500-600 tapes from it which made for quite the heavy cart.


While at an IBM facility I saw a product only 3 customers in the world had. It was a chute that would allow tapes to go from one library to another over top of the isle.   I guess not too many people pushed for 40 frame systems to make it become more popular.   lol

 

I just got rid of a 3500 was a very solid machine. 

We had a look at this chute or bridge (don't remember the name of it) for this customer. But we decided against it, because the transport of the tape via this thing was rather slow.

But it was an interesting idea… 😀


offsite tapes are all well and good …..but how many people have either a) tested restores or b) actually had to restore in anger, 2-3 years down the line?

 

I have a client that I deployed tape to this summer.  Next week I’m going to be replacing their SAN (and upgraded their NAS’s that are used as Veeam Repo’s), but before we pull out the old SAN, they had this crazy/not so crazy idea to do a full restore of all VM’s from the tape to the old SAN to verify all is well.  

And how did the restore go? flawless?


offsite tapes are all well and good …..but how many people have either a) tested restores or b) actually had to restore in anger, 2-3 years down the line?

 

I have a client that I deployed tape to this summer.  Next week I’m going to be replacing their SAN (and upgraded their NAS’s that are used as Veeam Repo’s), but before we pull out the old SAN, they had this crazy/not so crazy idea to do a full restore of all VM’s from the tape to the old SAN to verify all is well.  

That is a crazy/not so crazy idea I may have to try.

 

I’ve made and presented some educated guesses with testing about how long a REAL DR situation will take us to get functional, semi functional and fully back to normal.  Between SRM, Veeam, restores and different scenarios wither it’s site down, ransomware etc. 

 

I often keep our old SAN’s for temp space, landing areas for things, “Unsupported” risky areas for people to stack things up in testing and labs etc.  I’ve even gone as far as running backups on these unsupported areas, but warn them that things can go south in a hurry if disks or controllers fail and at your own risk.

 

What I haven't done is a full tape restore of our production environment and timed it.   The restore alone will take quite a bit of time, and that doesn’t confirm anything is going to work when booted (apps talking to DC’s, talking to DB’s etc) but you could verify the tape jobs and have a time frame. 

 

It’s a good way to test the load on your servers and make sure your Veeam, SAN, NW and fiber infrastructure can handle it as well as that is a ton of data.  When I started at a previous job they did SRM “Tests” all the time and they passed with ease but never had to use it. One day I set up a test VM and volume with SAN replication like they had. Created a protection group and figured I’d flip the VM to the other side and it failed so bad it broke a few things. It ended up being wonky networking that wouldn’t show up in the testing. I’d rather know this BEFORE being in a critical situation and spending the time fixing issues that didn’t need to be there. 


offsite tapes are all well and good …..but how many people have either a) tested restores or b) actually had to restore in anger, 2-3 years down the line?

 

I have a client that I deployed tape to this summer.  Next week I’m going to be replacing their SAN (and upgraded their NAS’s that are used as Veeam Repo’s), but before we pull out the old SAN, they had this crazy/not so crazy idea to do a full restore of all VM’s from the tape to the old SAN to verify all is well.  

And how did the restore go? flawless?

Don't know yet.  SAN is getting installed in a couple hours.


can’t wait the LTO14! I just hope that the humidity conditions will not drop like with LTO9. more capacity but more sensitive!

I am convinced LTO development is pushed by hyperscaler (thx aws...) 🙂. we can consider that the majority of the data is cold so the power consumption is reduced with the use of tape.

I have quantum librairies (I6000) that are 15 years old and still updated with news drives etc, quite a profitable investment. I hope drives connectivities will be increase more than 8gb/s in new generation.

I love new object storage solutions with direct to tape or tiering.  Welcome replication between sites and performances with tiering!


can’t wait the LTO14! I just hope that the humidity conditions will not drop like with LTO9. more capacity but more sensitive!

I am convinced LTO development is pushed by hyperscaler (thx aws...) 🙂. we can consider that the majority of the data is cold so the power consumption is reduced with the use of tape.

I have quantum librairies (I6000) that are 15 years old and still updated with news drives etc, quite a profitable investment. I hope drives connectivities will be increase more than 8gb/s in new generation.

I love new object storage solutions with direct to tape or tiering.  Welcome replication between sites and performances with tiering!

With LTO10 being about 1100MB/s it’s tough to say if there will even be a requirement to go above 8 gig. Based on averages of previous gens, LTO13 will be about 3300MB/s, which means 14 could end up between 3300-5500.    still no requirement for 16 gig fiber but I’d assume they switch by then as 8 will be obsolete for most devices.   I really don’t like to  have more than 2 different speeds in my fabric if possible but it seems tape is that last 8 gig device kicking around :)

 


Mhh, no need for the fibre connection to one tape drive.

But in normal cases several tape sessions run over one fibre connection of a server. So, you could be in need of faster connections there….

And with disk storage with fibre connections this a complete different case, too….


Mhh, no need for the fibre connection to one tape drive.

But in normal cases several tape sessions run over one fibre connection of a server. So, you could be in need of faster connections there….

Right, but that isn’t limited right now.  I run 32 and 16 for most of my servers currently. I have 16’s in my tape proxy servers.  

 

BertrandFR said “. I hope drives connectivities will be increase more than 8gb/s in new generation.”

I personally don’t think it will be required until 15, but it might get implemented in 13\14 due to fiber having a negative effect if you are mixing speeds more than 2 generations apart.   Slapping a bunch of  8 gig LTO tape drives to 64gb fiber in the servers is not best practice. 

 

 


I’ll add, slower storage and faster hosts usually isn’t as much of an issues as faster storage and slower hosts. Especially if your zoning is good and you are not using ISL’s.

 

I often see people go out and buy an all flash SAN with 32 gig ports and connecting servers with 8 gig to it though. That can cause an issue. Here is 2 vids explaining.  It’s pretty dry so grab a coffee. haha

https://mediacenter.ibm.com/media/Mixing+Fibre-Channel+Speeds+on+the+Same+Fabric/1_6ecwy6ij/172212232

 

 


Yes, all ok, I am with you. My statement was about a SAN with all 8 or 16 or whatever connections. 😎
Mixing isn’t a good idea at all...

In such an environment it is no problem to have several connections to tape drives on on server connection...


Comment