ADP(v1) and ADPv2 in a nutshell, it’s delicious!

Ever since clustered Data ONTAP went mainstream over 7-Mode, the dedicated root aggregate tax has been a bone of contention for many, especially for those entry-level systems with internal drives. Can you imagine buying a brand new FAS2220 or FAS2520 and being told that not only are you going to lose two drives as spares, but also another six to your root aggregates? This effectively left you with four drives for your data aggregate, two of which would be devoted to parity. I don’t think so. Now, this is a bit of an extreme example that was seldom deployed. Hopefully you had a deployment engineer who cared about the end result and would use RAID-4 for the root aggregates and maybe not even assign a spare to one controller, giving you seven whole disks for your active-passive deployment. Still, this was kind of a shaft. In a 24-disk system deployed active-active, you’d likely get something like this:

Traditional cDOT

Enter ADP.

In the first version of ADP introduced in version 8.3, clustered Data ONTAP gained the ability to partition drives on systems with internal drives as well as the first two shelves of drives on All Flash FAS systems. What this meant was the dedicated root aggregate tax got a little less painful. In this first version of ADP, clustered Data ONTAP carved each disk into two partitions: a small one for the root aggregates and a larger one for the data aggregate(s). This was referred to as root-data or R-D partitioning. The smaller partition’s size depended on how many drives existed. You could technically buy a system with fewer than 12 drives, but the ADP R-D minimum was eight drives. By default, both partitions on a disk were owned by the same controller, splitting overall disk ownership in half.

8.3 ADP, R-D


You could change this with some advanced command-line trickery to still build active-passive systems and gain two more drive partitions’ worth of data. Since you were likely only building one large aggregate on your system, you could also accomplish this in System Setup if you told it to create one large pool. This satisfied the masses for a while, but then those crafty engineers over at NetApp came up with something better.

Enter ADPv2.

Starting with ONTAP 9, not only did ONTAP get a name change (7-Mode hasn’t been an option since version 8.2.3), but it also gained ADPv2 which carves the aforementioned data partition in half, or R-D2 (Root-Data,Data) sharing for SSDs. Take note of the aforementioned SSDs there, as spinning disks aren’t eligible for this secondary partitioning. In this new version, you get one drive back that you would have allocated to be a spare, and you also get two of the parity drives back, lessening the pain of the RAID tax. With a minimum requirement of eight drives and a maximum of 48, here are the three main scenarios for this type of partitioning.

12 Drives:

ADPv2, R-D2 ½ shelf

24 Drives:

ADPv2, R-D2 1 shelf

48 Drives:

ADPv2, R-D2 2 shelves

As you can see, this is a far more efficient way of allocating your storage that yields up to ~17% more usable space on your precious SSDs.

So that’s ADP and ADPv2 in a nutshell—a change for the better. Interestingly enough, the ability to partition disks has lead to a radical change in the FlashPool world called “Storage Pools,” but that’s a topic for another day.

3 thoughts on “ADP(v1) and ADPv2 in a nutshell, it’s delicious!

  1. A.

    Very nice article!

    Just for my understanding: what would happen if you went with the ADPv2 “24-drives” configuration and ran out of disk-space and needed to add a shelf?
    With that in mind would you:
    1) choose the “48-drives” setup from the get-go (splitting the SSDs between nodes) so you can add a shelf and continue on that road
    2) just add another shelf and create the same “24-drives layout” (sharing the SSDs between nodes) ?

    1. chrismaki Post author

      Sorry about the delay Aleks, you posted that just as I was going on holidays for the break. I haven’t been able to find any documentation on specifically this, but I would assume if you were expanding your aggregate, that ONTAP would just create a whole new aggregate since technically the disks being used aren’t the same size. I thought I once read you could add bigger drives to an existing aggregate and not have them right-sized, but I can’t find anything in writing. I’ll research this some more.

    2. chrismaki Post author

      Hi Aleks, I’ve had some time to research this and when using ADPv2, no matter what the initial compliment of disks, all disks added up until the first 48 will be carved up into R1D2, this is to conserve spares. Because of that, the root slices will be larger because that size is determined by the initial compliment of disks.


Leave a Reply

Your email address will not be published. Required fields are marked *