Skip to main content
NetApp Knowledge Base

What is an aggregate?

Views:
18,566
Visibility:
Public
Votes:
56
Category:
data-ontap-8
Specialty:
CORE
Last Updated:

 

Applies to

  • Clustered Data ONTAP
  • Data ONTAP 7-Mode

Answer

What is an aggregate?

An aggregate is a collection of disks (or partitions) arranged into one or more RAID groups.  It is the most basic storage object within ONTAP and is required to allow for the provisioning of space for connected hosts.

How is an aggregate created?
  • In 7-Mode, use the aggr create command with the desired parameters.
  • In Clustered Data ONTAP, use the storage aggregate create command from the cluster shell with the desired parameters

Note that in order to create an aggregate on a given node, all of the required disks must be owned by the respective node.

Can an aggregate contain disks from different shelves?

Yes, an aggregate can be created using disks across different shelves.

How do you view aggregate details?

There are several commands available to to see aggregate details.  The specific commands are determined by the type of information required.  The most common commands are as shown.

  • To determine the physical size of an aggregate, use df -A <aggregate name>  to see the space usage for the aggregate in question.
  • To see the physical disks that are in a given aggregate, the aggr status -r <aggregate name> command can be run with an output similar to :
7-Mode:

Example:

controller> aggr status -r aggr0
Aggregate aggr0 (online, raid_dp) (block checksums)
  Plex /aggr0/plex0 (online, normal, active)
    RAID group /aggr0/plex0/rg0 (normal)

      RAID Disk Device          HA  SHELF BAY CHAN Pool Type  RPM  Used (MB/blks)    Phys (MB/blks)
      --------- ------          ------------- ---- ---- ---- ----- --------------    --------------
      dparity   0a.16           0a    1   0   FC:A   -  FCAL 15000 136000/278528000  137104/280790184
      parity    0a.17           0a    1   1   FC:A   -  FCAL 15000 136000/278528000  137104/280790184
      data      0c.00.1         0c    0   1   SA:B   -  SAS  15000 136000/278528000  280104/573653840

Clustered Data ONTAP:

Example:

cluster1::> aggr status -r aggr1_node1
Owner_Node: cluster1-01
 Aggregate: aggr1_node1 (online, raid_dp) (block checksums)
  Plex: /aggr1_node1/plex0 (online, normal, active, pool0)
   RAID Group /aggr1_node1/plex0/rg0 (normal, block checksums)
                                                              Usable Physical
     Position Disk                        Pool Type     RPM     Size     Size Status
     -------- --------------------------- ---- ----- ------ -------- -------- ----------
     dparity  1.1.0                        0   BSAS    7200   2.42TB   2.43TB (normal)
     parity   1.1.1                        0   BSAS    7200   2.42TB   2.43TB (normal)
     data     1.1.2                        0   BSAS    7200   2.42TB   2.43TB (normal)
     data     1.1.3                        0   BSAS    7200   2.42TB   2.43TB (normal)
     data     1.1.5                        0   BSAS    7200   2.42TB   2.43TB (normal)
     data     1.1.6                        0   BSAS    7200   2.42TB   2.43TB (normal)
6 entries were displayed.​

To see specific aggregate details, use the following commands:

7-Mode:
controller> aggr status -v <aggr name>

Clustered Data ONTAP:
cluster1::> node run -node <node name> -command "aggr status -v <aggr name>"

or

cluster1::> aggr show <aggr name> -instance

What to do with a failed aggregate?

An aggregate may have failed when one or more RAID groups have gone offline. Before the aggregate can be brought back online, a root cause will need to be determined and fixed.
Please call NetApp Technical Support for assistance.

What to do with a foreign aggregate?

In clustered Data ONTAP, while adding a shelf from a different cluster (or a 7-Mode system) to a node, any existing aggregates, or parts of aggregates; on that shelf will be flagged as foreign and the drives belonging to that aggregate as orphaned disks. This issue can also be encountered after replacing a failed disk.

For example, an aggregate can show up as follows under the aggr status command:

aggr0(1) failed

This is caused due to the disks not being zeroed (and containing a RAID label) and therefore retaining the aggregate information from an earlier/different system. These foreign aggregates are visible only when using the node shell.

I created an aggregate but it does not show up in the list for creating volumes in OnCommand System Manager.

Check to see if the new aggregate is present in the list of aggregates available for this SVM:

cluster1::> vserver show -vserver <name> -fields aggr-list

What causes a degraded aggregate?

When a disk fails, Data ONTAP can continue to serve data, but it must reconstruct the data from the failed disk using RAID parity. If there are no available spare disks to start the reconstruction the aggregate will be degraded.

A RAID group goes into degraded mode in the following scenarios:

  • A single disk fails in a RAID4 group.
  • One or two disks fail in a RAID-DP RAID group.

Note: If an aggregate contains at least one RAID group with double disk failure that is not being reconstructed (RAID-DP aggregate only) it is double degraded.

  • A disk is taken offline by Data ONTAP (i.e. Disk Firmware Updates). 

After the offline disk is brought back online, the RAID group returns to normal mode.

How do I identify a degraded aggregate?

The following examples show a raid group with 1 failed disk and no available spares for reconstruction.

Clustered Data ONTAP:

Cluster::*> aggr show aggr1 -fields state,raidstatus
aggregate raidstatus                     state
--------- ------------------------------ ------
aggr1     raid_dp, reconstruct, degraded online


7-Mode:

Node> aggr status aggr1
           Aggr State           Status                Options
          aggr1 online          raid_dp, aggr         nosnap=on
                                reconstruct
                                degraded
                                64-bit

                Plex /aggr1/plex0: online, normal, active
                    RAID group /aggr1/plex0/rg0: reconstruction 5% completed, block checksums
    
    
Node> sysconfig -r                  (Note: aggr status -r provides the same output)
Aggregate aggr1 (online, raid_dp, reconstruct, degraded) (block checksums)
  Plex /aggr1/plex0 (online, normal, active, pool0)
    RAID group /aggr1/plex0/rg0 (reconstruction 17% completed, block checksums)


The following example shows a degraded aggregate with a degraded and double degraded raid group. 
(Note: to view the double degraded raid group you must run the commands from the nodeshell. These are the same commands used in 7-mode)

Cluster::*> aggr show aggr2 -fields state,raidstatus
aggregate raidstatus        state
--------- ----------------- ------
aggr2     raid_dp, degraded online

Cluster::*> node run -node node_name aggr status -r aggr2
Aggregate aggr2 (online, raid_dp, degraded) (block checksums)
  Plex /aggr2/plex0 (online, normal, active, pool0)
    RAID group /aggr2/plex0/rg0 (degraded, block checksums)

      RAID Disk Device          HA  SHELF BAY CHAN Pool Type  RPM  Used (MB/blks)    Phys (MB/blks)
      --------- ------          ------------- ---- ---- ---- ----- --------------    --------------
      dparity   FAILED                  N/A                        2538546/ -
      parity    3c.10.11        3c    10  11  SA:B   0  BSAS  7200 2538546/5198943744 2543634/5209362816
      data      3c.10.12        3c    10  12  SA:B   0  BSAS  7200 2538546/5198943744 2543634/5209362816
      data      3c.10.13        3c    10  13  SA:B   0  BSAS  7200 2538546/5198943744 2543634/5209362816
      data      3c.10.14        3c    10  14  SA:B   0  BSAS  7200 2538546/5198943744 2543634/5209362816

    RAID group /aggr2/plex0/rg1 (double degraded, block checksums)

      RAID Disk Device          HA  SHELF BAY CHAN Pool Type  RPM  Used (MB/blks)    Phys (MB/blks)
      --------- ------          ------------- ---- ---- ---- ----- --------------    --------------
      dparity   FAILED                  N/A                        2538546/ -
      parity    3c.10.16        3c    10  16  SA:B   0  BSAS  7200 2538546/5198943744 2543634/5209362816
      data      FAILED                  N/A                        2538546/ -
      data      3c.10.18        3c    10  18  SA:B   0  BSAS  7200 2538546/5198943744 2543634/5209362816
      data      3c.10.19        3c    10  19  SA:B   0  BSAS  7200 2538546/5198943744 2543634/5209362816

What to do if have degraded aggr?

Replace the failed disks. After the disks are reconstructed to a spare, the RAID group returns to normal mode.

 

NetApp provides no representations or warranties regarding the accuracy or reliability or serviceability of any information or recommendations provided in this publication or with respect to any results that may be obtained by the use of the information or observance of any recommendations provided herein. The information in this document is distributed AS IS and the use of this information or the implementation of any recommendations or techniques herein is a customer's responsibility and depends on the customer's ability to evaluate and integrate them into the customer's operational environment. This document and the information contained herein may be used solely in connection with the NetApp products discussed in this document.