Skip to main content
NetApp Knowledge Base

What are common performance terms?

Views:
3,491
Visibility:
Public
Votes:
6
Category:
data-ontap-8
Specialty:
Perf
Last Updated:

 

Applies to

  • Data ONTAP 8.X
  • ONTAP 9.X
  • Performance concepts and terms used generally in multiple ontap subsystems along with offbox metrics.

Answer

  • Throughput - Rate of data transmitted over a communication channel, often interchanged or confused with Bandwidth
    • Units
      • Ops/sec, or IOPS
      • Bytes/sec
      • MB/sec
      • GB/sec
  • Bandwidth - The maximum possible rate of data that can be transmitted over a communication channel, often interchanged or confused with Throughput
    • Units
      • Bytes/sec
      • MB/sec
      • GB/sec
  • Latency - The total time since an input or command is issued and the response is received
    • Units
      • seconds (sec)
      • milliseconds (ms)
      • microseconds (us)
  • Utilization - A measurement of the amount of time in a sample period that a given resource was utilized; utilization is a useful metric of performance, but for Data ONTAP should not be the primary metric
    • Units
      • %
  • Bottleneck - The point of congestion in a computing system that impacts performance, there might be more than one bottleneck in an environment
    • NetApp Technical Support looks to address the bottleneck contributing the most to overall latency first
  • Service Center - A processing point within a servicing system and the time associated to queue and process a request in that center.
  • Delay Center - The queue time in that delay center.
  • Network Processing - The network and protocol processing part of ONTAP.
    • Historically this was also called N-Blade, as it was a separate blade server before being combined with the D-Blade.
    • The N-Blade term survives and may be used interchangably, but is a legacy term.
    • The Network Exempt CPU domain is the primary processing point here.
  • Data Processing - The WAFL processing of operations, or the file system level handing of data.
    • The Network Processing node of ONTAP sends requests to the Data Processing node via the local Network Processing node.
      • Think of the node that owns the Data LIF as remote and the node that owns the disks as local.
    • Historically, this was called D-Blade for the same reason as N-Blade, and survives to this day in some documentation.
    • WAFL_Exempt is the primary CPU domain handled by Data Processing, but some other CPU domains are part of the Data Processing part of ONTAP.
  • Concurrency - Measurement of the parallelism of workload in a computing system
    • The more parallelism there is in a workload, the more simultaneous operations are “in flight” at any point in time
    • This allows the system to be more efficient in processing work, and complete more operations in less time even with the same latency per op as a low concurrency workload
    • Little’s Law shows the relationship between throughput, latency and concurrency in a steady state. Though it looks intuitively easy, it’s quite a remarkable result:
      • Throughput = Concurrency / Latency
      • Latency is controlled by Data ONTAP
      • Concurrency is controlled by the clients/applications
      • In order to achieve the best throughput, it should be considered to lower the latency and/or increase the concurrency

Example:

Assume a request that takes 1 milliseconds (ms) to complete. An application using one thread with one outstanding read or write operation should achieve 1000 IOPS (1 second or 1000 ms / 1 ms per request). Theoretically, if the thread count is doubled, then the application should be able to achieve 2000 IOPS. If the outstanding asynchronous read or write operations for each thread are doubled, then the application should be able to achieve 4000 IOPS. In practice, request rates do not always scale so linearly, due to overhead in the client from task scheduling, context switching, and so forth.

Note: This is an example showing how to optimize the throughput by increasing the concurrency from the client side, assuming that 1ms latency is already good enough and there is no room for further improvement from a latency perspective.

  • I/O or Block Size - Refers to the size of an Input/Output operation, which can be calculated by following equation:

I/O size = Throughput/IOPS

Higher the I/O size, the higher the throughput will be.

  • Randomness - Refers to a workload that is performed in an unpredicted sequence, with no order or pattern
  • Sequentiality - Refers to a workload that is performed in a predetermined, ordered sequence. Many patterns can be detected: forward, backward, skip counts, etc.

sequential_vs_random

Additional Information

 

 

Sign in to view the entire content of this KB article.

New to NetApp?

Learn more about our award-winning Support

NetApp provides no representations or warranties regarding the accuracy or reliability or serviceability of any information or recommendations provided in this publication or with respect to any results that may be obtained by the use of the information or observance of any recommendations provided herein. The information in this document is distributed AS IS and the use of this information or the implementation of any recommendations or techniques herein is a customer's responsibility and depends on the customer's ability to evaluate and integrate them into the customer's operational environment. This document and the information contained herein may be used solely in connection with the NetApp products discussed in this document.