xorbits.numpy.percentile#

xorbits.numpy.percentile(a, q, axis=None, out=None, overwrite_input=False, interpolation='linear', keepdims=False)[source]#

Compute the q-th percentile of the data along the specified axis.

Returns the q-th percentile(s) of the array elements.

Parameters
  • a (array_like of real numbers) – Input array or object that can be converted to an array.

  • q (array_like of float) – Percentage or sequence of percentages for the percentiles to compute. Values must be between 0 and 100 inclusive.

  • axis ({int, tuple of int, None}, optional) –

    Axis or axes along which the percentiles are computed. The default is to compute the percentile(s) along a flattened version of the array.

    Changed in version 1.9.0(numpy): A tuple of axes is supported

  • out (ndarray, optional) – Alternative output array in which to place the result. It must have the same shape and buffer length as the expected output, but the type (of the output) will be cast if necessary.

  • overwrite_input (bool, optional) – If True, then allow the input array a to be modified by intermediate calculations, to save memory. In this case, the contents of the input a after this function completes is undefined.

  • method (str, optional (Not supported yet)) –

    This parameter specifies the method to use for estimating the percentile. There are many different methods, some unique to NumPy. See the notes for explanation. The options sorted by their R type as summarized in the H&F paper 1 are:

    1. ’inverted_cdf’

    2. ’averaged_inverted_cdf’

    3. ’closest_observation’

    4. ’interpolated_inverted_cdf’

    5. ’hazen’

    6. ’weibull’

    7. ’linear’ (default)

    8. ’median_unbiased’

    9. ’normal_unbiased’

    The first three methods are discontinuous. NumPy further defines the following discontinuous variations of the default ‘linear’ (7.) option:

    • ’lower’

    • ’higher’,

    • ’midpoint’

    • ’nearest’

    Changed in version 1.22.0(numpy): This argument was previously called “interpolation” and only offered the “linear” default and last four options.

  • keepdims (bool, optional) –

    If this is set to True, the axes which are reduced are left in the result as dimensions with size one. With this option, the result will broadcast correctly against the original array a.

    New in version 1.9.0(numpy).

  • interpolation (str, optional) –

    Deprecated name for the method keyword argument.

    Deprecated since version 1.22.0(numpy).

Returns

percentile – If q is a single percentile and axis=None, then the result is a scalar. If multiple percentiles are given, first axis of the result corresponds to the percentiles. The other axes are the axes that remain after the reduction of a. If the input contains integers or floats smaller than float64, the output data-type is float64. Otherwise, the output data-type is the same as that of the input. If out is specified, that array is returned instead.

Return type

scalar or ndarray

See also

mean

median

equivalent to percentile(..., 50)

nanpercentile

quantile

equivalent to percentile, except q in the range [0, 1].

Notes

Given a vector V of length n, the q-th percentile of V is the value q/100 of the way from the minimum to the maximum in a sorted copy of V. The values and distances of the two nearest neighbors as well as the method parameter will determine the percentile if the normalized ranking does not match the location of q exactly. This function is the same as the median if q=50, the same as the minimum if q=0 and the same as the maximum if q=100.

The optional method parameter specifies the method to use when the desired percentile lies between two indexes i and j = i + 1. In that case, we first determine i + g, a virtual index that lies between i and j, where i is the floor and g is the fractional part of the index. The final result is, then, an interpolation of a[i] and a[j] based on g. During the computation of g, i and j are modified using correction constants alpha and beta whose choices depend on the method used. Finally, note that since Python uses 0-based indexing, the code subtracts another 1 from the index internally.

The following formula determines the virtual index i + g, the location of the percentile in the sorted sample:

\[i + g = (q / 100) * ( n - alpha - beta + 1 ) + alpha\]

The different methods then work as follows

inverted_cdf:

method 1 of H&F 1. This method gives discontinuous results:

  • if g > 0 ; then take j

  • if g = 0 ; then take i

averaged_inverted_cdf:

method 2 of H&F 1. This method give discontinuous results:

  • if g > 0 ; then take j

  • if g = 0 ; then average between bounds

closest_observation:

method 3 of H&F 1. This method give discontinuous results:

  • if g > 0 ; then take j

  • if g = 0 and index is odd ; then take j

  • if g = 0 and index is even ; then take i

interpolated_inverted_cdf:

method 4 of H&F 1. This method give continuous results using:

  • alpha = 0

  • beta = 1

hazen:

method 5 of H&F 1. This method give continuous results using:

  • alpha = 1/2

  • beta = 1/2

weibull:

method 6 of H&F 1. This method give continuous results using:

  • alpha = 0

  • beta = 0

linear:

method 7 of H&F 1. This method give continuous results using:

  • alpha = 1

  • beta = 1

median_unbiased:

method 8 of H&F 1. This method is probably the best method if the sample distribution function is unknown (see reference). This method give continuous results using:

  • alpha = 1/3

  • beta = 1/3

normal_unbiased:

method 9 of H&F 1. This method is probably the best method if the sample distribution function is known to be normal. This method give continuous results using:

  • alpha = 3/8

  • beta = 3/8

lower:

NumPy method kept for backwards compatibility. Takes i as the interpolation point.

higher:

NumPy method kept for backwards compatibility. Takes j as the interpolation point.

nearest:

NumPy method kept for backwards compatibility. Takes i or j, whichever is nearest.

midpoint:

NumPy method kept for backwards compatibility. Uses (i + j) / 2.

Examples

>>> a = np.array([[10, 7, 4], [3, 2, 1]])  
>>> a  
array([[10,  7,  4],
       [ 3,  2,  1]])
>>> np.percentile(a, 50)  
3.5
>>> np.percentile(a, 50, axis=0)  
array([6.5, 4.5, 2.5])
>>> np.percentile(a, 50, axis=1)  
array([7.,  2.])
>>> np.percentile(a, 50, axis=1, keepdims=True)  
array([[7.],
       [2.]])
>>> m = np.percentile(a, 50, axis=0)  
>>> out = np.zeros_like(m)  
>>> np.percentile(a, 50, axis=0, out=out)  
array([6.5, 4.5, 2.5])
>>> m  
array([6.5, 4.5, 2.5])
>>> b = a.copy()  
>>> np.percentile(b, 50, axis=1, overwrite_input=True)  
array([7.,  2.])
>>> assert not np.all(a == b)  

The different methods can be visualized graphically:

(Source code, png, hires.png, pdf)

../../../_images/xorbits-numpy-percentile-1.png

References

1(1,2,3,4,5,6,7,8,9,10)

R. J. Hyndman and Y. Fan, “Sample quantiles in statistical packages,” The American Statistician, 50(4), pp. 361-365, 1996

This docstring was copied from numpy.