# Source code for kuibit.series

```
#!/usr/bin/env python3
# Copyright (C) 2020-2024 Gabriele Bozzola
#
# This program is free software; you can redistribute it and/or modify it under
# the terms of the GNU General Public License as published by the Free Software
# Foundation; either version 3 of the License, or (at your option) any later
# version.
#
# This program is distributed in the hope that it will be useful, but WITHOUT
# ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
# FOR A PARTICULAR PURPOSE. See the GNU General Public License for more
# details.
#
# You should have received a copy of the GNU General Public License along with
# this program; if not, see <https://www.gnu.org/licenses/>.
"""The :py:mod:`~.series` module provides a base class :py:class:`~.BaseSeries`
for representing and handling series (from which :py:class:`~.TimeSeries` and
:py:class:`~.FrequencySeries` are derived).
:py:class:`~.BaseSeries` handles series that have a independent variable ``x``
and a dependent variable ``y``. The derived classes have to implement setters
and getters if they need to rename these variables (e.g. ``x -> t``). The
independent variable has to be monotonically increasing.
:py:class:`~.BaseSeries` implements several methods for operations on series.
Most of these methods are available in two flavors: those that return a new
:py:class:`~.BaseSeries`, and those which modify the object in place. The latter
have names with imperative verbs.
This module also provides the useful function :py:func:`~.sample_common`, which
takes a list of series and resamples them to their common points.
"""
import warnings
from typing import Tuple
import numpy as np
import numpy.typing as npty
from scipy import integrate, interpolate, signal
from kuibit.attr_dict import AttributeDictionary
from kuibit.numerical import BaseNumerical
from kuibit.tensor import Tensor
class _AttributeDictionaryNumPy(AttributeDictionary):
"""Class that maps a dictionary to attributes and that has a method
``to_numpy``.
This is used only to fake being Pandas so that Series can be plotted with
matplotlib. See comments in BaseSeries.
"""
def to_numpy(self):
# This method only makes sense in the context of faking pandas. See
# comments in BaseSeries about this.
return self.values
# Note, we test this class testing its derived class TimeSeries
[docs]class BaseSeries(BaseNumerical):
"""Base class (not intended for direct use) for generic series data in
which the independendent variable x is sorted.
This class is already rich of features.
.. note:
Derived class should define setters and getters to handle ``x``
and ``y``. This is where the data is stored.
The idea is the following. The actual data is stored in the
``BaseSeries` properties ``data_x`` and ``data_y``. These are
accessible from the derived classes. However, we don't want the
derived classes to use directly ``data_x`` and ``data_y``: they
should use something that clearly inform the user of their meaning,
like ``t`` or ``f`` (time or frequency). To do this, we have to
define getters and setters that access and modify ``data_x``
and ``y`` but use more meaningful names. To define a getters,
simply use the ``@property`` decorator:
.. code-block:: python
@property
def t(self):
return self.data_x
With these, ``.t`` will return ``self.data_x``. For a setter,
.. code-block:: python
@t.setter
def t(self, t):
# This is defined BaseClass
self.data_x = t
This is called when with ``.t = something``. Once these are defined,
the derived classes should use their getters and setters.
:ivar data_x: Independent variable.
:vartype data_x: 1D NumPy array or float
:ivar y: Dependent variable.
:vartype y: 1D NumPy array or float
:ivar spline_real: Coefficients for a spline represent of the real part
of y.
:vartype spline_real: Tuple
:ivar spline_imag: Coefficients for a spline represent of the real part
of y.
:vartype spline_imag: Tuple
"""
@staticmethod
def _make_array(x):
"""Return a NumPy array version of x (if x is not already an array)."""
return np.atleast_1d(x) if not isinstance(x, np.ndarray) else x
def _return_array_if_monotonic(self, x_array):
"""Return the import array if it has length 1 or if it is
monotonically increasing. Otherwise return error.
We assume x_array is an array. We will not check for this, it is up to
the developer to guarantee this. If this is not true, some errors will
be thrown.
:param x_array: Array to check if it is monotonically increasing.
:type x_array: 1d NumPy array
:returns: Input array, if increasing monotonically.
:rtype: 1d NumPy array
"""
if len(x_array) > 1:
# Here we compute directly the diff because it seems faster
# than using np.diff
# Example:
# self.x = [1,2,3]
# self.x[1:] = [2, 3]
# self.x[:-1] = [1, 2]
# dx = [1,1]
dx = x_array[1:] - x_array[:-1]
if dx.min() <= 0:
# HACK: To provide more useful information we assume
# that the derived classes are named like TimeSeries.
# Then, we remove the "series"
name = type(self).__name__
x_name = name[:-6]
raise ValueError(f"{x_name} not monotonically increasing")
return x_array
def __init__(self, x, y, guarantee_x_is_monotonic=False):
"""When guarantee_x_is_monotonic is True no checks will be perform to
make sure that x is monotonically increasing (increasing performance).
This should is used internally whenever a new series is returned from
self, since we have already checked that data_x is good.
:param x: Independent variable.
:type x: 1d NumPy array or list
:param y: Dependent variable.
:type y: 1d NumPy array or list
:param guarantee_x_is_monotonic: Whether we can skip the check on monotonicity.
:param guarantee_x_is_monotonic: bool
"""
x_array = self._make_array(x)
y_array = self._make_array(y)
if len(x_array) != len(y_array):
raise ValueError("Data length mismatch")
if len(x_array) == 0:
raise ValueError("Trying to construct empty Series.")
if not guarantee_x_is_monotonic:
x_array = self._return_array_if_monotonic(x_array)
# The copy is because we don't want to change the input values
self.__data_x = x_array.copy()
self.__data_y = y_array.copy()
# The data is stored in the members self.data_x and self.data_y. We
# will never access these directly. We have setters and getters to that
# we can do stuff when variables change. For example, we want to
# compute or update splines. The setter and getters are for x and y
# We keep this flag around to know when we have to recompute the
# splines. Operations that invalidate the splines MUST reset this flag
# to True.
self.invalid_spline = True
# Here we also define the splines as empty objects so that we know
# that they are attributes of the class and they are not uninitialized
self.spline_real = None
self.spline_imag = None
@property
def x(self):
return self.__data_x
@x.setter
def x(self, x):
x_array = self._make_array(x)
if len(x_array) != len(self.x):
raise ValueError("You cannot change the length of the series")
x_array = self._return_array_if_monotonic(x_array)
# This series should own the data, so we copy (to avoid accidentally
# changing some other variable).
# If you do self.x = z
# and then self.x = *2
# z will change (if we don't copy)
self.__data_x = x_array.copy()
# Invalidate the spline
self.invalid_spline = True
@property
def y(self):
return self.__data_y
@y.setter
def y(self, y):
y_array = self._make_array(y)
if len(y_array) != len(self.__data_y):
raise ValueError("You cannot change the length of the series")
# This series should own the data, so we copy (to avoid accidentally
# changing some other variable)
self.__data_y = y_array.copy()
# Invalidate the spline
self.invalid_spline = True
# Here is where we pretend to be pandas. We want to be able to plot our
# series with matplotlib. Unfortunately, there is no easy way to provide a
# custom object to the plot functions. However, matplotlib has a special
# hook for pandas in the function matplotlib.cbook.index_of. In this
# function is checked if the index property is available, in which case,
# index.values and values are returned. We use this to make our objects
# plottable.
# The function in matplotlib < 3.5.3 is:
# try:
# return y.index.values, y.values
# except AttributeError:
# y = _check_1d(y)
# return np.arange(y.shape[0], dtype=float), y
#
# It got updated to
# try:
# return y.index.to_numpy(), y.to_numpy()
# except AttributeError:
# pass
# try:
# y = _check_1d(y)
# except (np.VisibleDeprecationWarning, ValueError):
# # NumPy 1.19 will warn on ragged input, and we can't actually use it.
# pass
# else:
# return np.arange(y.shape[0], dtype=float), y
# raise ValueError('Input could not be cast to an at-least-1D NumPy array')
#
# Let us try to be compatible with both (luckily, we can easily do that).
#
# For the older version, if we provide index.values and values, we can
# return x and y.
#
# For the newer version, we also have to provide the correct methods named
# `to_numpy` to both this class and the object that is return by the index
# attribute. So, we define a _AttributeDictionaryNumPy class that does that.
[docs] def to_numpy(self):
"""Return the data as NumPy array. Equivalent to ``self.y``.
This function is here to enable compatibility matplotlib.
"""
return self.y
@property
def values(self):
"""Fake pandas properties, to make Series objects plottable by
matplotlib.
"""
return self.y
@property
def index(self):
"""Fake pandas properties, to make Series objects plottable by
matplotlib.
"""
return _AttributeDictionaryNumPy({"values": self.x})
@property
def mask(self):
"""Return where the data is valid (according to the mask).
:returns: Array of True/False of the same length of the data.
False where the data is valid, true where is not.
:rtype: 1D array of bool
"""
if self.is_masked():
return np.ma.getmask(self.y)
return np.zeros(len(self), dtype=bool)
@property
def xmin(self):
"""Return the minimum of the independent variable x.
:rvalue: Minimum of x.
:rtype: float
"""
return self.x[0]
@property
def xmax(self):
"""Return the maximum of the independent variable x.
:rvalue: Maximum of x
:rtype: float
"""
return self.x[-1]
[docs] def is_regularly_sampled(self):
"""Return whether the series is regularly sampled.
If the series is only one point, an error is raised.
:returns: Is the series regularly sampled?
:rtype: bool
"""
if len(self) == 1:
raise RuntimeError(
"Series is only one point, "
"it does not make sense to compute dx"
)
dx = self.x[1:] - self.x[:-1]
return np.allclose(dx, dx[0], atol=1e-14)
def __len__(self):
"""The number of data points."""
return len(self.x)
def __iter__(self):
for x, y in zip(self.x, self.y):
yield x, y
[docs] def is_complex(self):
"""Return whether the data is complex.
:returns: True if the data is complex, false if it is not.
:rtype: bool
"""
return issubclass(self.y.dtype.type, complex)
[docs] def is_masked(self) -> bool:
"""Return whether the x or y are masked.
:returns: True if the x or y are masked, false if it is not.
:rtype: bool
"""
return isinstance(self.y, np.ma.MaskedArray) or isinstance(
self.x, np.ma.MaskedArray
)
[docs] def x_at_maximum_y(self):
"""Return the value of x when y is maximum.
:returns: Value of x when y is maximum.
:rtype: float
"""
return self.x[np.argmax(self.y)]
[docs] def x_at_minimum_y(self):
"""Return the value of x when y is minimum.
:returns: Value of x when y is minimum.
:rtype: float
"""
return self.x[np.argmin(self.y)]
[docs] def x_at_abs_maximum_y(self):
"""Return the value of x when abs(y) is maximum.
:returns: Value of x when abs(y) is maximum.
:rtype: float
"""
return self.x[np.argmax(np.abs(self.y))]
[docs] def x_at_abs_minimum_y(self):
"""Return the value of x when abs(y) is minimum.
:returns: Value of x when abs(y) is minimum.
:rtype: float
"""
return self.x[np.argmin(np.abs(self.y))]
def _local_extrema(
self, y: npty.NDArray, *args, include_edges: bool = True, **kwargs
) -> Tuple[npty.NDArray]:
"""Use SciPy's ``find_peaks`` to find the local minima and maxima.
Unkown arguments are passed to ``find_peaks``.
If the signal is complex, the absolute value is taken.
If ``include_edges`` is True, the edges are considered among the
possible extrema.
:returns: Coordinate and value of the peaks.
:rtype: Tuple of NumPy arrays
"""
# TODO (FEATURE): Improve accuracy
#
# We can use splines or quadratic approximations to improve the location
# the peak (ie, find it within a cell)
offset = 0
if include_edges:
# https://stackoverflow.com/a/60096220
min_y = (min(y),)
y = np.concatenate((min_y, y, min_y))
offset = 1
peak_indices, _ = signal.find_peaks(y, *args, **kwargs)
# If we included the edges, we added an extra point, se we have to
# remove the offset
peak_indices = peak_indices - offset
return self.x[peak_indices], self.y[peak_indices]
[docs] def local_maxima(
self, *args, include_edges: bool = True, **kwargs
) -> Tuple[npty.NDArray]:
"""Use SciPy's ``find_peaks`` to find the local maxima.
Unkown arguments are passed to ``find_peaks``.
If the signal is complex, the absolute value is taken.
If ``include_edges`` is True, the edges are considered among the
possible maxima.
:returns: Coordinate and value of the peaks.
:rtype: Tuple of NumPy arrays
"""
y = abs(self.y) if self.is_complex() else self.y
return self._local_extrema(
y, *args, include_edges=include_edges, **kwargs
)
[docs] def local_minima(
self, *args, include_edges: bool = True, **kwargs
) -> Tuple[npty.NDArray]:
"""Use SciPy's ``find_peaks`` to find the local minima.
Unkown arguments are passed to ``find_peaks``.
If the signal is complex, the absolute value is taken.
If ``include_edges`` is True, the edges are considered among the
possible minima.
:returns: Coordinate and value of the minima.
:rtype: Tuple of NumPy arrays
"""
y = -abs(self.y) if self.is_complex() else -self.y
return self._local_extrema(
y, *args, include_edges=include_edges, **kwargs
)
def _make_spline(self, *args, k=3, s=0, **kwargs):
"""Private function to make spline representation of the data.
This function is not meant to be called directly.
``k`` is the degree of the spline fit. It is recommended to use cubic
splines. Even values of ``k`` should be avoided especially with small ``s``
values. 1 <= k <= 5
Unknown arguments are pass to ``scipy.interpolate.splrep``.
:param k: Order of the spline representation.
:type k: int
:param s: Smoothing of the spline.
:type s: float
"""
if len(self) < k:
raise ValueError(
f"Too few points to compute a spline of order {k}"
)
if self.is_masked():
raise RuntimeError("Splines with masked data are not supported.")
self.spline_real = interpolate.splrep(
self.x, self.y.real, k=k, s=s, *args, **kwargs
)
if self.is_complex():
self.spline_imag = interpolate.splrep(
self.x, self.y.imag, k=k, s=s, *args, **kwargs
)
self.invalid_spline = False
[docs] def evaluate_with_spline(self, x, ext=2):
"""Evaluate the spline on the points ``x``.
Values outside the interval are extrapolated if ``ext=0``, set to 0 if
``ext=1``, raise a ``ValueError`` if ``ext=2``, or if ``ext=3``, return
the boundary value.
This method is meant to be used only if you want to use a different ext
for a specific call, otherwise, just use __call__.
:param x: Array of x where to evaluate the series or single x.
:type x: 1D NumPy array of float
:param ext: How to deal values outside the bounaries. Values outside the
interval are extrapolated if ``ext=0``, set to 0 if
``ext=1``, raise a ValueError if ``ext=2``, or if ``ext=3``,
return the boundary value.
:type ext: int
:returns: Values of the series evaluated on the input x.
:rtype: 1D NumPy array or float
"""
if self.invalid_spline:
self._make_spline()
y_real = interpolate.splev(x, self.spline_real, ext=ext)
if self.is_complex():
y_imag = interpolate.splev(x, self.spline_imag, ext=ext)
ret = y_real + 1j * y_imag
else:
ret = y_real
# When this method is called with a scalar input, at this point, ret
# would be a 0d NumPy scalar array. What's that? - you may ask. I have
# no idea, but the user is expecting a scalar as output. Hence, we cast
# the 0d array into at "at_least_1d" array, then we can see its length
# and act consequently.
ret = np.atleast_1d(ret)
return ret if len(ret) > 1 else ret[0]
def __call__(self, x):
"""Evaluate the spline on the points x. If the value is outside the
range, a ValueError will be raised.
"""
# We call the spline only if we need to.
# TODO (REFACTORING): This is not a Pythonic way to write this function.
#
# The main problem is that it is is not vectorized. It is also not
# efficient, we are going over the array a lot of times, re-checking the
# same elements over and over. Also, we are not considering the floating
# point arithmetic, we should allow for some tolerance.
# First we consider the scalar case
if not hasattr(x, "__len__"):
if x in self.x:
return self.y[np.searchsorted(self.x, x)]
return self.evaluate_with_spline(x, ext=2)
ret = np.zeros(len(x), dtype=type(self.y[0]))
# Hash-maps are more efficient than searching every time through the
# array, but there is some overhead cost in defining the dictionary.
# Experiments show that it is sill more performant.
dic_data = dict(zip(self.x, self.y))
for index, elem in enumerate(x):
if elem in self.x:
ret[index] = dic_data[elem]
else:
ret[index] = self.evaluate_with_spline(elem, ext=2)
return ret
[docs] def copy(self):
"""Return a deep copy.
:returns: Deep copy of the series.
:rtype: :py:class:`~.BaseSeries` or derived class
"""
# The following is more complicated copy constructor that is designed
# to copy also the spline information without re-computing it.
# This can speed up some computations.
copied = type(self).__new__(self.__class__)
# We don't use the setters
# skipcq PYL-W0212
copied.__data_x = self.__data_x.copy()
# skipcq PYL-W0212
copied.__data_y = self.__data_y.copy()
if not self.invalid_spline:
# splines are tuples, with a direct call to the function
# tuple() we make a deep copy
copied.spline_real = tuple(self.spline_real)
if self.is_complex():
copied.spline_imag = tuple(self.spline_imag)
copied.invalid_spline = False
copied.invalid_spline = True
return copied
[docs] def resampled(self, new_x, ext=2, piecewise_constant=False):
"""Return a new series resampled from this to new_x.
You can specify the details of the spline with the method make_spline.
If you want to resample without using the spline, and you want a nearest
neighbor resampling, pass the keyword ``piecewise_constant=True``.
This may be a good choice for data with large discontinuities, where the
splines are ineffective.
:param new_x: New independent variable.
:type new_x: 1D NumPy array or list of float
:param ext: How to handle points outside the data interval.
:type ext: 0 for extrapolation, 1 for returning zero, 2 for ``ValueError``,
3 for extending the boundary
:param piecewise_constant: Do not use splines, use the nearest neighbors.
:type piecewise_constant: bool
:returns: Resampled series.
:rtype: :py:class:`~.BaseSeries` or derived class
"""
# If x is the same, there's no need to resample
if len(self.x) == len(new_x) and np.allclose(
self.x, new_x, atol=1e-14
):
return self.copy()
# Unfortunately there is no nearest neighor resampling in SciPy's splines.
# Hence, we use directly the method interp1d.
if piecewise_constant:
interp_function = interpolate.interp1d(
self.x, self.y, kind="nearest", assume_sorted=True
)
new_y = interp_function(new_x)
else:
new_y = self.evaluate_with_spline(new_x, ext=ext)
return type(self)(new_x, new_y)
[docs] def resample(self, new_x, ext=2, piecewise_constant=False):
"""Resample the series to new independent variable new_x.
If you want to resample without using the spline, and you want a nearest
neighbor resampling, pass the keyword ``piecewise_constant=True``.
This may be a good choice for data with large discontinuities, where the
splines are ineffective.
:param new_x: New independent variable.
:type new_x: 1D NumPy array or list of float
:param ext: How to handle points outside the interval.
:type ext: 0 for extrapolation, 1 for returning zero, 2 for ValueError,
3 for extending the boundary
:param piecewise_constant: Do not use splines, use the nearest neighbors.
:type piecewise_constant: bool
"""
self._apply_to_self(
self.resampled,
new_x,
ext=ext,
piecewise_constant=piecewise_constant,
)
def _apply_binary(self, other, function, *args, **kwargs):
"""This is an abstract function that is used to implement mathematical
operations with other series (if they have the same x) or
scalars.
:py:meth:`~._apply_binary` takes another object that can be of the same
type or a scalar, and applies ``function(self.y, other.y)``, performing type
checking.
:param other: Other object.
:type other: :py:class:`~.BaseSeries` or derived class or float
:param function: Dyadic function (function that takes two arguments).
:type function: callable
:returns: Return value of ``function`` when called with self and other.
:rtype: :py:class:`~.BaseSeries` or derived class (typically)
"""
# If the other object is of the same type
if isinstance(other, type(self)):
if (len(self.x) != len(other.x)) or (
not np.allclose(other.x, self.x, atol=1e-14)
):
raise ValueError("The objects do not have the same x!")
return type(self)(
self.x,
function(self.y, other.y, *args, **kwargs),
True,
)
# If it is a number
if isinstance(other, (int, float, complex)):
return type(self)(
self.x, function(self.y, other, *args, **kwargs), True
)
# If it is a Tensor of type(self), we have to return a Tensor
if isinstance(other, Tensor) and type(self) is other.type:
# We keep this at the high level
return type(other).from_shape_and_flat_data(
other.shape,
[
function(ot, self, *args, **kwargs)
for ot in other.flat_data
],
)
# If we are here, it is because we cannot add the two objects
raise TypeError("I don't know how to combine these objects")
def __eq__(self, other):
"""Check for equality up to numerical precision."""
if isinstance(other, type(self)):
return np.allclose(self.x, other.x, atol=1e-14) and np.ma.allclose(
self.y, other.y, atol=1e-14
)
return False
# From Python's docs: In order to conform to the object model, classes that
# define their own equality method should also define their own hash method,
# or be unhashable.
# We consider series unhashable. We could make them hashable by taking
# the combined hash of all the points, but this would not be useful.
__hash__ = None
def _apply_to_self(self, f, *args, **kwargs):
"""Apply the method ``f`` to ``self``, modifying ``self``.
This is used to transform the commands from returning an object to
modifying ``self``. The function ``f`` has to return a new copy of the
object (not a reference).
"""
ret = f(*args, **kwargs)
# We avoid the setters to avoid checking for consistency because this
# was already done
self.__data_x, self.__data_y = ret.x, ret.y
# We have to recompute the splines
self.invalid_spline = True
[docs] def save(self, file_name, *args, **kwargs):
"""Saves into simple ASCII format with 2 columns ``(x, y)``
for real valued data and 3 columns ``(x, Re(y), Im(y))``
for complex valued data.
Unknown arguments are passed to ``NumPy.savetxt``.
:param file_name: Path (with extension) of the output file.
:type file_name: str
"""
if self.is_masked():
warnings.warn(
"Discarding mask information.",
RuntimeWarning,
)
if self.is_complex():
np.savetxt(
file_name,
np.transpose(
(self.x, self.y.real, self.y.imag),
*args,
**kwargs,
),
)
else:
np.savetxt(
file_name,
np.transpose((self.x, self.y), *args, **kwargs),
)
[docs] def nans_removed(self):
"""Filter out nans/infinite values.
Return a new series with finite values only.
:returns: A new series with only finite values.
:rtype: :py:class:`~.BaseSeries` or derived class
"""
mask = np.isfinite(self.y)
return type(self)(self.x[mask], self.y[mask], True)
[docs] def nans_remove(self):
"""Filter out nans/infinite values."""
self._apply_to_self(self.nans_removed)
[docs] def mask_removed(self):
"""Remove masked value.
Return a new series with valid values only.
:returns: A new series with only valid values.
:rtype: :py:class:`~.BaseSeries` or derived class
"""
if self.is_masked():
mask = np.invert(self.mask)
return type(self)(
np.ma.compressed(self.x[mask]),
np.ma.compressed(self.y[mask]),
True,
)
# We can copy the spline
return self.copy()
[docs] def mask_applied(self, mask, ignore_existing=False):
"""Return a new series with given mask applied to the data.
If a previous mask already exists, the new mask will be added on top,
unless ``ignore_existing`` is True.
:param mask: Array of booleans that identify where the data is invalid.
This can be obtained with the method :py:meth:`~.mask`.
:type mask: 1D NumPy array
:param ignore_existing: If True, overwrite any previously existing mask.
:type ignore_existing: bool
:returns: New series with mask applied.
:rtype: :py:class:`~.BaseSeries`
"""
if self.is_masked() and not ignore_existing:
mask = np.ma.mask_or(mask, self.mask)
return type(self)(self.x, np.ma.MaskedArray(self.y, mask=mask), True)
[docs] def mask_apply(self, mask, ignore_existing=False):
"""Apply given mask.
If a previous mask already exists, the new mask will be added on top,
unless ``ignore_existing`` is True.
:param mask: Array of booleans that identify where the data is invalid.
This can be obtained with the method :py:meth:`~.mask`.
:type mask: 1D NumPy array
:param ignore_existing: If True, overwrite any previously existing mask.
:type ignore_existing: bool
"""
self._apply_to_self(
self.mask_applied, mask, ignore_existing=ignore_existing
)
[docs] def integrated(self, dx=None):
"""Return a series that is the integral computed with method of
the rectangles.
The spacing ``dx`` can be optionally provided. If provided, it will be
used (increasing performance), otherwise it will be computed internally.
:param dx: Delta x in the independent variable. If None it will be
computed internally.
:type dx: float or None
:returns: New series with the cumulative integral.
:rtype: :py:class:`~.BaseSeries` or derived class
"""
if self.is_masked():
raise RuntimeError(
"Integration with masked data is not supported."
)
# We pass self.x only if dx was not provided
passing_x = self.x if dx is None else None
return type(self)(
self.x,
integrate.cumtrapz(self.y, x=passing_x, dx=dx, initial=0),
True,
)
[docs] def integrate(self, dx=None):
"""Integrate series with method of the rectangles.
The spacing ``dx`` can be optionally provided. If provided, it will be
used (increasing performance), otherwise it will be computed internally.
"""
self._apply_to_self(self.integrated, dx=dx)
[docs] def spline_differentiated(self, order=1):
"""Return a series that is the derivative of the current one using
the spline representation.
The optional parameter ``order`` specifies the order of the derivative.
.. warning::
The values at the boundary are typically not accurate.
:param order: Order of derivative (e.g. 2 = second derivative).
:type order: int
:returns: New series with derivative
:rtype: :py:class:`~.BaseSeries` or derived class
"""
if (order > 3) or (order < 0):
raise ValueError(f"Cannot compute differential of order {order}")
if self.invalid_spline:
self._make_spline()
if self.is_complex():
ret_value = interpolate.splev(
self.x, self.spline_real, der=order
) + 1j * interpolate.splev(self.x, self.spline_imag, der=order)
else:
ret_value = interpolate.splev(self.x, self.spline_real, der=order)
return type(self)(self.x, ret_value, True)
[docs] def spline_differentiate(self, order=1):
"""Differentiate the series using the spline representation.
The optional parameter ``order`` specifies the order of the derivative.
.. warning::
The values at the boundary are typically not accurate.
:param order: Order of derivative (e.g. 2 = second derivative).
:type order: int
"""
self._apply_to_self(self.spline_differentiated, order)
[docs] def differentiated(self, order=1):
"""Return a series that is the numerical order-differentiation of
the present series.
The optional parameter ``order`` specifies the order of the derivative.
The derivative is calulated as centered differencing in the interior
and one-sided derivatives at the boundaries. Higher orders are computed
applying the same rule recursively.
:param order: Order of derivative (e.g. 2 = second derivative).
:type order: int
:returns: New series with derivative.
:rtype: :py:class:`~.BaseSeries` or derived class
"""
if self.is_masked():
raise RuntimeError(
"Differentiation with masked data is not supported."
)
ret_value = self.y
for _num_deriv in range(order):
ret_value = np.gradient(ret_value, self.x, edge_order=2)
return type(self)(self.x, ret_value, True)
[docs] def differentiate(self, order=1):
"""Differentiate with the numerical order-differentiation.
The optional parameter ``order`` specifies the order of the derivative.
The derivative is calulated as centered differencing in the interior
and one-sided derivatives at the boundaries. Higher orders are computed
applying the same rule recursively.
:param order: Order of derivative (e.g. 2 = second derivative).
:type order: int
"""
self._apply_to_self(self.differentiated, order)
[docs] def savgol_smoothed(self, window_size, order=3):
"""Return a smoothed series with a Savitzky-Golay filter with
window of size ``window_size`` and order ``order``.
This is just like a regular "Moving average" filter, but instead of
just calculating the average, a polynomial (usually 2nd or 4th order)
fit is made for every point, and only the "middle" point is chosen.
Since 2nd (or 4th) order information is concerned at every point, the
bias introduced in "moving average" approach at local maxima or minima,
is circumvented.
:param window_size: Number of points of the smoothing window (needs to
be odd).
:type window_size: int
:param order: Order of the filter.
:type order: int
:returns: New smoothed series.
:rtype: :py:class:`~.BaseSeries` or derived class
"""
if self.is_masked():
raise RuntimeError("Smoothing with masked data is not supported.")
if self.is_complex():
return type(self)(
self.x,
signal.savgol_filter(self.y.imag, window_size, order)
+ 1j * signal.savgol_filter(self.y.real, window_size, order),
True,
)
return type(self)(
self.x,
signal.savgol_filter(self.y, window_size, order),
True,
)
[docs] def savgol_smooth(self, window_size, order=3):
"""Smooth the series with a Savitzky-Golay filter with window of
size ``window_size`` and order ``order``.
This is just like a regular "Moving average" filter, but instead of
just calculating the average, a polynomial (usually 2nd or 4th order)
fit is made for every point, and only the "middle" point is chosen.
Since 2nd (or 4th) order information is concerned at every point, the
bias introduced in "moving average" approach at local maxima or minima,
is circumvented.
:param window_size: Number of points of the smoothing window (needs to
be odd).
:type window_size: int
:param order: Order of the filter.
:type order: int
"""
self._apply_to_self(self.savgol_smoothed, window_size, order)
[docs] def cropped(self, init=None, end=None):
"""Return a series with data removed outside the interval ``[init, end]``. If
``init`` or ``end`` are not specified or None, it does not remove
anything from this side.
:param init: Data with ``x <= init`` will be removed.
:type init: float or None
:param end: Data with ``x >= end`` will be removed.
:type end: float or None
:returns: Series with enforced minimum and maximum
:rtype: :py:class:`~.BaseSeries` or derived class
"""
x = self.x
y = self.y
if init is not None:
m = x >= init
x = x[m]
y = y[m]
if end is not None:
m = x <= end
x = x[m]
y = y[m]
return type(self)(x, y, True)
[docs] def crop(self, init=None, end=None):
"""Remove data outside the the interval ``[init, end]``. If
``init`` or ``end`` are not specified or None, it does not remove
anything from this side.
:param init: Data with ``x <= init`` will be removed.
:type init: float or None
:param end: Data with ``x >= end`` will be removed.
:type end: float or None
"""
self._apply_to_self(self.cropped, init, end)
# Define aliases
clip = crop
clipped = cropped
def _apply_unary(self, function, *args, **kwargs):
"""Apply a unary function to the data.
:param function: Function to apply to the series.
:type function: callable
:return: New series with function applied to the data.
:rtype: :py:class:`~.BaseSeries` or derived class
"""
return type(self)(self.x, function(self.y, *args, **kwargs), True)
def _apply_reduction(self, reduction, *args, **kwargs):
"""Apply a reduction to the data.
:param function: Function to apply to the series.
:type function: callable
:return: Reduction applied to the data
:rtype: float
"""
return reduction(self.y, *args, **kwargs)
[docs]def sample_common(series, resample=False, piecewise_constant=False):
"""Take a list of series and return new ones so that they are all defined on the
same points.
If ``resample`` is False (default), take as input a list of series and
return a new list with the same series but only defined on those points that
are common to all the lists. If ``resample`` is True, instead of removing
points, find the common interval of definition, and resample all the series
on that internal. The number of sample points is the minimum over all
series. Additionally, if ``piecewise_constant=True``, the approximant used
for resampling is a piecewise constant function, splines are not used,
instead, the nearest neighbors are used. Use this when you have series with
discontinuities.
:param series: The series to resample or redefine on the common points
:type series: list of :py:class:`~.Series`
:param resample: Whether to resample the series, or just find the common
points.
:type resample: bool
:param piecewise_constant: Whether to use the nearest neighbor resampling
method instead of splines.
If ``piecewise_constant=True``, the approximant used
for resampling is a piecewise constant function.
:type piecewise_constant: bool
:returns: Resampled series so that they are all defined in
the same interval.
:rtype: list of :py:class:`~.Series`
"""
# In many cases there is no real need for resampling because the array
# already have the desired shape. It is worth checking if we need to
# resample. If the series are regularly sampled, it is easy to check
# if the are the same. We also need to check that they are regularly
# sampled, to do this, we check that the first is regularly sampled,
# and that all the other ones have the same x.
s1, *s_others = series
if s1.is_regularly_sampled():
for s in s_others:
if not (len(s) == len(s1)):
break
if not np.allclose(s1.x, s.x, atol=1e-14):
break
# This is an else to the for loop
else:
# We have to copy, otherwise one can accidentally modify input data
return [ss.copy() for ss in series]
if resample:
# Find the series with max xmin
s_xmin = max(series, key=lambda x: x.xmin)
# Find the series with min xmax
s_xmax = min(series, key=lambda x: x.xmax)
# Find the series with min number of points
s_ns = min(series, key=len)
x = np.linspace(s_xmin.xmin, s_xmax.xmax, len(s_ns))
return [
s.resampled(x, piecewise_constant=piecewise_constant)
for s in series
]
def float_intersection(array_1, array_2):
"""Here we find the intersection between the two arrays also
considering the floating points.
"""
# https://stackoverflow.com/a/32516182
return array_2[
np.isclose(array_1[:, None], array_2, atol=1e-14).any(0)
]
# Here we find the common intersection between all the x, starting with
# the first one
x = s1.x
for s in s_others:
x = float_intersection(x, s.x)
if len(x) == 0:
raise ValueError("Series do not have any point in common")
return [s.resampled(x) for s in series]
```