hyperspy.learn.mva module

class hyperspy.learn.mva.LearningResults

Bases: object

Stores the parameters and results from a decomposition.

crop_decomposition_dimension(n, compute=False)

Crop the score matrix up to the given number.

It is mainly useful to save memory and reduce the storage size

Parameters
  • n (int) – Number of components to keep.

  • compute (bool, default False) – If True and the decomposition results are lazy, also compute the results.

load(filename)

Load the results of a previous decomposition and demixing analysis.

Parameters

filename (string) – Path to load the results from.

save(filename, overwrite=None)

Save the result of the decomposition and demixing analysis.

Parameters
  • filename (string) – Path to save the results to.

  • overwrite ({True, False, None}, default None) – If True, overwrite the file if it exists. If None (default), prompt user if file exists.

summary()

Summarize the decomposition and demixing parameters.

Returns

String summarizing the learning parameters.

Return type

str

class hyperspy.learn.mva.MVA

Bases: object

Multivariate analysis capabilities for the Signal1D class.

_calculate_recmatrix(components=None, mva_type='decomposition')

Rebuilds data from selected components.

Parameters
  • components (None, int, or list of ints) –

    • If None, rebuilds signal instance from all components

    • If int, rebuilds signal instance from components in range 0-given int

    • If list of ints, rebuilds signal instance from only components in given list

  • mva_type (str {'decomposition', 'bss'}) – Decomposition type (not case sensitive)

Returns

Data built from the given components.

Return type

Signal instance

_cluster_analysis(scaled_data, algorithm)

Cluster analysis of a scaled data - internal

Parameters
  • n_clusters (int) – Number of clusters to find.

  • scaled_data (numpy array - (number_of_samples,number_of_features)) –

  • algorithm (scikit learn clustering object) –

  • **kwargs – Additional parameters passed to the clustering algorithm. This may include n_init, the number of times the algorithm is restarted to optimize results.

Returns

return the sklearn.cluster object

Return type

alg

_distances_within_cluster(cluster_data, memberships, squared=True, summed=False)

Return inter cluster distances.

Parameters
  • cluster_data (ndarray) – scaled cluster data

  • memberships (ndarray) – cluster labels

  • squared (bool, optional) – square distance measurement. The default is True.

  • summed (bool, optional) – If False returns array showing sum of distance from a given point to all other points in the cluster. If True returns a sum of all distances within a cluster. The results are scaled by 2*number of cluster points. The default is False.

Returns

result – list of distances for within the cluster

Return type

list

_get_cluster_algorithm(algorithm, **kwargs)

Convenience method to lookup cluster algorithm if algorithm is a string and instantiates it with n_clusters or if it’s an object check that the object has a fit method

_get_cluster_preprocessing_algorithm(algorithm, **kwargs)

Convenience method to lookup method if algorithm is a string or if it’s an object check that the object has a fit_transform method

_get_cluster_signal(cluster_source, number_of_components=None, navigation_mask=None, signal_mask=None)

A cluster source can be an external signal, the signal data or the decomposition or bss results Return a flatten version of the data, nav and signal mask

Parameters
  • cluster_source (str or BaseSignal) – “decomposition”, “bss”, “signal” or a Signal

  • number_of_components (int, optional) – Number of components to use with decomposition sources. The default is None.

  • navigation_mask (ndarray, optional) – mask used to select regions of the cluster_source to use. The default is None.

  • signal_mask (ndarray, optional) – mask used to select regions of the cluster_source signal. For decomposition or bss this is not used. The default is None.

  • reproject (bool, optional) – If False the and the cluster_source is decomposition or bss the loadings are returned. If True the factor @ loadings result is used. The default is False.

Returns

toreturn – Returns an unfolded dataset from the selected cluster_source

Return type

ndarray

_get_number_of_components_for_clustering()

Returns the number of components

_scale_data_for_clustering(cluster_signal, preprocessing='norm', preprocessing_kwargs={})

Scale data for cluster analysis

Results are stored in learning_results.

Parameters
  • cluster_signal ({"bss", "decomposition", "signal", Signal}) – If “bss” the blind source separation results are used If “decomposition” the decomposition results are used if “signal” the signal data is used (signal should be unfolded)

  • preprocessing ({"standard","norm","minmax",None or scikit learn preprocessing method}) – default: ‘norm’ Preprocessing the data before cluster analysis requires preprocessing the data to be clustered to similar scales. Standard preprocessing adjusts each feature to have uniform variation. Norm preprocessing adjusts treats the set of features like a vector and each measurement is scaled to length 1. You can also pass a cikit-learn preprocessing object See scaling methods in scikit-learn preprocessing for further details.

  • preprocessing_kwargs – Additional parameters passed to the cluster preprocessing algorithm. See sklearn.preprocessing preprocessing methods for further details

See also

Returns

  • scaled_data (numpy array - unfolded array of shape (number_of_samples,)

  • no_of_features) scaled according to the selected algorithm

blind_source_separation(number_of_components=None, algorithm='sklearn_fastica', diff_order=1, diff_axes=None, factors=None, comp_list=None, mask=None, on_loadings=False, reverse_component_criterion='factors', whiten_method='PCA', return_info=False, print_info=True, **kwargs)

Apply blind source separation (BSS) to the result of a decomposition.

The results are stored in self.learning_results.

Read more in the User Guide.

Parameters
  • number_of_components (int or None) – Number of principal components to pass to the BSS algorithm. If None, you must specify the comp_list argument.

  • algorithm ({"sklearn_fastica", "orthomax", "FastICA", "JADE", "CuBICA", "TDSEP", custom object}, default "sklearn_fastica") – The BSS algorithm to use. If algorithm is an object, it must implement a fit_transform() method or fit() and transform() methods, in the same manner as a scikit-learn estimator.

  • diff_order (int, default 1) – Sometimes it is convenient to perform the BSS on the derivative of the signal. If diff_order is 0, the signal is not differentiated.

  • diff_axes (None or list of ints or strings) –

    • If None and on_loadings is False, when diff_order is greater than 1 and signal_dimension is greater than 1, the differences are calculated across all signal axes

    • If None and on_loadings is True, when diff_order is greater than 1 and navigation_dimension is greater than 1, the differences are calculated across all navigation axes

    • Otherwise the axes can be specified in a list.

  • factors (BaseSignal or numpy array) – Factors to decompose. If None, the BSS is performed on the factors of a previous decomposition. If a Signal instance, the navigation dimension must be 1 and the size greater than 1.

  • comp_list (None or list or numpy array) – Choose the components to apply BSS to. Unlike number_of_components, this argument permits non-contiguous components.

  • mask (BaseSignal or subclass) – If not None, the signal locations marked as True are masked. The mask shape must be equal to the signal shape (navigation shape) when on_loadings is False (True).

  • on_loadings (bool, default False) – If True, perform the BSS on the loadings of a previous decomposition, otherwise, perform the BSS on the factors.

  • reverse_component_criterion ({"factors", "loadings"}, default "factors") – Use either the factors or the loadings to determine if the component needs to be reversed.

  • whiten_method ({"PCA", "ZCA", None}, default "PCA") – How to whiten the data prior to blind source separation. If None, no whitening is applied. See whiten_data() for more details.

  • return_info (bool, default False) – The result of the decomposition is stored internally. However, some algorithms generate some extra information that is not stored. If True, return any extra information if available. In the case of sklearn.decomposition objects, this includes the sklearn Estimator object.

  • print_info (bool, default True) – If True, print information about the decomposition being performed. In the case of sklearn.decomposition objects, this includes the values of all arguments of the chosen sklearn algorithm.

  • **kwargs (extra keyword arguments) – Any keyword arguments are passed to the BSS algorithm.

Returns

return_info

  • If True and ‘algorithm’ is an sklearn Estimator, returns the Estimator object.

  • Otherwise, returns None

Return type

sklearn.Estimator or None

See also

  • plot_bss_factors()

  • plot_bss_loadings()

  • plot_bss_results()

cluster_analysis(cluster_source, source_for_centers=None, preprocessing=None, preprocessing_kwargs={}, number_of_components=None, navigation_mask=None, signal_mask=None, algorithm=None, return_info=False, **kwargs)

Cluster analysis of a signal or decomposition results of a signal Results are stored in learning_results.

Parameters
  • cluster_source ({"bss", "decomposition", "signal", BaseSignal}) – If “bss” the blind source separation results are used If “decomposition” the decomposition results are used if “signal” the signal data is used Note that using the signal or BaseSignal can be memory intensive and is only recommended if the Signal dimension is small BaseSignal must have the same navigation dimensions as the signal.

  • source_for_centers ({None,"decomposition","bss","signal",BaseSignal},) – default : None If None the cluster_source is used If “bss” the blind source separation results are used If “decomposition” the decomposition results are used if “signal” the signal data is used BaseSignal must have the same navigation dimensions as the signal.

  • preprocessing ({"standard","norm","minmax",None or scikit learn preprocessing method}) – default: ‘norm’ Preprocessing the data before cluster analysis requires preprocessing the data to be clustered to similar scales. Standard preprocessing adjusts each feature to have uniform variation. Norm preprocessing adjusts treats the set of features like a vector and each measurement is scaled to length 1. You can also pass one of the scikit-learn preprocessing scale_method = import sklearn.processing.StandadScaler() preprocessing = scale_method See preprocessing methods in scikit-learn preprocessing for further details.

  • preprocessing_kwargs (dict) – Additional parameters passed to the supported sklearn preprocessing methods. See sklearn.preprocessing scaling methods for further details

  • number_of_components (int, default None) – If you are getting the cluster centers using the decomposition results (cluster_source_for_centers=”decomposition”) you can define how many components to use. If set to None the method uses the estimate of significant components found in the decomposition step using the elbow method and stored in the learning_results.number_significant_components attribute. This applies to both bss and decomposition results.

  • navigation_mask (boolean numpy array) – The navigation locations marked as True are not used.

  • signal_mask (boolean numpy array) – The signal locations marked as True are not used in the clustering for “signal” or Signals supplied as cluster source. This is not applied to decomposition results or source_for_centers (as it may be a different shape to the cluster source)

  • algorithm ({ "kmeans" | "agglomerative" | "minibatchkmeans" | "spectralclustering"}) – See scikit-learn documentation. Default “kmeans”

  • return_info (bool, default False) – The result of the cluster analysis is stored internally. However, the cluster class used contain a number of attributes. If True (the default is False) return the cluster object so the attributes can be accessed.

  • **kwargs (dict optional, default - empty) – Additional parameters passed to the clustering class for initialization. For example, in case of the “kmeans” algorithm, n_init can be used to define the number of times the algorithm is restarted to optimize results.

  • n_clusters (int) – Number of clusters to find using the one of the pre-defined methods “kmeans”,”agglomerative”,”minibatchkmeans”,”spectralclustering” See sklearn.cluster for details

See also

Returns

  • If ‘return_info’ is True returns the Scikit-learn cluster object

  • used for clustering. Useful if you wish to

  • examine inertia or other outputs.

decomposition(normalize_poissonian_noise=False, algorithm='SVD', output_dimension=None, centre=None, auto_transpose=True, navigation_mask=None, signal_mask=None, var_array=None, var_func=None, reproject=None, return_info=False, print_info=True, svd_solver='auto', copy=True, **kwargs)

Apply a decomposition to a dataset with a choice of algorithms.

The results are stored in self.learning_results.

Read more in the User Guide.

Parameters
  • normalize_poissonian_noise (bool, default False) – If True, scale the signal to normalize Poissonian noise using the approach described in [Keenan2004].

  • algorithm ({"SVD", "MLPCA", "sklearn_pca", "NMF", "sparse_pca", "mini_batch_sparse_pca", "RPCA", "ORPCA", "ORNMF", custom object}, default "SVD") – The decomposition algorithm to use. If algorithm is an object, it must implement a fit_transform() method or fit() and transform() methods, in the same manner as a scikit-learn estimator.

  • output_dimension (None or int) – Number of components to keep/calculate. Default is None, i.e. min(data.shape).

  • centre ({None, "navigation", "signal"}, default None) –

    • If None, the data is not centered prior to decomposition.

    • If “navigation”, the data is centered along the navigation axis. Only used by the “SVD” algorithm.

    • If “signal”, the data is centered along the signal axis. Only used by the “SVD” algorithm.

  • auto_transpose (bool, default True) – If True, automatically transposes the data to boost performance. Only used by the “SVD” algorithm.

  • navigation_mask (boolean numpy array or BaseSignal) – The navigation locations marked as True are not used in the decomposition.

  • signal_mask (boolean numpy array or BaseSignal) – The signal locations marked as True are not used in the decomposition.

  • var_array (numpy array) – Array of variance for the maximum likelihood PCA algorithm. Only used by the “MLPCA” algorithm.

  • var_func (None or function or numpy array, default None) –

    • If None, ignored

    • If function, applies the function to the data to obtain var_array. Only used by the “MLPCA” algorithm.

    • If numpy array, creates var_array by applying a polynomial function defined by the array of coefficients to the data. Only used by the “MLPCA” algorithm.

  • reproject ({None, "signal", "navigation", "both"}, default None) – If not None, the results of the decomposition will be projected in the selected masked area.

  • return_info (bool, default False) – The result of the decomposition is stored internally. However, some algorithms generate some extra information that is not stored. If True, return any extra information if available. In the case of sklearn.decomposition objects, this includes the sklearn Estimator object.

  • print_info (bool, default True) – If True, print information about the decomposition being performed. In the case of sklearn.decomposition objects, this includes the values of all arguments of the chosen sklearn algorithm.

  • svd_solver ({"auto", "full", "arpack", "randomized"}, default "auto") –

    If auto:

    The solver is selected by a default policy based on data.shape and output_dimension: if the input data is larger than 500x500 and the number of components to extract is lower than 80% of the smallest dimension of the data, then the more efficient “randomized” method is enabled. Otherwise the exact full SVD is computed and optionally truncated afterwards.

    If full:

    run exact SVD, calling the standard LAPACK solver via scipy.linalg.svd(), and select the components by postprocessing

    If arpack:

    use truncated SVD, calling ARPACK solver via scipy.sparse.linalg.svds(). It requires strictly 0 < output_dimension < min(data.shape)

    If randomized:

    use truncated SVD, calling sklearn.utils.extmath.randomized_svd() to estimate a limited number of components

  • copy (bool, default True) –

    • If True, stores a copy of the data before any pre-treatments such as normalization in s._data_before_treatments. The original data can then be restored by calling s.undo_treatments().

    • If False, no copy is made. This can be beneficial for memory usage, but care must be taken since data will be overwritten.

  • **kwargs (extra keyword arguments) – Any keyword arguments are passed to the decomposition algorithm.

Returns

return_info

  • If True and ‘algorithm’ in [‘RPCA’, ‘ORPCA’, ‘ORNMF’], returns the low-rank (X) and sparse (E) matrices from robust PCA/NMF.

  • If True and ‘algorithm’ is an sklearn Estimator, returns the Estimator object.

  • Otherwise, returns None

Return type

tuple(numpy array, numpy array) or sklearn.Estimator or None

References

Keenan2004(1,2)

M. Keenan and P. Kotula, “Accounting for Poisson noise in the multivariate analysis of ToF-SIMS spectrum images”, Surf. Interface Anal 36(3) (2004): 203-212.

See also

estimate_elbow_position(explained_variance_ratio=None, log=True, max_points=20)

Estimate the elbow position of a scree plot curve.

Used to estimate the number of significant components in a PCA variance ratio plot or other “elbow” type curves.

Find a line between first and last point on the scree plot. With a classic elbow scree plot, this line more or less defines a triangle. The elbow should be the point which is the furthest distance from this line. For more details, see [Satopää2011].

Parameters
  • explained_variance_ratio ({None, numpy array}) – Explained variance ratio values that form the scree plot. If None, uses the explained_variance_ratio array stored in s.learning_results, so a decomposition must have been performed first.

  • max_points (int) – Maximum number of points to consider in the calculation.

Returns

elbow position – Index of the elbow position in the input array. Due to zero-based indexing, the number of significant components is elbow_position + 1.

Return type

int

References

Satopää2011

V. Satopää, J. Albrecht, D. Irwin, and B. Raghavan. “Finding a “Kneedle” in a Haystack: Detecting Knee Points in System Behavior,. 31st International Conference on Distributed Computing Systems Workshops, pp. 166-171, June 2011.

estimate_number_of_clusters(cluster_source, max_clusters=10, preprocessing=None, preprocessing_kwargs={}, number_of_components=None, navigation_mask=None, signal_mask=None, algorithm=None, metric='gap', n_ref=4, **kwargs)

Performs cluster analysis of a signal for cluster sizes ranging from n_clusters =2 to max_clusters ( default 12) Note that this can be a slow process for large datasets so please consider reducing max_clusters in this case. For each cluster it evaluates the silhouette score which is a metric of how well separated the clusters are. Maximima or peaks in the scores indicate good choices for cluster sizes.

Parameters
  • cluster_source ({"bss", "decomposition", "signal" or Signal}) – If “bss” the blind source separation results are used If “decomposition” the decomposition results are used if “signal” the signal data is used Note that using the signal can be memory intensive and is only recommended if the Signal dimension is small. Input Signal must have the same navigation dimensions as the signal instance.

  • max_clusters (int, default 10) – Max number of clusters to use. The method will scan from 2 to max_clusters.

  • preprocessing ({"standard","norm","minmax" or sklearn-like preprocessing object}) – default: ‘norm’ Preprocessing the data before cluster analysis requires preprocessing the data to be clustered to similar scales. Standard preprocessing adjusts each feature to have uniform variation. Norm preprocessing adjusts treats the set of features like a vector and each measurement is scaled to length 1. You can also pass an instance of a sklearn preprocessing module. See preprocessing methods in scikit-learn preprocessing for further details.

  • preprocessing_kwargs (dict, default empty) – Additional parameters passed to the cluster preprocessing algorithm. See sklearn.preprocessing preprocessing methods for further details

  • number_of_components (int, default None) – If you are getting the cluster centers using the decomposition results (cluster_source_for_centers=”decomposition”) you can define how many PCA components to use. If set to None the method uses the estimate of significant components found in the decomposition step using the elbow method and stored in the learning_results.number_significant_components attribute.

  • navigation_mask (boolean numpy array, default : None) – The navigation locations marked as True are not used in the clustering.

  • signal_mask (boolean numpy array, default : None) – The signal locations marked as True are not used in the clustering. Applies to “signal” or Signal cluster sources only.

  • metric ({'elbow','silhouette','gap'} default 'gap') – Use distance,silhouette analysis or gap statistics to estimate the optimal number of clusters. Gap is believed to be, overall, the best metric but it’s also the slowest. Elbow measures the distances between points in each cluster as an estimate of how well grouped they are and is the fastest metric. For elbow the optimal k is the knee or elbow point. For gap the optimal k is the first k gap(k)>= gap(k+1)-std_error For silhouette the optimal k will be one of the “maxima” found with this method

  • n_ref (int, default 4) – Number of references to use in gap statistics method Gap statistics compares the results from clustering the data to clustering uniformly distributed data. As clustering has a random variation it is typically averaged n_ref times to get an statistical average

  • **kwargs (dict {} default empty) – Parameters passed to the clustering algorithm.

  • n_clusters (int) – Number of clusters to find using the one of the pre-defined methods “kmeans”,”agglomerative”,”minibatchkmeans”,”spectralclustering” See sklearn.cluster for details

Returns

best_k – Estimate of the best cluster size

Return type

int

See also

get_bss_model(components=None, chunks='auto')

Generate model with the selected number of independent components.

Parameters

components ({None, int, list of ints}, default None) –

  • If None, rebuilds signal instance from all components

  • If int, rebuilds signal instance from components in range 0-given int

  • If list of ints, rebuilds signal instance from only components in given list

Returns

A model built from the given components.

Return type

Signal instance

get_decomposition_model(components=None)

Generate model with the selected number of principal components.

Parameters

components ({None, int, list of ints}, default None) –

  • If None, rebuilds signal instance from all components

  • If int, rebuilds signal instance from components in range 0-given int

  • If list of ints, rebuilds signal instance from only components in given list

Returns

A model built from the given components.

Return type

Signal instance

get_explained_variance_ratio()

Return explained variance ratio of the PCA components as a Signal1D.

Read more in the User Guide.

Returns

s – Explained variance ratio.

Return type

Signal1D

See also

normalize_bss_components(target='factors', function=<function sum>)

Normalize BSS components.

Parameters
  • target ({"factors", "loadings"}) – Normalize components based on the scale of either the factors or loadings.

  • function (numpy universal function, default np.sum) – Each target component is divided by the output of function(target). The function must return a scalar when operating on numpy arrays and must have an axis argument.

normalize_decomposition_components(target='factors', function=<function sum>)

Normalize decomposition components.

Parameters
  • target ({"factors", "loadings"}) – Normalize components based on the scale of either the factors or loadings.

  • function (numpy universal function, default np.sum) – Each target component is divided by the output of function(target). The function must return a scalar when operating on numpy arrays and must have an axis argument.

normalize_poissonian_noise(navigation_mask=None, signal_mask=None)

Normalize the signal under the assumption of Poisson noise.

Scales the signal using to “normalize” the Poisson data for subsequent decomposition analysis [Keenan2004].

Parameters
  • navigation_mask ({None, boolean numpy array}, default None) – Optional mask applied in the navigation axis.

  • signal_mask ({None, boolean numpy array}, default None) – Optional mask applied in the signal axis.

plot_cluster_metric()
Plot the cluster metrics calculated

using evaluate_number_of_clusters method

See also

plot_cumulative_explained_variance_ratio(n=50)

Plot cumulative explained variance up to n principal components.

Parameters

n (int) – Number of principal components to show.

Returns

ax – Axes object containing the cumulative explained variance plot.

Return type

matplotlib.axes

plot_explained_variance_ratio(n=30, log=True, threshold=0, hline='auto', vline=False, xaxis_type='index', xaxis_labeling=None, signal_fmt=None, noise_fmt=None, fig=None, ax=None, **kwargs)

Plot the decomposition explained variance ratio vs index number.

This is commonly known as a scree plot.

Read more in the User Guide.

Parameters
  • n (int or None) – Number of components to plot. If None, all components will be plot

  • log (bool, default True) – If True, the y axis uses a log scale.

  • threshold (float or int) – Threshold used to determine how many components should be highlighted as signal (as opposed to noise). If a float (between 0 and 1), threshold will be interpreted as a cutoff value, defining the variance at which to draw a line showing the cutoff between signal and noise; the number of signal components will be automatically determined by the cutoff value. If an int, threshold is interpreted as the number of components to highlight as signal (and no cutoff line will be drawn)

  • hline ({'auto', True, False}) – Whether or not to draw a horizontal line illustrating the variance cutoff for signal/noise determination. Default is to draw the line at the value given in threshold (if it is a float) and not draw in the case threshold is an int, or not given. If True, (and threshold is an int), the line will be drawn through the last component defined as signal. If False, the line will not be drawn in any circumstance.

  • vline (bool, default False) – Whether or not to draw a vertical line illustrating an estimate of the number of significant components. If True, the line will be drawn at the the knee or elbow position of the curve indicating the number of significant components. If False, the line will not be drawn in any circumstance.

  • xaxis_type ({'index', 'number'}) – Determines the type of labeling applied to the x-axis. If 'index', axis will be labeled starting at 0 (i.e. “pythonic index” labeling); if 'number', it will start at 1 (number labeling).

  • xaxis_labeling ({'ordinal', 'cardinal', None}) – Determines the format of the x-axis tick labels. If 'ordinal', “1st, 2nd, …” will be used; if 'cardinal', “1, 2, …” will be used. If None, an appropriate default will be selected.

  • signal_fmt (dict) – Dictionary of matplotlib formatting values for the signal components

  • noise_fmt (dict) – Dictionary of matplotlib formatting values for the noise components

  • fig (matplotlib figure or None) – If None, a default figure will be created, otherwise will plot into fig

  • ax (matplotlib ax (subplot) or None) – If None, a default ax will be created, otherwise will plot into ax

  • **kwargs – remaining keyword arguments are passed to matplotlib.figure()

Returns

ax – Axes object containing the scree plot

Return type

matplotlib.axes

Example

To generate a scree plot with customized symbols for signal vs. noise components and a modified cutoff threshold value:

>>> s = hs.load("some_spectrum_image")
>>> s.decomposition()
>>> s.plot_explained_variance_ratio(n=40,
>>>                                 threshold=0.005,
>>>                                 signal_fmt={'marker': 'v',
>>>                                             's': 150,
>>>                                             'c': 'pink'}
>>>                                 noise_fmt={'marker': '*',
>>>                                             's': 200,
>>>                                             'c': 'green'})

See also

reverse_bss_component(component_number)

Reverse the independent component.

Parameters

component_number (list or int) – component index/es

Examples

>>> s = hs.load('some_file')
>>> s.decomposition(True) # perform PCA
>>> s.blind_source_separation(3)  # perform ICA on 3 PCs
>>> s.reverse_bss_component(1) # reverse IC 1
>>> s.reverse_bss_component((0, 2)) # reverse ICs 0 and 2
reverse_decomposition_component(component_number)

Reverse the decomposition component.

Parameters

component_number (list or int) – component index/es

Examples

>>> s = hs.load('some_file')
>>> s.decomposition(True) # perform PCA
>>> s.reverse_decomposition_component(1) # reverse IC 1
>>> s.reverse_decomposition_component((0, 2)) # reverse ICs 0 and 2
undo_treatments()

Undo Poisson noise normalization and other pre-treatments.

Only valid if calling s.decomposition(..., copy=True).

hyperspy.learn.mva._get_derivative(signal, diff_axes, diff_order)

Calculate the derivative of a signal.

hyperspy.learn.mva._normalize_components(target, other, function=<function sum>)

Normalize components according to a function.