Source code for mindspore.train.callback._summary_collector

# Copyright 2020-2022 Huawei Technologies Co., Ltd
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
# ============================================================================
"""Summary collector callback."""
from __future__ import absolute_import

import os
import stat
import re
import json
from json.decoder import JSONDecodeError

from importlib import import_module
from collections import defaultdict
from collections.abc import Iterable

import numpy as np

from mindspore import log as logger
from mindspore import context
from mindspore.common.tensor import Tensor
from mindspore.common.parameter import Parameter
from mindspore.train.summary.summary_record import SummaryRecord, process_export_options
from mindspore.train.summary.enums import PluginEnum, ModeEnum
from mindspore.train.callback import Callback, ModelCheckpoint
from mindspore.train import lineage_pb2
from mindspore.train.serialization import save_checkpoint
from mindspore.train.callback._dataset_graph import DatasetGraph
from mindspore.nn.optim.optimizer import Optimizer
from mindspore.nn.loss.loss import LossBase
from mindspore.train._utils import check_value_type, _make_directory
from mindspore._c_expression import security

HYPER_CONFIG_ENV_NAME = "MINDINSIGHT_HYPER_CONFIG"
HYPER_CONFIG_LEN_LIMIT = 100000


class LineageMetadata:
    """Initialize parameters used in model lineage management."""
    train_dataset_path = 'train_dataset_path'
    valid_dataset_path = 'valid_dataset_path'
    train_network = 'train_network'
    loss_function = 'loss_function'
    loss = 'loss'
    optimizer = 'optimizer'
    learning_rate = 'learning_rate'
    epoch = 'epoch'
    step_num = 'step_num'
    parallel_mode = 'parallel_mode'
    device_num = 'device_num'
    batch_size = 'batch_size'
    model_path = 'model_path'
    model_ckpt = 'model_ckpt'
    model_size = 'model_size'
    metrics = 'metrics'
    train_dataset_size = 'train_dataset_size'
    valid_dataset_size = 'valid_dataset_size'


[docs]class SummaryCollector(Callback): """ SummaryCollector can help you to collect some common information. It can help you to collect loss, learning late, computational graph and so on. SummaryCollector also enables the summary operator to collect data to summary files. Note: 1. When using SummaryCollector, you need to run the code in `if __name__ == "__main__"` . 2. Multiple SummaryCollector instances in callback list are not allowed. 3. Not all information is collected at the training phase or at the eval phase. 4. SummaryCollector always record the data collected by the summary operator. 5. SummaryCollector only supports Linux systems. 6. The Summary is not supported when compile source with `-s on` option. Args: summary_dir (str): The collected data will be persisted to this directory. If the directory does not exist, it will be created automatically. collect_freq (int): Set the frequency of data collection, it should be greater than zero, and the unit is `step`. If a frequency is set, we will collect data when (current steps % freq) equals to 0, and the first step will be collected at any time. It is important to note that if the data sink mode is used, the unit will become the `epoch`. It is not recommended to collect data too frequently, which can affect performance. Default: 10. collect_specified_data (Union[None, dict]): Perform custom operations on the collected data. By default, if set to None, all data is collected as the default behavior. You can customize the collected data with a dictionary. For example, you can set {'collect_metric': False} to control not collecting metrics. The data that supports control is shown below. Default: None. - collect_metric (bool): Whether to collect training metrics, currently only the loss is collected. The first output will be treated as the loss and it will be averaged. Default: True. - collect_graph (bool): Whether to collect the computational graph. Currently, only training computational graph is collected. Default: True. - collect_train_lineage (bool): Whether to collect lineage data for the training phase, this field will be displayed on the `lineage page \ <https://www.mindspore.cn/mindinsight/docs/en/r2.0.0-alpha/lineage_and_scalars_comparison.html>`_ of MindInsight. Default: True. - collect_eval_lineage (bool): Whether to collect lineage data for the evaluation phase, this field will be displayed on the lineage page of MindInsight. Default: True. - collect_input_data (bool): Whether to collect dataset for each training. Currently only image data is supported. If there are multiple columns of data in the dataset, the first column should be image data. Default: True. - collect_dataset_graph (bool): Whether to collect dataset graph for the training phase. Default: True. - histogram_regular (Union[str, None]): Collect weight and bias for parameter distribution page and displayed in MindInsight. This field allows regular strings to control which parameters to collect. It is not recommended to collect too many parameters at once, as it can affect performance. Note that if you collect too many parameters and run out of memory, the training will fail. Default: None, it means only the first five parameters are collected. - collect_landscape (Union[dict,None]): Whether to collect the parameters needed to create the loss landscape. If set to None, collect_landscape parameters will not be collected. All parameter information is collected by default and stored in file `{summary_dir}/ckpt_dir/train_metadata.json`. - landscape_size (int): Specify the image resolution of the generated loss landscape. For example, if it is set to 128, the resolution of the landscape is 128 * 128. The calculation time increases with the increase of resolution. Default: 40. Optional values: between 3 and 256. - unit (str): Specify the interval strength of the training process. Default: "step". Optional: epoch/step. - create_landscape (dict): Select how to create loss landscape. Training process loss landscape(train) and training result loss landscape(result). Default: {"train": True, "result": True}. Optional: True/False. - num_samples (int): The size of the dataset used to create the loss landscape. For example, in image dataset, You can set num_samples is 128, which means that 128 images are used to create loss landscape. Default: 128. - intervals (List[List[int]]): Specifies the interval in which the loss landscape. For example: If the user wants to create loss landscape of two training processes, they are 1-5 epoch and 6-10 epoch respectively. They anc set [[1, 2, 3, 4, 5], [6, 7, 8, 9, 10]]. Note: Each interval have at least three epochs. keep_default_action (bool): This field affects the collection behavior of the 'collect_specified_data' field. True: it means that after specified data is set, non-specified data is collected as the default behavior. False: it means that after specified data is set, only the specified data is collected, and the others are not collected. Default: True. custom_lineage_data (Union[dict, None]): Allows you to customize the data and present it on the MingInsight lineage page. In the custom data, the type of the key supports str, and the type of value supports str, int and float. Default: None, it means there is no custom data. collect_tensor_freq (Optional[int]): The same semantics as the `collect_freq`, but controls TensorSummary only. Because TensorSummary data is too large to be compared with other summary data, this parameter is used to reduce its collection. By default, The maximum number of steps for collecting TensorSummary data is 20, but it will not exceed the number of steps for collecting other summary data. For example, given `collect_freq=10`, when the total steps is 600, TensorSummary will be collected 20 steps, while other summary data 61 steps, but when the total steps is 20, both TensorSummary and other summary will be collected 3 steps. Also note that when in parallel mode, the total steps will be split evenly, which will affect the number of steps TensorSummary will be collected. Default: None, which means to follow the behavior as described above. max_file_size (Optional[int]): The maximum size in bytes of each file that can be written to the disk. For example, to write not larger than 4GB, specify `max_file_size=4*1024**3`. Default: None, which means no limit. export_options (Union[None, dict]): Perform custom operations on the export data. Note that the size of export files is not limited by the max_file_size. You can customize the export data with a dictionary. For example, you can set {'tensor_format': 'npy'} to export tensor as npy file. The data that supports control is shown below. Default: None, it means that the data is not exported. - tensor_format (Union[str, None]): Customize the export tensor format. Supports ["npy", None]. Default: None, it means that the tensor is not exported. - npy: export tensor as npy file. Raises: ValueError: The Summary is not supported, please without `-s on` and recompile source. Examples: >>> import mindspore as ms >>> import mindspore.nn as nn >>> from mindspore.train import Model, SummaryCollector >>> from mindspore.nn import Accuracy >>> >>> if __name__ == '__main__': ... # If the device_target is GPU, set the device_target to "GPU" ... ms.set_context(mode=ms.GRAPH_MODE, device_target="Ascend") ... mnist_dataset_dir = '/path/to/mnist_dataset_directory' ... # The detail of create_dataset method shown in model_zoo.official.cv.lenet.src.dataset.py ... ds_train = create_dataset(mnist_dataset_dir, 32) ... # The detail of LeNet5 shown in model_zoo.official.cv.lenet.src.lenet.py ... network = LeNet5(10) ... net_loss = nn.SoftmaxCrossEntropyWithLogits(sparse=True, reduction="mean") ... net_opt = nn.Momentum(network.trainable_params(), 0.01, 0.9) ... model = Model(network, net_loss, net_opt, metrics={"Accuracy": Accuracy()}, amp_level="O2") ... ... # Simple usage: ... summary_collector = SummaryCollector(summary_dir='./summary_dir') ... model.train(1, ds_train, callbacks=[summary_collector], dataset_sink_mode=False) ... ... # Do not collect metric and collect the first layer parameter, others are collected by default ... specified={'collect_metric': False, 'histogram_regular': '^conv1.*'} ... summary_collector = SummaryCollector(summary_dir='./summary_dir', collect_specified_data=specified) ... model.train(1, ds_train, callbacks=[summary_collector], dataset_sink_mode=False) """ _DEFAULT_SPECIFIED_DATA = { 'collect_metric': True, 'collect_graph': True, 'collect_train_lineage': True, 'collect_eval_lineage': True, 'collect_input_data': True, 'collect_dataset_graph': True, 'histogram_regular': None, 'collect_landscape': {'landscape_size': 40, 'unit': 'step', 'num_samples': 2048, 'create_landscape': {'train': True, 'result': True}, 'intervals': [[1, 2, 3, 4, 5, 6, 7, 8, 9], [10, 11, 12, 13, 14, 15, 16, 17, 18, 19]]} } def __init__(self, summary_dir, collect_freq=10, collect_specified_data=None, keep_default_action=True, custom_lineage_data=None, collect_tensor_freq=None, max_file_size=None, export_options=None): if security.enable_security(): raise ValueError('The Summary is not supported, please without `-s on` and recompile source.') super(SummaryCollector, self).__init__() self._summary_dir = _make_directory(summary_dir, "summary_dir") self._record = None self._check_positive('collect_freq', collect_freq) self._collect_freq = collect_freq self._check_positive('collect_tensor_freq', collect_tensor_freq, allow_none=True) self._collect_tensor_freq = collect_tensor_freq self._tensor_collect_range = None self._check_positive('max_file_size', max_file_size, allow_none=True) self._max_file_size = max_file_size self._export_options = process_export_options(export_options) self._check_action(keep_default_action) self._collect_specified_data = self._process_specified_data(collect_specified_data, keep_default_action) msg = f"For 'collect_specified_data' the value after processing is: {self._collect_specified_data}." logger.info(msg) landscape = self._collect_specified_data.get('collect_landscape', None) check_value_type('collect_landscape', landscape, [dict, type(None)]) self._is_collect_landscape = bool(landscape) if self._is_collect_landscape: self._check_collect_landscape_data(landscape) self._ckpt_dir = os.path.join(self._summary_dir, 'ckpt_dir') _make_directory(self._ckpt_dir) self._model_params_file_map = {} self._epoch_group = defaultdict(list) intervals = landscape.get('intervals') self._create_epoch_group(intervals) self._custom_lineage_data = self._process_custom_lineage_data(custom_lineage_data) self._temp_optimizer = None self._has_saved_graph = False self._has_saved_custom_data = False self._is_parse_loss_success = True self._first_step = True self._dataset_sink_mode = True def __enter__(self): self._record = SummaryRecord(log_dir=self._summary_dir, max_file_size=self._max_file_size, raise_exception=False, export_options=self._export_options) self._first_step, self._dataset_sink_mode = True, True return self def __exit__(self, *err): self._record.close() def _check_positive(self, name, value, allow_none=False): """Check if the value to be int type and positive.""" if allow_none and value is None: return check_value_type(name, value, int) if value <= 0: raise ValueError(f'For "{self.__class__.__name__}", the value of `{name}` should be greater than 0, ' f'but got `{value}`.') def _create_epoch_group(self, intervals): """Create epoch group.""" for i, interval in enumerate(intervals): for j in interval: self._epoch_group[i].append(j) def _process_custom_lineage_data(self, custom_lineage_data): """ Check user custom lineage data. Args: custom_lineage_data (dict): The user custom defined data. Raises: TypeError: If the type of parameters is invalid. """ if custom_lineage_data is None: custom_lineage_data = {} self._check_custom_lineage_type('custom_lineage_data', custom_lineage_data) auto_custom_lineage_data = self._collect_optimizer_custom_lineage_data() self._check_custom_lineage_type('auto_custom_lineage_data', auto_custom_lineage_data) # the priority of user defined info is higher than auto collected info auto_custom_lineage_data.update(custom_lineage_data) custom_lineage_data = auto_custom_lineage_data return custom_lineage_data @staticmethod def _check_custom_lineage_type(param_name, custom_lineage): """Check custom lineage type.""" check_value_type(param_name, custom_lineage, [dict, type(None)]) for key, value in custom_lineage.items(): check_value_type(f'{param_name} -> {key}', key, str) check_value_type(f'the value of {param_name} -> {key}', value, (int, str, float)) @staticmethod def _collect_optimizer_custom_lineage_data(): """Collect custom lineage data if mindoptimizer has set the hyper config.""" auto_custom_lineage_data = {} hyper_config = os.environ.get(HYPER_CONFIG_ENV_NAME) if hyper_config is None: logger.debug("Hyper config is not in system environment.") return auto_custom_lineage_data if len(hyper_config) > HYPER_CONFIG_LEN_LIMIT: logger.warning("The 'MINDINSIGHT_HYPER_CONFIG' of environment variable is too long. The length limit " "is %s, the length of hyper_config is %s." % (HYPER_CONFIG_LEN_LIMIT, len(hyper_config))) return auto_custom_lineage_data try: hyper_config = json.loads(hyper_config) except (TypeError, JSONDecodeError) as exc: logger.warning("The 'MINDINSIGHT_HYPER_CONFIG' of environment variable decode error. " "Detail: %s." % str(exc)) return auto_custom_lineage_data custom_lineage_data = hyper_config.get("custom_lineage_data") if custom_lineage_data is None: logger.info("No custom lineage data in hyper config. Please check the custom lineage data " "if custom parameters exist in the configuration file.") auto_custom_lineage_data = custom_lineage_data if custom_lineage_data is not None else {} return auto_custom_lineage_data @staticmethod def _check_action(action): """Check action type.""" check_value_type('keep_default_action', action, bool) def _check_landscape_size(self, landscape_size): """Check landscape size type and value.""" check_value_type('landscape_size', landscape_size, int) # landscape size should be between 3 and 256. if landscape_size < 3 or landscape_size > 256: raise ValueError(f'For "{self.__class__.__name__}", the "landscape_size" in collect_specified_data ' f'should be less than 256 and more than 3, but got the: {landscape_size}') def _check_unit(self, unit): """Check unit type and value.""" check_value_type('unit', unit, str) if unit not in ["step", "epoch"]: raise ValueError(f'For "{self.__class__.__name__}", unit in collect_specified_data should be step ' f'or epoch, but got the: {unit}.') def _check_create_landscape(self, create_landscape): """Check create landscape type and value.""" check_value_type('create_landscape', create_landscape, dict) for param, value in create_landscape.items(): if param not in ["train", "result"]: raise ValueError(f'For "{self.__class__.__name__}", the key to create landscape should be in ' f'["train", "result"], but got the: {param}.') check_value_type(param, value, bool) def _check_intervals(self, intervals): """Check intervals type and value.""" check_value_type('intervals', intervals, list) for _, interval in enumerate(intervals): check_value_type('each interval inintervals', interval, list) if len(interval) < 3: raise ValueError(f'For "{self.__class__.__name__}", the length of each list in "intervals" should not ' f'be less than three, but got the: {interval}') for j in interval: if not isinstance(j, int): raise TypeError(f'For "{self.__class__.__name__}", the value of each list in "intervals" should ' f'be int, but got the: {type(j)}') def _check_collect_landscape_data(self, collect_landscape): """Check collect landscape data type and value.""" unexpected_params = set(collect_landscape) - set(self._DEFAULT_SPECIFIED_DATA.get("collect_landscape")) if unexpected_params: raise ValueError(f'For "{self.__class__.__name__}", the keys {unexpected_params} of `collect_landscape` ' f'are unsupported, expect the follow keys: ' f'{list(self._DEFAULT_SPECIFIED_DATA.get("collect_landscape").keys())}') landscape_size = collect_landscape.get("landscape_size", 40) self._check_landscape_size(landscape_size) unit = collect_landscape.get("unit", "step") self._check_unit(unit) num_samples = collect_landscape.get("num_samples", 2048) check_value_type("num_samples", num_samples, int) create_landscape = collect_landscape.get("create_landscape", {"train": True, "result": True}) self._check_create_landscape(create_landscape) intervals = collect_landscape.get("intervals") self._check_intervals(intervals) def _process_specified_data(self, specified_data, action): """Check specified data type and value.""" check_value_type('collect_specified_data', specified_data, [dict, type(None)]) if specified_data is None: if action: return dict(self._DEFAULT_SPECIFIED_DATA) return dict() for param_name in specified_data: check_value_type(param_name, param_name, [str]) unexpected_params = set(specified_data) - set(self._DEFAULT_SPECIFIED_DATA) if unexpected_params: raise ValueError(f'For "{self.__class__.__name__}", the keys {unexpected_params} of ' f'`collect_specified_data` are unsupported, ' f'expect the follow keys: {list(self._DEFAULT_SPECIFIED_DATA.keys())}') if 'histogram_regular' in specified_data: regular = specified_data.get('histogram_regular') check_value_type('histogram_regular', regular, (str, type(None))) if isinstance(regular, str): try: re.match(regular, '') except re.error as exc: raise ValueError(f'For `collect_specified_data`, the value of `histogram_regular` ' f'is not a valid regular expression. Detail: {str(exc)}.') from exc bool_items = set(self._DEFAULT_SPECIFIED_DATA) - {'histogram_regular', 'collect_landscape'} for item in bool_items: if item in specified_data: check_value_type(item, specified_data.get(item), bool) if action: result = dict(self._DEFAULT_SPECIFIED_DATA) result.update(specified_data) else: result = specified_data return result def begin(self, run_context): cb_params = run_context.original_args() self._check_callbacks(cb_params) if cb_params.mode not in ModeEnum.to_list(): raise ValueError('Only support `train` (model.train) and `eval` (model.eval) mode, ' 'but got `{cb_params.mode}` mode.') self._record.set_mode(cb_params.mode) self._dataset_sink_mode = cb_params.dataset_sink_mode def step_end(self, run_context): cb_params = run_context.original_args() if cb_params.mode != ModeEnum.TRAIN.value: return if not self._has_saved_graph: self._collect_graphs(cb_params) self._collect_dataset_graph(cb_params) self._has_saved_graph = True self._record.record(cb_params.cur_step_num) if self._custom_lineage_data and not self._has_saved_custom_data: packaged_custom_data = self._package_custom_lineage_data(self._custom_lineage_data) self._record.add_value('custom_lineage_data', 'custom_lineage_data', packaged_custom_data) self._has_saved_custom_data = True self._record.record(cb_params.cur_step_num) if self._first_step: self._tensor_collect_range = self._get_tensor_collect_range(cb_params, self._dataset_sink_mode) self._collect_at_step_end(cb_params, plugin_filter=None) self._first_step = False self._record.flush() else: current = cb_params.cur_epoch_num if self._dataset_sink_mode else cb_params.cur_step_num if current % self._collect_freq == 0 and current in self._tensor_collect_range: self._collect_at_step_end(cb_params, plugin_filter=None) elif current in self._tensor_collect_range: self._collect_at_step_end(cb_params, lambda plugin: plugin == PluginEnum.TENSOR.value) elif current % self._collect_freq == 0: self._collect_at_step_end(cb_params, lambda plugin: plugin != PluginEnum.TENSOR.value) collect_landscape = self._collect_specified_data.get('collect_landscape') if collect_landscape is not None: intervals = collect_landscape.get('intervals') collect_interval = False for interval in intervals: if "cur_step_num" in cb_params: if cb_params.cur_step_num in interval: collect_interval = True break if collect_landscape and collect_landscape.get('unit', 'step') == 'step' and collect_interval: self._save_model_params_for_landscape(cb_params) def _get_tensor_collect_range(self, cb_params, dataset_sink_mode): """Get tensor collect range.""" total_step = cb_params.epoch_num if not dataset_sink_mode: total_step *= cb_params.batch_num if self._collect_tensor_freq is not None: # `total_step + 1`: `total_step` would be a value of `cb_params.cur_step_num`. return range(0, total_step + 1, self._collect_tensor_freq) summary_to_collect = len(range(0, total_step + 1, self._collect_freq)) default_tensor_summary_limit = 20 if summary_to_collect > default_tensor_summary_limit: tensor_freq = total_step // (default_tensor_summary_limit - 1) if tensor_freq > 1: return range(0, total_step + 1, tensor_freq)[:default_tensor_summary_limit] # `cb_params.cur_step_num` counting from `1`, when `1` is in the range, take `1` more steps. return range(0, total_step + 1)[:default_tensor_summary_limit + 1] return range(0, total_step + 1, self._collect_freq) def _collect_at_step_end(self, cb_params, plugin_filter): self._collect_input_data(cb_params) self._collect_metric(cb_params) self._collect_histogram(cb_params) self._record.record(cb_params.cur_step_num, plugin_filter=plugin_filter) def epoch_end(self, run_context): cb_params = run_context.original_args() collect_landscape = self._collect_specified_data.get('collect_landscape') if collect_landscape is not None: intervals = collect_landscape.get('intervals') collect_interval = False for interval in intervals: if "cur_epoch_num" in cb_params and cb_params.cur_epoch_num in interval: collect_interval = True break if collect_landscape and collect_landscape.get('unit', 'step') == 'epoch' and collect_interval: self._save_model_params_for_landscape(cb_params) self._record.flush() def end(self, run_context): cb_params = run_context.original_args() if cb_params.mode == ModeEnum.TRAIN.value: self._collect_train_lineage(cb_params) else: self._collect_eval_lineage(cb_params) # This is a workaround to avoid record '_summary_tensor_cache'. self._record.set_mode('eval') # There's nothing special about setting step to 0 here, just to satisfy the interface call self._record.record(step=0) collect_landscape = self._collect_specified_data.get('collect_landscape') if cb_params.mode == ModeEnum.TRAIN.value and collect_landscape: unit = collect_landscape.get('unit', 'step') num_samples = collect_landscape.get('num_samples', 2048) landscape_size = collect_landscape.get('landscape_size', 40) create_landscape = collect_landscape.get('create_landscape', {'train': True, 'result': True}) self._save_metadata(cb_params.batch_num, unit, num_samples, landscape_size, create_landscape) def _save_metadata(self, step_per_epoch, unit, num_samples, landscape_size, create_landscape): """Save meta data to json file.""" data = { "epoch_group": self._epoch_group, "model_params_file_map": self._model_params_file_map, "step_per_epoch": step_per_epoch, "unit": unit, "num_samples": num_samples, "landscape_size": landscape_size, "create_landscape": create_landscape } meta_path = os.path.join(self._ckpt_dir, 'train_metadata.json') try: with open(meta_path, 'w') as file: json.dump(data, file) os.chmod(meta_path, stat.S_IRUSR) except OSError as e: logger.error("Write meta data %s failed, detail: %s" % (meta_path, str(e))) def _save_model_params(self, cur_num, unit, backbone): """Save model params.""" param_list = [] for param in backbone.get_parameters(): param.init_data() param_data = param.data if isinstance(param.data, Tensor) else Tensor(param.data) param_list.append(dict( name=param.name, data=param_data )) ckpt_file_name = f"{type(backbone).__name__}_{cur_num}_{unit}.ckpt" file_path = os.path.join(self._ckpt_dir, ckpt_file_name) try: save_checkpoint(param_list, file_path) except OSError as e: logger.error(str(e)) self._model_params_file_map[str(cur_num)] = file_path def _save_model_params_for_landscape(self, cb_params): """Save model params for landscape.""" if cb_params.mode == ModeEnum.TRAIN.value: backbone = self._get_backbone(cb_params.train_network) while True: if 'optimizer' in backbone.name_cells() and 'network' in backbone.name_cells(): backbone = backbone.network break if 'optimizer' not in backbone.name_cells() and 'network' in backbone.name_cells(): backbone = backbone.network else: break collect_landscape = self._collect_specified_data.get('collect_landscape') unit = collect_landscape.get('unit', 'step') cur_num = cb_params.cur_epoch_num if collect_landscape and unit == 'epoch' else cb_params.cur_step_num logger.info("Save model params, %s: %s." % (unit, cur_num)) self._save_model_params(cur_num, unit, backbone) def _check_callbacks(self, cb_params): """Check there if there are duplicate instances of SummaryCollector.""" callbacks = cb_params.list_callback is_find = False for callback in callbacks: if type(callback).__name__ == self.__class__.__name__: if not is_find: is_find = True continue raise ValueError(f"There are more than one {self.__class__.__name__} instance in callback list," f"but expected only one {self.__class__.__name__} instance.") @staticmethod def _package_custom_lineage_data(custom_lineage_data): """ Package user-defined lineage data into binary data. Args: custom_lineage_data (dict): User custom lineage data. Returns: UserDefinedInfo, a object of lineage_pb2.UserDefinedInfo. """ user_defined_info = lineage_pb2.UserDefinedInfo() for key, value in custom_lineage_data.items(): if isinstance(value, int): attr_name = "map_int32" elif isinstance(value, float): attr_name = "map_double" else: attr_name = "map_str" user_info = user_defined_info.user_info.add() getattr(user_info, attr_name)[key] = value return user_defined_info def _collect_input_data(self, cb_params): """Only support to collect image data.""" if not self._is_allowed_to_collect_input_data(cb_params): return input_data = getattr(cb_params, 'train_dataset_element', None) if isinstance(input_data, (list, tuple)) and input_data: input_data = input_data[0] try: self._record.add_value(PluginEnum.IMAGE.value, 'input_data/auto', input_data) except (TypeError, ValueError): logger.warning('The input data of network are not image, so will not collect by SummaryCollector.') self._collect_specified_data['collect_input_data'] = False return def _is_allowed_to_collect_input_data(self, cb_params): """Check if the input data is allowed to be collected.""" if not self._collect_specified_data.get('collect_input_data'): return False if self._dataset_sink_mode and (context.get_context('device_target') in ('Ascend', 'GPU')): logger.warning("On Ascend or GPU device, SummaryCollector is not supported to " "record input data in dataset sink mode.") self._collect_specified_data['collect_input_data'] = False return False input_data = getattr(cb_params, 'train_dataset_element', None) if not isinstance(input_data, (Tensor, list, tuple)): self._collect_specified_data['collect_input_data'] = False logger.warning("The type of input data is not Tensor/list/tuple, " "so SummaryCollector will not collect input data.") return False if not isinstance(input_data, Tensor) and not input_data: self._collect_specified_data['collect_input_data'] = False logger.warning("The 'train_dataset_element' in RunContext.original_args is empty, " "so SummaryCollector will not record the input data. ") return False return True def _collect_dataset_graph(self, cb_params): """Only collect train dataset graph.""" if not self._collect_specified_data.get('collect_dataset_graph'): return # After analysis, we think that the validated dataset graph and the training dataset graph # should be consistent under normal scenarios, so only the training dataset graph is collected. if cb_params.mode == ModeEnum.TRAIN.value: train_dataset = cb_params.train_dataset dataset_graph = DatasetGraph() graph_bytes = dataset_graph.package_dataset_graph(train_dataset) if graph_bytes is None: return self._record.add_value('dataset_graph', 'train_dataset', graph_bytes) def _collect_graphs(self, cb_params): """Collect the graph of train network and eval network.""" if not self._collect_specified_data.get('collect_graph'): return network = cb_params.train_network if cb_params.mode == ModeEnum.TRAIN.value else cb_params.eval_network graph_proto = network.get_func_graph_proto() if graph_proto is None: logger.warning("Can not get graph proto, it may not be 'GRAPH_MODE' in context currently, " "so SummaryCollector will not collect graph.") return self._record.add_value(PluginEnum.GRAPH.value, 'train_network/auto', graph_proto) def _collect_metric(self, cb_params): """Collect metric, currently only collection Loss is supported.""" if not self._collect_specified_data.get('collect_metric'): return loss = self._get_loss(cb_params) if loss is None: return try: self._record.add_value(PluginEnum.SCALAR.value, 'loss/auto', loss) except ValueError: logger.warning("The output of network is not a scalar, so SummaryCollector will not collect loss.") self._collect_specified_data['collect_metric'] = False def _get_loss(self, cb_params): """ Get loss from the network output. Args: cb_params (_InternalCallbackParam): Callback parameters. Returns: Union[Tensor, None], if parse loss success, will return a Tensor value(shape is [1]), else return None. """ if not self._is_parse_loss_success: # If parsing has failed before, avoid repeating it return None output = cb_params.net_outputs if output is None: logger.warning("Can not find any output by this network, so SummaryCollector will not collect loss.") self._is_parse_loss_success = False return None if isinstance(output, (int, float, Tensor)): loss = output elif isinstance(output, (list, tuple)) and output: # If the output is a list, since the default network returns loss first, # we assume that the first one is loss. loss = output[0] else: logger.warning("The output type could not be identified, expect type is one of " "[int, float, Tensor, list, tuple], so no loss was recorded in SummaryCollector.") self._is_parse_loss_success = False return None if not isinstance(loss, Tensor): loss = Tensor(loss) loss = Tensor(np.mean(loss.asnumpy())) return loss def _get_optimizer(self, cb_params): """ Get optimizer from the cb_params or parse from the network. Args: cb_params (_InternalCallbackParam): Callback parameters. Returns: Union[Optimizer, None], if parse optimizer success, will return a optimizer, else return None. """ # 'optimizer_failed' means find optimizer failed, so we will not collect data about optimizer. optimizer_failed = 'Failed' if self._temp_optimizer == optimizer_failed: return None if self._temp_optimizer is not None: return self._temp_optimizer optimizer = cb_params.optimizer if optimizer is None: network = cb_params.train_network if cb_params.mode == 'train' else cb_params.eval_network optimizer = self._parse_optimizer_by_network(network) if optimizer is None or not isinstance(optimizer, Optimizer): logger.warning("Can not find optimizer in network, or the optimizer does not inherit MindSpore's " "optimizer, so we will not collect data about optimizer in SummaryCollector.") optimizer = None self._temp_optimizer = optimizer if optimizer is not None else optimizer_failed return optimizer @staticmethod def _parse_optimizer_by_network(network): """Parse optimizer from network, if parse success will return a optimizer, else return None.""" optimizer = None for _, cell in network.cells_and_names(): if isinstance(cell, Optimizer): return cell try: optimizer = getattr(cell, 'optimizer') except AttributeError: continue if not isinstance(optimizer, Optimizer): continue # Optimizer found successfully break return optimizer def _collect_histogram(self, cb_params): """Collect histogram data, contain the parameter weight and bias.""" # Note: if there is not a key named `histogram_regular` in `self._collect_specified_data`, # it means we will not collect histogram data. if 'histogram_regular' not in self._collect_specified_data: return optimizer = self._get_optimizer(cb_params) if optimizer is None: return parameters = optimizer.parameters regular = self._collect_specified_data.get('histogram_regular') if regular is not None: for parameter in parameters: if re.match(regular, parameter.name): self._record.add_value(PluginEnum.HISTOGRAM.value, parameter.name+'/auto', parameter.data) return # Note: If `histogram_regular` in `self._collect_specified_data` and the value is None, # we will collect the first five parameters. default_parameter_count = 5 for parameter in parameters[:default_parameter_count]: self._record.add_value(PluginEnum.HISTOGRAM.value, parameter.name+'/auto', parameter.data) @staticmethod def _get_learning_rate(optimizer): """ Parse the learning rate from optimizer. Args: optimizer (Optimizer): A optimizer which inherit the MindSpore Optimizer class. Returns: Union[Tensor, None], if parse learning rate success, will return a Tensor, else return None. """ learning_rate = optimizer.learning_rate if not isinstance(learning_rate, Parameter): logger.warning("The learning rate detected in the optimizer is not a Parameter type, " "so it is not recorded. Its type is %r.", type(learning_rate).__name__) return None return learning_rate.data def _collect_train_lineage(self, cb_params): """Collect train lineage data, the detail refer to lineage_pb2.TrainLineage.""" if not self._collect_specified_data.get('collect_train_lineage'): return train_lineage = {} loss = self._get_loss(cb_params) if loss is not None: loss_numpy = loss.asnumpy() loss = float(np.atleast_1d(loss_numpy)[0]) train_lineage[LineageMetadata.loss] = loss else: train_lineage[LineageMetadata.loss] = None optimizer = self._get_optimizer(cb_params) learning_rate = self._get_learning_rate(optimizer) if optimizer is not None else None if learning_rate is not None: train_lineage[LineageMetadata.learning_rate] = list(np.atleast_1d(learning_rate.asnumpy()))[0] else: train_lineage[LineageMetadata.learning_rate] = None train_lineage[LineageMetadata.optimizer] = type(optimizer).__name__ if optimizer else None train_lineage[LineageMetadata.train_network] = type(cb_params.network).__name__ loss_fn = self._get_loss_fn(cb_params) train_lineage[LineageMetadata.loss_function] = type(loss_fn).__name__ if loss_fn else None train_lineage[LineageMetadata.epoch] = cb_params.epoch_num train_lineage[LineageMetadata.step_num] = cb_params.cur_step_num train_lineage[LineageMetadata.parallel_mode] = cb_params.parallel_mode train_lineage[LineageMetadata.device_num] = cb_params.device_number ckpt_file_path = self._get_ckpt_file_path(cb_params) train_lineage[LineageMetadata.model_path] = json.dumps(dict(ckpt=ckpt_file_path)) model_size = os.path.getsize(ckpt_file_path) if ckpt_file_path else 0 train_lineage[LineageMetadata.model_size] = model_size self._parse_dataset(cb_params, train_lineage) train_lineage_message = self._package_train_lineage_message(train_lineage) self._record.add_value(PluginEnum.TRAIN_LINEAGE.value, 'train_lineage', train_lineage_message) @staticmethod def _package_train_lineage_message(train_lineage): """ Package train lineage data into binary data. Args: train_lineage (dict): The train lineage dict, refer to the attribute of `_collect_train_lineage` method. Returns: TrainLineage, a object of lineage_pb2.TrainLineage. """ lineage_message = lineage_pb2.TrainLineage() if train_lineage.get(LineageMetadata.train_network) is not None: lineage_message.algorithm.network = train_lineage.get(LineageMetadata.train_network) if train_lineage.get(LineageMetadata.loss) is not None: lineage_message.algorithm.loss = train_lineage.get(LineageMetadata.loss) # Construct train_dataset message. if train_lineage.get(LineageMetadata.train_dataset_path) is not None: lineage_message.train_dataset.train_dataset_path = train_lineage.get(LineageMetadata.train_dataset_path) if train_lineage.get(LineageMetadata.train_dataset_size) is not None: lineage_message.train_dataset.train_dataset_size = train_lineage.get(LineageMetadata.train_dataset_size) # Construct model message lineage_message.model.path = train_lineage.get(LineageMetadata.model_path) lineage_message.model.size = train_lineage.get(LineageMetadata.model_size) # Construct hyper_parameters message. if train_lineage.get(LineageMetadata.learning_rate) is not None: lineage_message.hyper_parameters.learning_rate = train_lineage.get(LineageMetadata.learning_rate) if train_lineage.get(LineageMetadata.optimizer) is not None: lineage_message.hyper_parameters.optimizer = train_lineage.get(LineageMetadata.optimizer) if train_lineage.get(LineageMetadata.loss_function) is not None: lineage_message.hyper_parameters.loss_function = train_lineage.get(LineageMetadata.loss_function) if train_lineage.get(LineageMetadata.parallel_mode) is not None: lineage_message.hyper_parameters.parallel_mode = train_lineage.get(LineageMetadata.parallel_mode) lineage_message.hyper_parameters.epoch = train_lineage.get(LineageMetadata.epoch) lineage_message.hyper_parameters.device_num = train_lineage.get(LineageMetadata.device_num) lineage_message.hyper_parameters.batch_size = train_lineage.get(LineageMetadata.batch_size) return lineage_message def _parse_dataset(self, cb_params, lineage_dict): """ Analyze Dataset to get the dataset path and dataset size. Args: cb_params (_InternalCallbackParam): Callback parameters. lineage_dict (dict): The lineage dict, refer to the attribute of `_collect_train_lineage` method or `_collect_eval_lineage`. Returns: dict, the lineage metadata. """ dataset = cb_params.train_dataset if cb_params.mode == ModeEnum.TRAIN.value else cb_params.valid_dataset try: dataset_path = self._get_dataset_path(dataset) except IndexError: dataset_path = None if dataset_path and os.path.isfile(dataset_path): dataset_dir = os.path.dirname(dataset_path) else: dataset_dir = dataset_path batch_num = dataset.get_dataset_size() batch_size = dataset.get_batch_size() dataset_size = int(batch_num * batch_size) lineage_dict[LineageMetadata.batch_size] = batch_size if cb_params.mode == ModeEnum.TRAIN.value: lineage_dict[LineageMetadata.train_dataset_path] = dataset_dir lineage_dict[LineageMetadata.train_dataset_size] = dataset_size else: lineage_dict[LineageMetadata.valid_dataset_path] = dataset_dir lineage_dict[LineageMetadata.valid_dataset_size] = dataset_size return lineage_dict def _get_dataset_path(self, output_dataset): """ Get dataset path of MindDataset object. Args: output_dataset (Union[Dataset, ImageFolderDataset, MnistDataset, Cifar10Dataset, Cifar100Dataset, VOCDataset, CelebADataset, MindDataset, ManifestDataset, TFRecordDataset, TextFileDataset]): Refer to mindspore.dataset.Dataset. Returns: str, dataset path. Raises: IndexError: it means get dataset path failed. """ dataset_package = import_module('mindspore.dataset') dataset_dir_set = (dataset_package.ImageFolderDataset, dataset_package.MnistDataset, dataset_package.Cifar10Dataset, dataset_package.Cifar100Dataset, dataset_package.VOCDataset, dataset_package.CelebADataset) dataset_files_set = (dataset_package.TFRecordDataset, dataset_package.TextFileDataset) dataset_path = '' if isinstance(output_dataset, dataset_package.ManifestDataset): dataset_path = output_dataset.dataset_file if isinstance(output_dataset, dataset_package.MindDataset): dataset_path = output_dataset.dataset_files if isinstance(output_dataset, dataset_dir_set): dataset_path = output_dataset.dataset_dir if isinstance(output_dataset, dataset_files_set): dataset_path = output_dataset.dataset_files[0] if dataset_path: if isinstance(dataset_path, str): return dataset_path if isinstance(dataset_path, Iterable): return list(dataset_path)[0] return self._get_dataset_path(output_dataset.children[0]) @staticmethod def _get_ckpt_file_path(cb_params): """ Get checkpoint file path from MindSpore callback list. Args: cb_params (_InternalCallbackParam): Callback parameters. Returns: Union[str, None], if parse success will checkpoint file absolute path, else return None. """ callbacks = cb_params.list_callback ckpt_file_path = None for callback in callbacks: if isinstance(callback, ModelCheckpoint): ckpt_file_path = callback.latest_ckpt_file_name if ckpt_file_path: ckpt_file_path = os.path.realpath(ckpt_file_path) return ckpt_file_path @classmethod def _get_backbone(cls, network): """ Get the backbone network. Args: network (Cell): The train network. Returns: Union[Cell, None], backbone network, if parse failed, will return None. """ backbone = None backbone_key = '_backbone' for _, cell in network.cells_and_names(): if hasattr(cell, backbone_key): backbone = getattr(cell, backbone_key) break backbone = network if backbone is None else backbone return backbone @staticmethod def _get_loss_fn(cb_params): """ Get loss function by cb_params and analyzing network. Args: cb_params (_InternalCallbackParam): Callback parameters. Returns: Union[Cell, None], a Cell object, if parse failed, will return None. """ loss_fn = cb_params.loss_fn if loss_fn is not None: return loss_fn if cb_params.mode == ModeEnum.TRAIN.value: network = cb_params.train_network else: network = cb_params.eval_network for _, cell in network.cells_and_names(): if isinstance(cell, LossBase): loss_fn = cell break return loss_fn def _collect_eval_lineage(self, cb_params): """Collect eval lineage data, the detail refer to lineage_pb2.EvaluationLineage.""" if not self._collect_specified_data.get('collect_eval_lineage'): return eval_lineage = dict() try: eval_lineage[LineageMetadata.metrics] = json.dumps(cb_params.metrics) except TypeError as exc: logger.warning("Summary cannot collect the type of metrics, currently support type: dict, list, tuple, " "str, int, float, bool and None. %s.", str(exc)) self._parse_dataset(cb_params, eval_lineage) eval_lineage_message = self._package_eval_lineage_message(eval_lineage) self._record.add_value(PluginEnum.EVAL_LINEAGE.value, 'eval_lineage', eval_lineage_message) @staticmethod def _package_eval_lineage_message(eval_lineage): """ Package eval lineage data into binary data. Args: eval_lineage (dict): The eval lineage dict, refer to the attribute of `_collect_eval_lineage` method. Returns: EvaluationLineage, a object of lineage_pb2.EvaluationLineage. """ lineage_message = lineage_pb2.EvaluationLineage() if eval_lineage.get(LineageMetadata.metrics) is not None: lineage_message.metric = eval_lineage.get(LineageMetadata.metrics) if eval_lineage.get(LineageMetadata.valid_dataset_path) is not None: lineage_message.valid_dataset.valid_dataset_path = eval_lineage.get(LineageMetadata.valid_dataset_path) if eval_lineage.get(LineageMetadata.valid_dataset_size) is not None: lineage_message.valid_dataset.valid_dataset_size = eval_lineage.get(LineageMetadata.valid_dataset_size) return lineage_message