Network Construction and Training Error Analysis

View Source On Gitee  

The following lists the common network construction and training errors in static graph mode.

Incorrect Context Configuration

When performing network training, you must run the following command to specify the backend device: set_context(device_target=device). MindSpore supports CPU, GPU, and Ascend. If a GPU backend device is incorrectly specified as Ascend by running set_context(device_target="Ascend"), the following error message will be displayed:

ValueError: For 'set_context', package type mindspore-gpu support 'device_target' type gpu or cpu, but got Ascend.

The running backend specified by the script must match the actual hardware device.

For details, visit the following website:

MindSpore Configuration Error - ‘set_context’ Configuration Error

For details about the context configuration, see ‘set_context’.

Syntax Errors

Incorrect Construct Parameter

In MindSpore, the basic unit of the neural network is nn.Cell. All the models or neural network layers must inherit this base class. As a member function of this base class, construct defines the calculation logic to be executed and must be rewritten in all inherited classes. The function prototype of construct is:

def construct(self, *inputs, **kwargs):

When the function is rewritten, the following error message may be displayed:

TypeError: The function construct needs 0 positional argument and 0 default argument, but provided 1

This is because the function parameter list is incorrect when the user-defined construct function is implemented, for example, "def construct(*inputs, **kwargs):", where self is missing. In this case, an error is reported when MindSpore parses the syntax.

For details, visit the following website:

MindSpore Syntax Error - ‘construct’ Definition Error

Incorrect Control Flow Syntax

In static graph mode, Python code is not executed by the Python interpreter. Instead, the code is built into a static computational graph for execution. The control flow syntax supported by MindSpore includes if, for, and while statements. The attributes of the objects returned by different branches of the if statement may be inconsistent. As a result, an error is reported. The error message is displayed as follows:

TypeError: Cannot join the return values of different branches, perhaps you need to make them equal.
Type Join Failed: dtype1 = Float32, dtype2 = Float16.

According to the error message, the return values of different branches of the if statement are of different types. One is float32, and the other is float16. As a result, a build error is reported.

ValueError: Cannot join the return values of different branches, perhaps you need to make them equal.
Shape Join Failed: shape1 = (2, 3, 4, 5), shape2 = ().

According to the error message, the dimension shapes of the return values of different branches of the if statement are different. One is a 4-bit tensor of 2*3*4*5, and the other is a scalar. As a result, a build error is reported.

For details, visit the following website:

MindSpore Syntax Error - Type (Shape) Join Failed

The number of loops of the for and while statements may exceed the permitted range. As a result, the function call stack exceeds the threshold. The error message is displayed as follows:

RuntimeError: Exceed function call depth limit 1000, (function call depth: 1001, simulate call depth: 997).

One solution to the problem that the function call stack exceeds the threshold is to simplify the network structure and reduce the number of loops. Another method is to use set_context(max_call_depth=value) to increase the threshold of the function call stack.

For details, visit the following website:

MindSpore Syntax Error - Exceed function call depth limit

Operator Build Errors

Operator build errors are mainly caused by input parameters that do not meet requirements or operator functions that are not supported.

For example, when the ReduceSum operator is used, the following error message is displayed if the input data exceeds eight dimensions:

RuntimeError: ({'errCode': 'E80012', 'op_name': 'reduce_sum_d', 'param_name': 'x', 'min_value': 0, 'max_value': 8, 'real_value': 10}, 'In op, the num of dimensions of input/output[x] should be in the range of [0, 8], but actually is [10].')

For details, visit the following website:

MindSpore Operator Build Error - ReduceSum Operator Does Not Support Input of More Than Eight Dimensions

For example, the Parameter parameter does not support automatic type conversion. When the Parameter operator is used, an error is reported during data type conversion. The error message is as follows:

RuntimeError: Data type conversion of 'Parameter' is not supported, so data type int32 cannot be converted to data type float32 automatically.

For details, visit the following website:

MindSpore Operator Build Error - Error Reported Due to Inconsistent ScatterNdUpdate Operator Parameter Types

In addition, sometimes errors such as Response is empty, Try to send request before Open() and Try to get response before Open() may appear during the operator compilation process, as shown below:

>       result = self._graph_executor.compile(obj, args_list, phase, self._use_vm_mode())
E       RuntimeError: Response is empty
E
E       ----------------------------------------------------
E       - C++ Call Stack: (For framework developers)
E       ----------------------------------------------------
E       mindspore/ccsrc/backend/common/session/kernel_build_client.h:100 Response

The direct cause of this problem is usually a timeout caused by the subprocess of operator compilation hanging or the call blocking, which can be investigated from the following aspects:

  1. Check the logs to see if there are any other error logs before this error. If yes, please resolve the previous errors first. Some operator related issues (for example, TBE package is not installed properly on Ascend and NVCC not available on GPU) can cause subsequent such errors;

  2. If the graph kernel fusion feature is used, it is possible that the AKG operator compilation of the graph is stuck and timed out, and you can try to turn off the graph kernel fusion feature;

  3. You can try to reduce the number of processes for parallel compilation of operators on Ascend by using the environment variable MS_BUILD_PROCESS_NUM setting, with a value range of 1-24;

  4. Check the memory and CPU usage of the host. It is possible that the host’s memory and CPU usage are too high. As a result, the operator compilation process cannot be started and the compilation fails. You can try to identify the processes that occupy too much memory or CPU and optimize them;

  5. If you encounter this issue in a training environment on the cloud, you can try restarting the kernel.

Operator Execution Errors

Operator execution errors are mainly caused by improper input data , operator implementation, or operator initialization. Generally, the analogy method can be used to analyze operator execution errors.

For details, see the following example:

MindSpore Operator Execution Error - nn.GroupNorm Operator Output Exception

Insufficient Resources

During network debugging, Out Of Memory errors often occur. MindSpore divides the memory into four layers for management on the Ascend device, including runtime, context, dual cursors, and memory overcommitment.

For details about memory management and FAQs of MindSpore on the Ascend device, see MindSpore Ascend Memory Management.