From b4755c5aa7ede517bf9bc559e9247c050c6711f2 Mon Sep 17 00:00:00 2001 From: liaogang Date: Mon, 14 Aug 2017 15:50:28 +0800 Subject: [PATCH 1/7] Demangle exception call stack for PADDLE_ENFORCE --- paddle/platform/enforce.h | 47 ++++++++++++++++++++++++++++++++------- 1 file changed, 39 insertions(+), 8 deletions(-) diff --git a/paddle/platform/enforce.h b/paddle/platform/enforce.h index 337a059fb1..aa0660df88 100644 --- a/paddle/platform/enforce.h +++ b/paddle/platform/enforce.h @@ -14,14 +14,20 @@ limitations under the License. */ #pragma once -#include +#include // for dladdr +#include // for backtrace #include #include #include #include + #include "paddle/string/printf.h" #include "paddle/string/to_string.h" +#ifdef __GNUC__ +#include // for __cxa_demangle +#endif + #ifndef PADDLE_ONLY_CPU #include "paddle/platform/dynload/cublas.h" @@ -39,6 +45,19 @@ limitations under the License. */ namespace paddle { namespace platform { +namespace { +#ifdef __GNUC__ +inline std::string demangle(std::string name) { + int status = -4; // some arbitrary value to eliminate the compiler warning + std::unique_ptr res{ + abi::__cxa_demangle(name.c_str(), NULL, NULL, &status), std::free}; + return (status == 0) ? res.get() : name; +} +#else +inline std::string demangle(std::string name) { return name; } +#endif +} + struct EnforceNotMet : public std::exception { std::exception_ptr exp_; std::string err_str_; @@ -48,15 +67,27 @@ struct EnforceNotMet : public std::exception { std::rethrow_exception(exp_); } catch (const std::exception& exp) { std::ostringstream sout; + sout << string::Sprintf("%s at [%s:%d]", exp.what(), f, l) << std::endl; - sout << "Call Stacks: " << std::endl; + sout << "PaddlePaddle Call Stacks: " << std::endl; + void* call_stack[TRACE_STACK_LIMIT]; - int sz = backtrace(call_stack, TRACE_STACK_LIMIT); - auto line = backtrace_symbols(call_stack, sz); - for (int i = 0; i < sz; ++i) { - sout << line[i] << std::endl; + auto size = backtrace(call_stack, TRACE_STACK_LIMIT); + auto symbols = backtrace_symbols(call_stack, size); + + Dl_info info; + for (int i = 0; i < size; ++i) { + if (dladdr(call_stack[i], &info)) { + auto demangled = demangle(info.dli_sname); + sout << string::Sprintf( + "%-3d %*0p %s + %zd\n", i, 2 + sizeof(void*) * 2, call_stack[i], + demangled, (char*)call_stack[i] - (char*)info.dli_saddr); + } else { + sout << string::Sprintf("%-3d %*0p %s\n", i, 2 + sizeof(void*) * 2, + call_stack[i]); + } } - free(line); + free(symbols); err_str_ = sout.str(); } } @@ -170,7 +201,7 @@ inline void throw_on_error(T e) { * PADDLE_ENFORCE_EQ(a, b); * * will raise an expression described as follows: - * "enforce a == b failed, 1 != 2" with detailed stack infomation. + * "enforce a == b failed, 1 != 2" with detailed stack information. * * extra messages is also supported, for example: * PADDLE_ENFORCE(a, b, "some simple enforce failed between %d numbers", 2) From 8210bcea6116be008aa027c6e868acbf5ebfeda6 Mon Sep 17 00:00:00 2001 From: liaogang Date: Mon, 14 Aug 2017 15:59:02 +0800 Subject: [PATCH 2/7] Using static_cast to replace (char*) --- paddle/platform/enforce.h | 8 +++++--- 1 file changed, 5 insertions(+), 3 deletions(-) diff --git a/paddle/platform/enforce.h b/paddle/platform/enforce.h index aa0660df88..cc38dc4ffe 100644 --- a/paddle/platform/enforce.h +++ b/paddle/platform/enforce.h @@ -79,9 +79,11 @@ struct EnforceNotMet : public std::exception { for (int i = 0; i < size; ++i) { if (dladdr(call_stack[i], &info)) { auto demangled = demangle(info.dli_sname); - sout << string::Sprintf( - "%-3d %*0p %s + %zd\n", i, 2 + sizeof(void*) * 2, call_stack[i], - demangled, (char*)call_stack[i] - (char*)info.dli_saddr); + auto addr_offset = static_cast(call_stack[i]) - + static_cast(info.dli_saddr); + sout << string::Sprintf("%-3d %*0p %s + %zd\n", i, + 2 + sizeof(void*) * 2, call_stack[i], + demangled, addr_offset); } else { sout << string::Sprintf("%-3d %*0p %s\n", i, 2 + sizeof(void*) * 2, call_stack[i]); From ffbb4c19634778d5380954dd55f4f56ec17b3859 Mon Sep 17 00:00:00 2001 From: liaogang Date: Mon, 14 Aug 2017 16:27:25 +0800 Subject: [PATCH 3/7] memory.h for unique_ptr --- paddle/platform/enforce.h | 1 + 1 file changed, 1 insertion(+) diff --git a/paddle/platform/enforce.h b/paddle/platform/enforce.h index cc38dc4ffe..15fdf7a94f 100644 --- a/paddle/platform/enforce.h +++ b/paddle/platform/enforce.h @@ -17,6 +17,7 @@ limitations under the License. */ #include // for dladdr #include // for backtrace #include +#include #include #include #include From 186fb0c1185b6b1b94a7eeac54fa1cbd001debfd Mon Sep 17 00:00:00 2001 From: Yu Yang Date: Mon, 14 Aug 2017 16:31:54 +0800 Subject: [PATCH 4/7] Remove input_format in backward.cc --- paddle/framework/backward.cc | 10 +++------- 1 file changed, 3 insertions(+), 7 deletions(-) diff --git a/paddle/framework/backward.cc b/paddle/framework/backward.cc index 315bdde76d..855e2cae20 100644 --- a/paddle/framework/backward.cc +++ b/paddle/framework/backward.cc @@ -127,11 +127,8 @@ std::shared_ptr BackwardRecursive( net->ops_[op_offset]->Rename(name, dup_outputs.back()); } insert_position.push_back( - {dup_op.back(), - OpRegistry::CreateOp( - "add", {{"X", {dup_outputs}}}, {{"Out", {name}}}, - {{"input_format", - std::vector{0, static_cast(dup_outputs.size())}}})}); + {dup_op.back(), OpRegistry::CreateOp("add", {{"X", {dup_outputs}}}, + {{"Out", {name}}}, {})}); } insert_position.sort( @@ -140,7 +137,6 @@ std::shared_ptr BackwardRecursive( for (auto& pos : insert_position) { net->InsertOp(pos.first + 1, pos.second); } - } else { std::shared_ptr grad_op = OpRegistry::CreateGradOp(forwardOp); @@ -176,7 +172,7 @@ std::shared_ptr BackwardRecursive( net->type_ = "@GENERATED_BACKWARD@"; net->CompleteAddOp(); return net; -} +} // namespace framework // See header for comments std::shared_ptr Backward( From 32a60971f05da4e65b913752608fd0ec68d028a0 Mon Sep 17 00:00:00 2001 From: wanghaoshuang Date: Mon, 14 Aug 2017 17:45:26 +0800 Subject: [PATCH 5/7] Fix pnpair_evaluator. --- .../trainer_config_helpers/evaluators.py | 20 ++++++++----------- 1 file changed, 8 insertions(+), 12 deletions(-) diff --git a/python/paddle/trainer_config_helpers/evaluators.py b/python/paddle/trainer_config_helpers/evaluators.py index 567521ee9d..e272f76a81 100644 --- a/python/paddle/trainer_config_helpers/evaluators.py +++ b/python/paddle/trainer_config_helpers/evaluators.py @@ -230,9 +230,8 @@ def auc_evaluator( def pnpair_evaluator( input, label, - info, - name=None, - weight=None, ): + weight, + name=None, ): """ Positive-negative pair rate Evaluator which adapts to rank task like learning to rank. This evaluator must contain at least three layers. @@ -241,27 +240,24 @@ def pnpair_evaluator( .. code-block:: python - eval = pnpair_evaluator(input, info, label) + eval = pnpair_evaluator(input, label, weight) - :param name: Evaluator name. - :type name: None|basestring :param input: Input Layer name. The output prediction of network. :type input: LayerOutput :param label: Label layer name. :type label: LayerOutput - :param info: Label layer name. (TODO, explaination) - :type info: LayerOutput :param weight: Weight Layer name. It should be a matrix with size [sample_num, 1]. (TODO, explaination) :type weight: LayerOutput + :param name: Evaluator name. + :type name: None|basestring """ evaluator_base( - name=name, - type="pnpair", input=input, + type="pnpair", label=label, - info=info, - weight=weight) + weight=weight, + name=name, ) @evaluator(EvaluatorAttribute.FOR_CLASSIFICATION) From 957aa691b49037bcf245f848706e85ac2649c56a Mon Sep 17 00:00:00 2001 From: wanghaoshuang Date: Mon, 14 Aug 2017 18:47:38 +0800 Subject: [PATCH 6/7] Fix pnpair_evaluator. 1. Append info into input after label and before weight. --- .../trainer_config_helpers/evaluators.py | 28 ++++++++++++------- 1 file changed, 18 insertions(+), 10 deletions(-) diff --git a/python/paddle/trainer_config_helpers/evaluators.py b/python/paddle/trainer_config_helpers/evaluators.py index 72cfbecf6d..57979db4de 100644 --- a/python/paddle/trainer_config_helpers/evaluators.py +++ b/python/paddle/trainer_config_helpers/evaluators.py @@ -297,7 +297,8 @@ def auc_evaluator( def pnpair_evaluator( input, label, - weight, + info, + weight=None, name=None, ): """ Positive-negative pair rate Evaluator which adapts to rank task like @@ -307,22 +308,29 @@ def pnpair_evaluator( .. code-block:: python - eval = pnpair_evaluator(input, label, weight) + eval = pnpair_evaluator(input, label, info) :param input: Input Layer name. The output prediction of network. :type input: LayerOutput :param label: Label layer name. :type label: LayerOutput + :param info: Info layer name. (TODO, explaination) + :type info: LayerOutput :param weight: Weight Layer name. It should be a matrix with size [sample_num, 1]. (TODO, explaination) :type weight: LayerOutput :param name: Evaluator name. :type name: None|basestring """ + if not isinstance(input, list): + input = [input] + if label: + input.append(label) + if info: + input.append(info) evaluator_base( input=input, type="pnpair", - label=label, weight=weight, name=name, ) @@ -425,12 +433,12 @@ def chunk_evaluator( .. code-block:: text - Scheme Description + Scheme Description plain Use the same label for the whole chunk. - IOB Two labels for chunk type X, B-X for chunk begining and I-X for chunk inside. + IOB Two labels for chunk type X, B-X for chunk begining and I-X for chunk inside. IOE Two labels for chunk type X, E-X for chunk ending and I-X for chunk inside. - IOBES Four labels for chunk type X, B-X for chunk begining, I-X for chunk inside, E-X for chunk end and S-X for single word chunk. - + IOBES Four labels for chunk type X, B-X for chunk begining, I-X for chunk inside, E-X for chunk end and S-X for single word chunk. + To make it clear, let's illustrate by an NER example. Assuming that there are three named entity types including ORG, PER and LOC which are called 'chunk type' here, if 'IOB' scheme were used, the label set will be extended to a set including B-ORG, I-ORG, B-PER, I-PER, B-LOC, I-LOC and O, @@ -447,7 +455,7 @@ def chunk_evaluator( tagType = label % numTagType chunkType = label / numTagType otherChunkType = numChunkTypes - + The following table shows the mapping rule between tagType and tag type in each scheme. .. code-block:: text @@ -471,7 +479,7 @@ def chunk_evaluator( O 6 In this example, chunkType has three values: 0 for ORG, 1 for PER, 2 for LOC, because the scheme is - "IOB" so tagType has two values: 0 for B and 1 for I. + "IOB" so tagType has two values: 0 for B and 1 for I. Here we will use I-LOC to explain the above mapping rules in detail. For I-LOC, the label id is 5, so we can get tagType=1 and chunkType=2, which means I-LOC is a part of NER chunk LOC and the tag is I. @@ -482,7 +490,7 @@ def chunk_evaluator( eval = chunk_evaluator(input, label, chunk_scheme, num_chunk_types) - + :param input: The input layers. :type input: LayerOutput :param label: An input layer containing the ground truth label. From 991c4d807959fc1fc9e54d17f545fd46e0226bbf Mon Sep 17 00:00:00 2001 From: Yan Chunwei Date: Mon, 14 Aug 2017 19:04:38 +0800 Subject: [PATCH 7/7] add some doc to backward (#3474) --- paddle/framework/backward.cc | 32 +++++++++++++++++++++++--------- 1 file changed, 23 insertions(+), 9 deletions(-) diff --git a/paddle/framework/backward.cc b/paddle/framework/backward.cc index 855e2cae20..2118c9d5d4 100644 --- a/paddle/framework/backward.cc +++ b/paddle/framework/backward.cc @@ -30,6 +30,7 @@ static void ForEachVarName(Map& names, T callback) { } } +// return whether all the names + suffixes in the set static bool AllInSet( const std::map>& names, const std::string& suffix, const std::unordered_set& set) { @@ -48,7 +49,7 @@ static std::shared_ptr NOP() { return net_op; } -// Get backward operator from a forward operator, recursively implementation. +// Get backward operator from a forward operator, a recursive implementation. // // no_grad_names the gradient variable names without gradient calculating. // @@ -56,27 +57,30 @@ static std::shared_ptr NOP() { // BackwardRecursive. use `uid = uniq_id++;` to get the unique index, and // pass `uniq_id` through recursive calling. // -// returns The backward operator. For simple situation, it is a simple -// operator. For complex situation, it is a NetOp. +// returns The backward operator. In a simple situation, it may be a simple +// operator, in a complex situation, it maybe a NetOp. // // See Backward.h for details static std::shared_ptr BackwardRecursive( const OperatorBase& forwardOp, std::unordered_set& no_grad_names, size_t& uniq_id); + std::shared_ptr BackwardRecursive( const OperatorBase& forwardOp, std::unordered_set& no_grad_names, size_t& uniq_id) { // If all input gradients of forwarding operator do not need to calculate, // just return an NOP. Not return null ptr because NOP does not take - // too much time for calculation, but it is useful for simplifying logic. - if (AllInSet(forwardOp.inputs_, kGradVarSuffix, no_grad_names)) { + // much time for calculation, but it is useful for simplifying logic. + if (AllInSet(forwardOp.inputs_ /*names*/, kGradVarSuffix /*suffix*/, + no_grad_names /*set*/)) { return NOP(); } // All output gradients of forwarding operator do not need to calculate. // Then all input gradients cannot be computed at all, and we put them into // `no_grad_names` set. Return an NOP. - if (AllInSet(forwardOp.outputs_, kGradVarSuffix, no_grad_names)) { + if (AllInSet(forwardOp.outputs_ /*names*/, kGradVarSuffix /*suffix*/, + no_grad_names /*set*/)) { ForEachVarName(forwardOp.inputs_, [&no_grad_names](const std::string& name) -> bool { no_grad_names.insert(GradVarName(name)); @@ -93,11 +97,11 @@ std::shared_ptr BackwardRecursive( auto& forwardNet = static_cast(forwardOp); // Map from output gradient variable name to operator's indices in - // backward net. That operator generates that variable. + // backward net's ops_. That operator generates that variable. std::unordered_map> dup_output_ops; size_t local_op_id = 0; - // reversely travel forwardNet + // reversely travel forwardNet and collect all duplicate outputs. for (auto it = forwardNet.ops_.rbegin(); it != forwardNet.ops_.rend(); ++it, ++local_op_id) { auto fwd = *it; @@ -112,25 +116,35 @@ std::shared_ptr BackwardRecursive( // Get unique ID for this method. auto uid = uniq_id++; // TODO(dzh): more comment + // multiple operators which have the same output (y for example) may + // overwrite the same y variable when backward, special operations are token + // to handle this case. For each duplicate output, rename it to an alias + // (original name with a offset), append an `add` op for its operator, + // and finally sum all the alias variable to the final output variable y. using Pos = std::pair>; std::list insert_position; for (auto& dup_output_op : dup_output_ops) { const std::string& name = dup_output_op.first; auto& dup_op = dup_output_op.second; + // no duplicate output if (dup_op.size() == 1) continue; - std::vector dup_outputs; + // process the duplicate outputs + std::vector dup_outputs; for (size_t i = 0; i < dup_op.size(); ++i) { + // rename each duplicate output to an alias auto op_offset = dup_op[i]; dup_outputs.push_back(name + "@RENAME@" + std::to_string(uid) + "@" + std::to_string(i)); net->ops_[op_offset]->Rename(name, dup_outputs.back()); } + // collect all the offset to append `add` op for each alias insert_position.push_back( {dup_op.back(), OpRegistry::CreateOp("add", {{"X", {dup_outputs}}}, {{"Out", {name}}}, {})}); } + // make sure the inserted `add` ops follow the BFS order. insert_position.sort( [](const Pos& l, const Pos& r) { return l.first > r.first; });