From b7e702c8c77885e0e89139b10c1a2db4368e0539 Mon Sep 17 00:00:00 2001 From: Avijit Date: Mon, 8 Apr 2019 22:27:00 -0700 Subject: [PATCH] Release 0.12.0-rc-4 prep (#493) * Release 0.12.0-rc-4 prep * Update README.md * Modify testrunner to print error and failure outputs along with summary --- README.md | 9 ++++---- python/setup.in.py | 2 +- src/version.cc | 2 +- test/python/tensorflow/tf_unittest_runner.py | 24 +++++++++++++------- 4 files changed, 23 insertions(+), 14 deletions(-) diff --git a/README.md b/README.md index 530f39d7..8290a159 100644 --- a/README.md +++ b/README.md @@ -1,6 +1,6 @@ [![License](https://img.shields.io/badge/License-Apache%202.0-blue.svg)](https://github.com/NervanaSystems/ngraph-tf/blob/master/LICENSE) -[![Build Status](https://badge.buildkite.com/f20db2a4be0e82e493faa08de85953d45b313b3be12abf4acf.svg)](https://buildkite.com/ngraph/ngtf-master-cpu) +[![Build Status](https://badge.buildkite.com/f20db2a4be0e82e493faa08de85953d45b313b3be12abf4acf.svg?branch=master)](https://buildkite.com/ngraph/ngtf-master-cpu) # Intel(R) nGraph(TM) Compiler and runtime for TensorFlow* @@ -47,9 +47,10 @@ a variety of nGraph-enabled backends: CPU, GPU, and custom silicon like the This will produce something like this: TensorFlow version: 1.13.1 - nGraph bridge version: b'0.12.0-rc3' - nGraph version used for this build: b'0.21.0-rc.0+b638705' + nGraph bridge version: b'0.12.0-rc4' + nGraph version used for this build: b'0.17.0-rc.1+045b71e' TensorFlow version used for this build: v1.13.1-0-g6612da8951 + CXX11_ABI flag used for this build: 1 Next you can try out the TensorFlow models by adding one line to your existing TensorFlow model scripts and running them the usual way: @@ -89,7 +90,7 @@ The installation prerequisites are the same as described in the TensorFlow git clone https://github.com/NervanaSystems/ngraph-tf.git cd ngraph-tf - git checkout v0.12.0-rc3 + git checkout v0.12.0-rc4 2. Next run the following Python script to build TensorFlow, nGraph and the bridge. Please use Python 3.5: diff --git a/python/setup.in.py b/python/setup.in.py index 96537612..5524a86c 100644 --- a/python/setup.in.py +++ b/python/setup.in.py @@ -39,7 +39,7 @@ def get_tag(self): setup( name='ngraph_tensorflow_bridge', - version='0.12.0rc3', + version='0.12.0rc4', description='Intel nGraph compiler and runtime for TensorFlow', long_description=long_description, long_description_content_type="text/markdown", diff --git a/src/version.cc b/src/version.cc index 28e727e7..e7fb97ce 100644 --- a/src/version.cc +++ b/src/version.cc @@ -30,7 +30,7 @@ // candidate such as v0.7.0-rc0 // The code in master will always have the last released version number // with a suffix of '-master' -#define NG_TF_VERSION_SUFFIX "-rc3" +#define NG_TF_VERSION_SUFFIX "-rc4" #define VERSION_STR_HELPER(x) #x #define VERSION_STR(x) VERSION_STR_HELPER(x) diff --git a/test/python/tensorflow/tf_unittest_runner.py b/test/python/tensorflow/tf_unittest_runner.py index ed6ea9f0..9ee8abf0 100644 --- a/test/python/tensorflow/tf_unittest_runner.py +++ b/test/python/tensorflow/tf_unittest_runner.py @@ -250,9 +250,9 @@ def run_test(test_list, xml_report, verbosity=2): if test_result.wasSuccessful(): succeeded.append(test) elif test_result.failures: - failures.append(test) + failures.append(test_result.failures) elif test_result.errors: - errors.append(test) + errors.append(test_result.errors) summary = {"PASSED": succeeded, "FAILED": failures, "ERRORS": errors} return summary @@ -275,9 +275,11 @@ def print_results(status_list, invalid_list): if key is "PASSED": print(test + '\033[92m' + ' ..PASS' + '\033[0m') if key is "FAILED": - print(test + '\033[91m' + ' ..FAIL' + '\033[0m') + print(test[0][0].id() + '\033[91m' + ' ..FAIL' + '\033[0m') + print(test[0][1]) if key is "ERRORS": - print(test + '\033[33m' + ' ..ERROR' + '\033[0m') + print(test[0][0].id() + '\033[33m' + ' ..ERROR' + '\033[0m') + print(test[0][1]) if (len(invalid_list) != 0): print('\033[1m' + '\nInvalid Tests' + '\033[0m') @@ -288,10 +290,16 @@ def print_results(status_list, invalid_list): test_class_name = {} test_name = status_list[key] for test in test_name: - module, classname, testcase = test.split('.') - module_classname = module + '.' + classname - test_class_name[module_classname] = test_class_name.get( - module_classname, 0) + 1 + if key is "PASSED": + module, classname, testcase = test.split('.') + module_classname = module + '.' + classname + test_class_name[module_classname] = test_class_name.get( + module_classname, 0) + 1 + if key is "FAILED" or key is "ERRORS": + module, classname, testcase = test[0][0].id().split('.') + module_classname = module + '.' + classname + test_class_name[module_classname] = test_class_name.get( + module_classname, 0) + 1 for k in test_class_name: print('Number of tests ' + key + ' ' + k, test_class_name[k])