| Commit message (Collapse) | Author | Age | Files | Lines |
|
|
|
|
|
|
|
|
|
|
|
|
| |
* Remove all openvino related recipes, tests and other data from meta-intel
layer as a new layer (meta-oepnvino) specific to openvino has been created.
* Update openvino documentation
meta-openvino layer URL:
https://github.com/intel/meta-openvino
Signed-off-by: Yogesh Tyagi <yogesh.tyagi@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
|
| |
Update test_mkldnn_rnn_api to run batch test 'test_rnn_all'.
Previously it was not ready and testcase was running 'test_rnn_small'.
test_rnn_small is no longer available.
Signed-off-by: sangeeta jain <sangeeta.jain@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
|
| |
Previously, the status return with 0 even though the file
required was not exist.
Fixed the file exist checking for onednn source file required
for compilation testing.
Signed-off-by: Yeoh Ee Peng <ee.peng.yeoh@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
| |
Update test pakcages checking and referencing as mkldnn
rename to onednn.
Signed-off-by: Yeoh Ee Peng <ee.peng.yeoh@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
| |
Enable checking of available inference engine devices.
Signed-off-by: Yeoh Ee Peng <ee.peng.yeoh@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
|
|
|
| |
Add automated tests that:
- test that mkl_dnn shared libraries and headers can be used to
compile mkl_dnn sample application and execute it
- test that mkl_dnn various api (eg. convolution, deconv, rnn, etc)
are executable
- add extra packages checking for test compiling mkldnn application
Signed-off-by: Yeoh Ee Peng <ee.peng.yeoh@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
Add sanity tests for inference engine:
- test inference engine c/cpp shared library
- test inference engine python api
- test inference engine cpu, gpu, myriad plugin
Add sanity tests for model optimizer
- test model optmizer can generate ir
Licenses:
- classification_sample.py
license: Apache 2.0
source: <install_root>/deployment_tools/inference_engine/samples/*
Signed-off-by: Yeoh Ee Peng <ee.peng.yeoh@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|