| Commit message (Collapse) | Author | Age | Files | Lines |
|
|
|
|
|
|
|
|
|
| |
Refresh patches and add zlib to DEPENDS. Also fetch pybind11 as it's
needed by ngraph.
Release notes:
https://software.intel.com/content/www/us/en/develop/articles/openvino-2021-4-lts-relnotes.html
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
| |
Use the convert-overrides.py to convert to new syntax and manually
fix some additional changes.
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
| |
When included in a project using find_package, it goes and looks for
headers and cmake module at an incorrect path.
Make sure that we are able to locate the headers and cmake module file.
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
| |
Temporarily disable opencl since intel-graphics-compiler and vc-intrinsics
compile failed since some api of llvm changed.
Signed-off-by: Changqing Li <changqing.li@windriver.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
|
|
| |
Update openvino to the latest one c5f7ad383e65 ("Fix
license header in Movidius sources") since it is include
a important license change from Intel private License to
Apache2.0 which is allowed external user to disclose it
without limit.
Signed-off-by: Liwei Song <liwei.song@windriver.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
| |
Signed-off-by: Naveen Saini <naveen.kumar.saini@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
|
| |
Include the header explicitly to avoid failures:
| inference-engine/src/plugin_api/ie_system_conf.h:21:31: error: 'exception_ptr' in namespace 'std' does not name a type; did you mean 'exception'?
| 21 | INFERENCE_ENGINE_API_CPP(std::exception_ptr&) CurrentException();
| | ^~~~~~~~~~~~~
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
|
| |
Release notes:
https://github.com/openvinotoolkit/openvino/releases/tag/2021.3
Remove a patch that isn't needed anymore and refresh the rest. Remove
dependency on onednn as it's using a forked version.
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
| |
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
|
|
|
| |
The EDK2 sources are DOS-style on server (contains CRLF) and this may
cause interop issues on Linux machines (patches do not apply clean) The
standard solution is to use the "dos2unix" bbclass, but that can modify
only files under the ${S} folder - meaning we have to move the edk2
checkout folder from ${S}/../edk2 to ${S}/edk2. Also, the LICENSE file
checksum is going to be impacted by the dos2unix conversion.
Signed-off-by: Robi Buranyi <rburanyi@google.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
| |
Ref:
https://github.com/intel/ipmctl/releases/tag/v03.00.00.0302
Signed-off-by: Naveen Saini <naveen.kumar.saini@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
ipmctl is a utility for configuring and managing Intel® Optane™ Persistent Memory modules (PMem).
It supports functionality to:
* Discover PMems on the platform.
* Provision the platform memory configuration.
* View and update the firmware on PMems.
* Configure data-at-rest security on PMems.
* Track health and performance of PMems.
* Debug and troubleshoot PMems.
* ipmctl refers to the following interface components:
libipmctl: An Application Programming Interface (API) library for managing PMems.
ipmctl: A Command Line Interface (CLI) application for configuring and managing PMems from the command line.
Signed-off-by: Naveen Saini <naveen.kumar.saini@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
| |
Drop a patch that has been merged and include another that was
accidently left off during the last upgrade.
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
| |
Make sure that ClDNN runtime is built by default. This would need clang
layer to be included as well.
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
|
| |
For release notes, see:
https://software.intel.com/content/www/us/en/develop/articles/openvino-relnotes.html
Refresh patches and use the now available option to build against system
version of pugixml.
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
| |
python2 is EOL, so switch shebang to python3
Signed-off-by: Changqing Li <changqing.li@windriver.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
| |
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
| |
Release notes:
https://software.intel.com/content/www/us/en/develop/articles/openvino-relnotes.html
Update the firmware files and refresh patches as well.
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
| |
Signed-off-by: sangeeta jain <sangeeta.jain@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
| |
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
For release notes, see:
https://software.intel.com/content/www/us/en/develop/articles/openvino-relnotes.html
The recipe has been changed to:
- disable unit tests as they rely on an older version of googletest.
- remove a patch as it was merged upstream.
- backport a patch instead of using a local version.
- refresh other patches.
- download mkl-dnn separately instead of using the submodule. The
project downloads a lot of other modules as well and we don't
want/use those.
- not download lfs documentation content.
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
| |
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
| |
Make sure that cmake files are installed at correct location so they
could be located using find_package.
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
|
|
|
| |
* Point to the renamed repository and change the project name to
openvino-inference-engine.
* Include ngraph as a submodule instead of using it as a separate
project.
* Use the new unit tests, the old ones have been deprecated.
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
fix compile error:
inference-engine/src/mkldnn_plugin/mkldnn_memory_solver.hpp:60:9: error: 'int64_t' does not name a type
| 60 | int64_t size;
| | ^~~~~~~
include stdint.h to fix it.
Signed-off-by: Liwei Song <liwei.song@windriver.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
| |
Add open-model-zoo, version 2020.1
Signed-off-by: Liwei Song <liwei.song@windriver.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
| |
Add required packages from requirements.txt to list of packages on which
we need to RDEPENDS.
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
| |
We don't want to install sample sources. So delete them and install the
binaries instead. Also make sure that we're installing the python
samples and modules at correct location. Remove the tweaks from local
patch and copy in the recipe itself.
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
| |
Release notes:
https://software.intel.com/en-us/articles/OpenVINO-RelNotes
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
| |
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
|
| |
Fix the dependency to be added when ptests are not enabled. Fixes:
| ERROR: Nothing PROVIDES '0' (but /meta-intel/dynamic-layers/openembedded-layer/recipes-support/opencv/dldt-inference-engine_2019r3.1.bb DEPENDS on or otherwise requires it)
| ERROR: Required build target 'dldt-inference-engine' has no buildable providers.
| Missing or unbuildable dependency chain was: ['dldt-inference-engine', '0']
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
| |
Install the OpenCL kernels and cldnn_global_custom_kernels.xml to allow
specification of OpenCL kernels for custom layers.
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
| |
Signed-off-by: Chin Huat Ang <chin.huat.ang@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
| |
* Make sure that USB udev rules for Intel(R) Movidius(TM) Neural Compute
Stick and Intel(R) Neural Compute Stick 2 are packaged.
* Package vpu firmware only when it is enabled.
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* There is interesting issue in do_patch, I was debugging strange
behavior with .bbappend where I've added another small patch.
And it started failing to configure completely.
bitbake -e shows that all .patch files are in SRC_URI and
log.do_patch shows that all were applied, but git diff (as well as
patches/series) shows only the last one added from the bbappend
to be applied.
This was caused by 8 existing patches in .bb file using ;patchdir=../
and my patch in .bbappend using ;patchdir=.. without slash at the end,
it should be fixed in quilt (or how do_patch is using it), but for
now just drop the trailing slash, because 99.9% recipes use ;patchdir=..
without the slash.
It's easily reproducible by removing the slash from the last patch
(without any bbappend).
Signed-off-by: Martin Jansa <Martin.Jansa@gmail.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* use the files from SRC_URI instead of fetching them during
configure
* use the firmware files CMake would download during configure
when not provided already, I haven't found any changelog to
find out what's the difference between
firmware_ma2x8x_784.zip firmware_ma2x8x_mdk_R9.8.zip
firmware_ma2450_759W.zip firmware_ma2450_784.zip
firmware_mv0262_784.zip firmware_mv0262_mdk_R9.8.zip
so to be safe use whatever CMake does and what I was testing
on raspberrypi4 before (because of missing this fix).
-- Downloading from https://download.01.org/opencv/2019/openvinotoolkit/R3/inference_engine/firmware_ma2450_759W.zip to dldt-inference-engine/2019r3-r0/git/inference-engine/temp/download/firmware_ma2450_759W.zip ...
-- Downloading from https://download.01.org/opencv/2019/openvinotoolkit/R3/inference_engine/firmware_ma2x8x_mdk_R9.8.zip to dldt-inference-engine/2019r3-r0/git/inference-engine/temp/download/firmware_ma2x8x_mdk_R9.8.zip ...
-- Downloading from https://download.01.org/opencv/2019/openvinotoolkit/R3/inference_engine/firmware_mv0262_mdk_R9.8.zip to dldt-inference-engine/2019r3-r0/git/inference-engine/temp/download/firmware_mv0262_mdk_R9.8.zip ...
* add PACKAGECONFIG verbose for VERBOSE_BUILD so it's easier
to debug which firmware files are being used, in this case:
log.do_configure will show:
-- ma2450=../mvnc/MvNCAPI-ma2450.mvcmd
-- ma2x8x=../mvnc/MvNCAPI-ma2x8x.mvcmd
-- mv0262=../mvnc/MvNCAPI-mv0262.mvcmd
* the patch is very loosely based on the patch removed here:
http://git.yoctoproject.org/cgit/cgit.cgi/meta-intel/diff/dynamic-layers/openembedded-layer/recipes-support/opencv/files/0003-Supply-firmware-at-build-time.patch?id=649456856dc38086e7d43d740d68198e3395b067
but I've reworked it competely to simplify this mess :)
Signed-off-by: Martin Jansa <Martin.Jansa@gmail.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
| |
Signed-off-by: Martin Jansa <Martin.Jansa@gmail.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
| |
* add PACKAGECONFIG for vpu
* add extra package for firmware files
* tested on rpi4 with NCS2
Signed-off-by: Martin Jansa <Martin.Jansa@gmail.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
| |
share/doc
* otherwise components depending on them won't be able to find them
Signed-off-by: Martin Jansa <Martin.Jansa@gmail.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
| |
Signed-off-by: Martin Jansa <Martin.Jansa@gmail.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
|
|
| |
Instead of letting clDNN build against intel_ocl_icd prebuilt binaries
under clDNN/common/intel_ocl_icd, configure cmake build to pick up
opencl-icd-loader headers and libraries from staging directory.
Do not set CMAKE_INSTALL_LOCAL_ONLY as it is unused.
Signed-off-by: Chin Huat Ang <chin.huat.ang@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
| |
Refresh patches so that they apply cleanly on 2019r3.
Signed-off-by: Chin Huat Ang <chin.huat.ang@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
| |
VPU plugins are untested, temporarily disable them.
Signed-off-by: Chin Huat Ang <chin.huat.ang@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
| |
Inference engine is still downloading and building it's own copy of
mkl-dnn, so remove it from DEPENDS.
Signed-off-by: Chin Huat Ang <chin.huat.ang@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
Package libmock_engine.so as part of dldt-inference-engine-ptest and
update run-ptest to set LD_LIBRARY_PATH to fix the following
InferenceEngineUnitTests failures:
FAIL: 12 tests, listed below:
FAIL: PluginTest.canCreatePlugin
FAIL: PluginTest.canCreatePluginUsingSmartPtr
FAIL: PluginTest.shouldThrowExceptionIfPluginNotExist
FAIL: PluginTest.canCallErrorHandlerIfNecessary
FAIL: PluginTest.canForwardPluginEnginePtr
FAIL: PluginTest.canSetConfiguration
FAIL: PluginDispatcherTests.canLoadMockPlugin
FAIL: PluginDispatcherTests.returnsIfLoadSuccessfull
FAIL: SharedObjectLoaderTests.canLoadExistedPlugin
FAIL: SharedObjectLoaderTests.canFindExistedMethod
FAIL: SharedObjectLoaderTests.throwIfMethodNofFoundInLibrary
FAIL: SharedObjectLoaderTests.canCallExistedMethod
Signed-off-by: Chin Huat Ang <chin.huat.ang@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
|
|
|
| |
Add PACKAGECONFIG[python3] for building dldt-inference-engine-python3
package which contains the inference engine python API.
Also tweak recipe to inherit python3native instead of relying on host
python as building the python API requires python3-cython which might
not be available on the host.
Signed-off-by: Chin Huat Ang <chin.huat.ang@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
Install clDNN to /usr/lib to resolve the following inference engine
error when running with GPU plugin:
[ ERROR ] Failed to create plugin libclDNNPlugin.so for device GPU
Please, check your environment
Cannot load library 'libclDNNPlugin.so': libclDNNPlugin.so: cannot open
shared object file: No such file or directory
/usr/src/debug/dldt-inference-engine/2019r2-r0/git/inference-engine/include/details/os/lin_shared_object_loader.h:36
/usr/src/debug/dldt-inference-engine/2019r2-r0/git/inference-engine/src/inference_engine/ie_core.cpp:277
Signed-off-by: Chin Huat Ang <chin.huat.ang@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* Release notes:
https://software.intel.com/en-us/articles/OpenVINO-RelNotes
* Enable unit tests to be built and tested using ptest mechanism.
* Include patches from Clear Linux for build fixes.
* Switch to using python3 and threading to using TBB. Switch ENABLE_OPENCV
to off so opencv from system is used.
* Remove do_install and patch Makefiles instead to install libraries correctly.
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|
|
This recipe builds the inference engine from opencv/dldt 2019 R1.1
release.
OpenVINO™ toolkit, short for Open Visual Inference and Neural network
Optimization toolkit, provides developers with improved neural network
performance on a variety of Intel® processors and helps further unlock
cost-effective, real-time vision applications.
The toolkit enables deep learning inference and easy heterogeneous
execution across multiple Intel® platforms (CPU, Intel® Processor Graphics)—providing
implementations across cloud architectures to edge device.
For more details, see:
https://01.org/openvinotoolkit
The recipe needs components from meta-oe so move it to
dynamic-layers/openembedded-layer. GPU plugin support needs intel-compute-runtime
which can be built by including clang layer in the mix as well.
CPU and GPU plugins have been sanity tested to work using
classification_sample. Further fine-tuning is still needed to improve
the performance.
Original patch by Anuj Mittal.
Signed-off-by: Chin Huat Ang <chin.huat.ang@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
|