Commit Graph

14 Commits

Author SHA1 Message Date
Anuj Mittal
1937c72e02 openvino-inference-engine: upgrade 2020.3 -> 2020.4
For release notes, see:
https://software.intel.com/content/www/us/en/develop/articles/openvino-relnotes.html

The recipe has been changed to:
 - disable unit tests as they rely on an older version of googletest.
 - remove a patch as it was merged upstream.
 - backport a patch instead of using a local version.
 - refresh other patches.
 - download mkl-dnn separately instead of using the submodule. The
   project downloads a lot of other modules as well and we don't
   want/use those.
 - not download lfs documentation content.

Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
2020-08-14 09:48:04 +08:00
Anuj Mittal
9ddf617311 openvino-inference-engine: fix installation of ngraph cmake
Make sure that cmake files are installed at correct location so they
could be located using find_package.

Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
2020-06-24 09:55:44 +08:00
Anuj Mittal
78386dd80b openvino-inference-engine: upgrade 2020.1 -> 2020.3.0
* Point to the renamed repository and change the project name to
  openvino-inference-engine.

* Include ngraph as a submodule instead of using it as a separate
  project.

* Use the new unit tests, the old ones have been deprecated.

Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
2020-06-11 15:35:56 +08:00
LiweiSong
7497da8042 dldt-inference-engine: fix int64_t does not name a type compile error
fix compile error:

inference-engine/src/mkldnn_plugin/mkldnn_memory_solver.hpp:60:9: error: 'int64_t' does not name a type
|    60 |         int64_t size;
|       |         ^~~~~~~

include stdint.h to fix it.

Signed-off-by: Liwei Song <liwei.song@windriver.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
2020-06-03 15:47:47 +08:00
Anuj Mittal
af4fdcc321 dldt-inference-engine: fix installation of samples
We don't want to install sample sources. So delete them and install the
binaries instead. Also make sure that we're installing the python
samples and modules at correct location. Remove the tweaks from local
patch and copy in the recipe itself.

Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
2020-04-13 15:22:18 +08:00
Anuj Mittal
89c0fe5dce dldt: upgrade 2019r3.1 -> 2020r1
Release notes:
https://software.intel.com/en-us/articles/OpenVINO-RelNotes

Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
2020-04-10 11:29:22 +08:00
Anuj Mittal
9796a61261 dldt-inference-engine: use libdir when installing python modules
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
2020-02-25 15:03:03 +08:00
Martin Jansa
1a669124ec dldt-inference-engine: use firmware files from SRC_URI
* use the files from SRC_URI instead of fetching them during
  configure
* use the firmware files CMake would download during configure
  when not provided already, I haven't found any changelog to
  find out what's the difference between
  firmware_ma2x8x_784.zip   firmware_ma2x8x_mdk_R9.8.zip
  firmware_ma2450_759W.zip  firmware_ma2450_784.zip
  firmware_mv0262_784.zip   firmware_mv0262_mdk_R9.8.zip
  so to be safe use whatever CMake does and what I was testing
  on raspberrypi4 before (because of missing this fix).

  -- Downloading from https://download.01.org/opencv/2019/openvinotoolkit/R3/inference_engine/firmware_ma2450_759W.zip to dldt-inference-engine/2019r3-r0/git/inference-engine/temp/download/firmware_ma2450_759W.zip ...
  -- Downloading from https://download.01.org/opencv/2019/openvinotoolkit/R3/inference_engine/firmware_ma2x8x_mdk_R9.8.zip to dldt-inference-engine/2019r3-r0/git/inference-engine/temp/download/firmware_ma2x8x_mdk_R9.8.zip ...
  -- Downloading from https://download.01.org/opencv/2019/openvinotoolkit/R3/inference_engine/firmware_mv0262_mdk_R9.8.zip to dldt-inference-engine/2019r3-r0/git/inference-engine/temp/download/firmware_mv0262_mdk_R9.8.zip ...

* add PACKAGECONFIG verbose for VERBOSE_BUILD so it's easier
  to debug which firmware files are being used, in this case:
  log.do_configure will show:
  -- ma2450=../mvnc/MvNCAPI-ma2450.mvcmd
  -- ma2x8x=../mvnc/MvNCAPI-ma2x8x.mvcmd
  -- mv0262=../mvnc/MvNCAPI-mv0262.mvcmd

* the patch is very loosely based on the patch removed here:
  http://git.yoctoproject.org/cgit/cgit.cgi/meta-intel/diff/dynamic-layers/openembedded-layer/recipes-support/opencv/files/0003-Supply-firmware-at-build-time.patch?id=649456856dc38086e7d43d740d68198e3395b067
  but I've reworked it competely to simplify this mess :)

Signed-off-by: Martin Jansa <Martin.Jansa@gmail.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
2019-10-25 10:05:14 +08:00
Martin Jansa
fd9d57a992 dldt-inference-engine: install extension headers in includedir instead of share/doc
* otherwise components depending on them won't be able to find them

Signed-off-by: Martin Jansa <Martin.Jansa@gmail.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
2019-10-24 12:07:30 +08:00
Chin Huat Ang
649456856d dldt-inference-engine: update 2019r2 -> 2019r3
Refresh patches so that they apply cleanly on 2019r3.

Signed-off-by: Chin Huat Ang <chin.huat.ang@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
2019-10-21 07:20:21 +08:00
Chin Huat Ang
eff93c1db6 dldt-inference-engine: fix ptest failures
Package libmock_engine.so as part of dldt-inference-engine-ptest and
update run-ptest to set LD_LIBRARY_PATH to fix the following
InferenceEngineUnitTests failures:

  FAIL: 12 tests, listed below:
  FAIL: PluginTest.canCreatePlugin
  FAIL: PluginTest.canCreatePluginUsingSmartPtr
  FAIL: PluginTest.shouldThrowExceptionIfPluginNotExist
  FAIL: PluginTest.canCallErrorHandlerIfNecessary
  FAIL: PluginTest.canForwardPluginEnginePtr
  FAIL: PluginTest.canSetConfiguration
  FAIL: PluginDispatcherTests.canLoadMockPlugin
  FAIL: PluginDispatcherTests.returnsIfLoadSuccessfull
  FAIL: SharedObjectLoaderTests.canLoadExistedPlugin
  FAIL: SharedObjectLoaderTests.canFindExistedMethod
  FAIL: SharedObjectLoaderTests.throwIfMethodNofFoundInLibrary
  FAIL: SharedObjectLoaderTests.canCallExistedMethod

Signed-off-by: Chin Huat Ang <chin.huat.ang@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
2019-10-21 07:19:43 +08:00
Chin Huat Ang
611a831306 dldt-inference-engine: fix clDNN install directory
Install clDNN to /usr/lib to resolve the following inference engine
error when running with GPU plugin:

    [ ERROR ] Failed to create plugin libclDNNPlugin.so for device GPU
    Please, check your environment
    Cannot load library 'libclDNNPlugin.so': libclDNNPlugin.so: cannot open
    shared object file: No such file or directory
    /usr/src/debug/dldt-inference-engine/2019r2-r0/git/inference-engine/include/details/os/lin_shared_object_loader.h:36

    /usr/src/debug/dldt-inference-engine/2019r2-r0/git/inference-engine/src/inference_engine/ie_core.cpp:277

Signed-off-by: Chin Huat Ang <chin.huat.ang@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
2019-09-30 22:28:09 +08:00
Anuj Mittal
90b225ce8a dldt-inference-engine: upgrade 2019r1.1 -> 2019r2
* Release notes:
https://software.intel.com/en-us/articles/OpenVINO-RelNotes

* Enable unit tests to be built and tested using ptest mechanism.

* Include patches from Clear Linux for build fixes.

* Switch to using python3 and threading to using TBB. Switch ENABLE_OPENCV
to off so opencv from system is used.

* Remove do_install and patch Makefiles instead to install libraries correctly.

Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
2019-09-30 15:34:44 +08:00
Chin Huat Ang
096598691d dldt-inference-engine: add recipe
This recipe builds the inference engine from opencv/dldt 2019 R1.1
release.

OpenVINO™ toolkit, short for Open Visual Inference and Neural network
Optimization toolkit, provides developers with improved neural network
performance on a variety of Intel® processors and helps further unlock
cost-effective, real-time vision applications.

The toolkit enables deep learning inference and easy heterogeneous
execution across multiple Intel® platforms (CPU, Intel® Processor Graphics)—providing
implementations across cloud architectures to edge device.

For more details, see:
https://01.org/openvinotoolkit

The recipe needs components from meta-oe so move it to
dynamic-layers/openembedded-layer. GPU plugin support needs intel-compute-runtime
which can be built by including clang layer in the mix as well.

CPU and GPU plugins have been sanity tested to work using
classification_sample. Further fine-tuning is still needed to improve
the performance.

Original patch by Anuj Mittal.

Signed-off-by: Chin Huat Ang <chin.huat.ang@intel.com>
Signed-off-by: Anuj Mittal <anuj.mittal@intel.com>
2019-09-28 17:18:30 +08:00