代码拉取完成,页面将自动刷新
同步操作将从 src-openEuler/openvino 强制同步,此操作会覆盖自 Fork 仓库以来所做的任何修改,且无法恢复!!!
确定后同步将在后台操作,完成时将刷新页面,请耐心等待。
#
# Copyright (C) 2024 Intel Corporation
#
# Please submit issues or comments https://github.com/openvinotoolkit/openvino/issues
%define shlib lib%{name}
%define prj_name OpenVINO
%global openeuler_release 2
Name: openvino
Version: 2024.3.0
Release: %{openeuler_release}%{?dist}
Summary: A toolkit for optimizing and deploying AI inference
License: Apache-2.0 AND BSD-2-Clause AND BSD-3-Clause AND HPND AND JSON AND MIT AND OFL-1.1 AND Zlib
URL: https://github.com/openvinotoolkit/openvino
# The source tarball contains openvino all submodules initialization and update
# $ git clone --depth 1 --branch 2024.3.0 https://github.com/openvinotoolkit/openvino.git
# $ cd openvino && git submodule update --init –recursive
Source0: %{name}-%{version}.tar.gz
#These packages are only for x86_64 and aarch64
ExclusiveArch: x86_64 aarch64
BuildRequires: cmake
BuildRequires: fdupes
BuildRequires: gcc-c++
BuildRequires: tbb-devel
BuildRequires: pkgconfig
BuildRequires: zstd
BuildRequires: opencl-headers
BuildRequires: pkgconfig(snappy)
BuildRequires: scons
BuildRequires: libatomic
%description
OpenVINO is an open-source toolkit for optimizing and deploying AI inference.
#
%package -n %{shlib}
Summary: Shared library for OpenVINO toolkit
%description -n %{shlib}
OpenVINO is an open-source toolkit for optimizing and deploying AI inference.
This package provides the shared library for OpenVINO.
#
%package -n lib%{name}-devel
Summary: Headers and sources for OpenVINO toolkit
Requires: %{shlib} = %{version}
Requires: lib%{name}_ir_frontend = %{version}
Requires: lib%{name}_onnx_frontend = %{version}
Requires: lib%{name}_paddle_frontend = %{version}
Requires: lib%{name}_pytorch_frontend = %{version}
Requires: lib%{name}_tensorflow_frontend = %{version}
Requires: lib%{name}_tensorflow_lite_frontend = %{version}
Requires: tbb-devel
Recommends: lib%{name}-auto-batch-plugin = %{version}
Recommends: lib%{name}-auto-plugin = %{version}
Recommends: lib%{name}-hetero-plugin-%{version} = %{version}
%ifarch x86_64
Recommends: lib%{name}-intel-cpu-plugin = %{version}
Recommends: lib%{name}-intel-gpu-plugin = %{version}
%endif
%ifarch aarch64
Recommends: lib%{name}-arm-cpu-plugin = %{version}
%endif
%description -n lib%{name}-devel
OpenVINO is an open-source toolkit for optimizing and deploying AI inference.
This package provides the headers and sources for developing applications with
OpenVINO.
## Plugins ##
#
%package -n lib%{name}-auto-plugin
Summary: Auto / Multi software plugin for OpenVINO toolkit
%description -n lib%{name}-auto-plugin
OpenVINO is an open-source toolkit for optimizing and deploying AI inference.
This package provides the Auto / Multi software plugin for OpenVINO.
#
%package -n lib%{name}-auto-batch-plugin
Summary: Automatic batch software plugin for OpenVINO toolkit
%description -n lib%{name}-auto-batch-plugin
OpenVINO is an open-source toolkit for optimizing and deploying AI inference.
This package provides the automatic batch software plugin for OpenVINO.
#
%package -n lib%{name}-hetero-plugin
Summary: Hetero frontend for Intel OpenVINO toolkit
%description -n lib%{name}-hetero-plugin
OpenVINO is an open-source toolkit for optimizing and deploying AI inference.
This package provides the hetero frontend for OpenVINO.
%ifarch x86_64
#
%package -n lib%{name}-intel-cpu-plugin
Summary: Intel CPU plugin for OpenVINO toolkit
%description -n lib%{name}-intel-cpu-plugin
OpenVINO is an open-source toolkit for optimizing and deploying AI inference.
This package provides the intel CPU plugin for OpenVINO for %{x86_64} archs.
#
%package -n lib%{name}-intel-gpu-plugin
Summary: Intel GPU plugin for OpenVINO toolkit
Requires: ocl-icd
%description -n lib%{name}-intel-gpu-plugin
OpenVINO is an open-source toolkit for optimizing and deploying AI inference.
This package provides the intel GPU plugin for OpenVINO for %{x86_64} archs.
%endif
%ifarch aarch64
#
%package -n lib%{name}-arm-cpu-plugin
Summary: ARM CPU plugin for OpenVINO toolkit
%description -n lib%{name}-arm-cpu-plugin
OpenVINO is an open-source toolkit for optimizing and deploying AI inference.
This package provides the arm CPU plugin for OpenVINO for %{aarch} archs.
%endif
## Frontend shared libs ##
#
%package -n lib%{name}_ir_frontend
Summary: Paddle frontend for Intel OpenVINO toolkit
%description -n lib%{name}_ir_frontend
OpenVINO is an open-source toolkit for optimizing and deploying AI inference.
This package provides the ir frontend for OpenVINO.
#
%package -n lib%{name}_onnx_frontend
Summary: Onnx frontend for OpenVINO toolkit
%description -n lib%{name}_onnx_frontend
OpenVINO is an open-source toolkit for optimizing and deploying AI inference.
This package provides the onnx frontend for OpenVINO.
#
%package -n lib%{name}_paddle_frontend
Summary: Paddle frontend for Intel OpenVINO toolkit
%description -n lib%{name}_paddle_frontend
OpenVINO is an open-source toolkit for optimizing and deploying AI inference.
This package provides the paddle frontend for OpenVINO.
#
%package -n lib%{name}_pytorch_frontend
Summary: PyTorch frontend for OpenVINO toolkit
%description -n lib%{name}_pytorch_frontend
OpenVINO is an open-source toolkit for optimizing and deploying AI inference.
This package provides the pytorch frontend for OpenVINO.
#
%package -n lib%{name}_tensorflow_frontend
Summary: TensorFlow frontend for OpenVINO toolkit
%description -n lib%{name}_tensorflow_frontend
OpenVINO is an open-source toolkit for optimizing and deploying AI inference.
This package provides the tensorflow frontend for OpenVINO.
#
%package -n lib%{name}_tensorflow_lite_frontend
Summary: TensorFlow Lite frontend for OpenVINO toolkit
%description -n lib%{name}_tensorflow_lite_frontend
OpenVINO is an open-source toolkit for optimizing and deploying AI inference.
This package provides the tensorflow-lite frontend for OpenVINO.
## Samples/examples ##
#
%package -n %{name}-samples
Summary: Samples for use with OpenVINO toolkit
BuildArch: noarch
%description -n %{name}-samples
OpenVINO is an open-source toolkit for optimizing and deploying AI inference.
This package provides some samples for use with openVINO.
%global debug_package %{nil}
#
%prep
# download source0
# wget -qO %{SOURCE0} https://user-repo.openeuler.openatom.cn/lfs-tar/openvino/openvino-2024.3.0.tar.gz
%autosetup -p1
%build
# Otherwise intel_cpu plugin declares an executable stack
%ifarch x86_64
%define build_ldflags -Wl,-z,noexecstack
%endif
%cmake \
-DCMAKE_CXX_STANDARD=17 \
-DBUILD_SHARED_LIBS=ON \
-DENABLE_OV_ONNX_FRONTEND=ON \
-DENABLE_OV_PADDLE_FRONTEND=ON \
-DENABLE_OV_PYTORCH_FRONTEND=ON \
-DENABLE_OV_IR_FRONTEND=ON \
-DENABLE_OV_TF_FRONTEND=ON \
-DENABLE_OV_TF_LITE_FRONTEND=ON \
-DENABLE_INTEL_NPU=OFF \
-DENABLE_JS=OFF \
-DENABLE_SYSTEM_TBB=ON \
-DBUILD_TYPE=Release \
-DCPACK_GENERATOR=RPM \
%{nil}
%make_build
%install
%make_install DESTDIR=%{buildroot}
# Unnecessary if we get our package dependencies and lib paths right!
rm -fr %{buildroot}%{_prefix}/install_dependencies \
%{buildroot}%{_prefix}/setupvars.sh
%fdupes %{buildroot}%{_datadir}/
# We do not use bundled thirdparty libs
rm -rf %{buildroot}%{_datadir}/licenses/*
rm -rf %{buildroot}%{_datadir}/doc
%ldconfig_scriptlets -n %{shlib}
%ldconfig_scriptlets -n lib%{name}_ir_frontend
%ldconfig_scriptlets -n lib%{name}_onnx_frontend
%ldconfig_scriptlets -n lib%{name}_paddle_frontend
%ldconfig_scriptlets -n lib%{name}_pytorch_frontend
%ldconfig_scriptlets -n lib%{name}_tensorflow_lite_frontend
%ldconfig_scriptlets -n lib%{name}_tensorflow_frontend
%files -n %{shlib}
%license LICENSE
%{_libdir}/libopenvino.so.*
%{_libdir}/libopenvino_c.so.*
%files -n lib%{name}-auto-batch-plugin
%dir %{_libdir}/%{name}-%{version}
%{_libdir}/%{name}-%{version}/libopenvino_auto_batch_plugin.so
%files -n lib%{name}-auto-plugin
%dir %{_libdir}/%{name}-%{version}
%{_libdir}/%{name}-%{version}/libopenvino_auto_plugin.so
%ifarch x86_64
%files -n lib%{name}-intel-gpu-plugin
%dir %{_libdir}/%{name}-%{version}
%{_libdir}/%{name}-%{version}/libopenvino_intel_gpu_plugin.so
%{_libdir}/%{name}-%{version}/cache.json
%files -n lib%{name}-intel-cpu-plugin
%dir %{_libdir}/%{name}-%{version}
%{_libdir}/%{name}-%{version}/libopenvino_intel_cpu_plugin.so
%endif
%ifarch aarch64
%files -n lib%{name}-arm-cpu-plugin
%dir %{_libdir}/%{name}-%{version}
%{_libdir}/%{name}-%{version}/libopenvino_arm_cpu_plugin.so
%endif
%files -n lib%{name}-hetero-plugin
%dir %{_libdir}/%{name}-%{version}
%{_libdir}/%{name}-%{version}/libopenvino_hetero_plugin.so
%files -n lib%{name}_onnx_frontend
%{_libdir}/libopenvino_onnx_frontend.so.*
%files -n lib%{name}_ir_frontend
%{_libdir}/libopenvino_ir_frontend.so.*
%files -n lib%{name}_paddle_frontend
%{_libdir}/libopenvino_paddle_frontend.so.*
%files -n lib%{name}_pytorch_frontend
%{_libdir}/libopenvino_pytorch_frontend.so.*
%files -n lib%{name}_tensorflow_frontend
%{_libdir}/libopenvino_tensorflow_frontend.so.*
%files -n lib%{name}_tensorflow_lite_frontend
%{_libdir}/libopenvino_tensorflow_lite_frontend.so.*
%files -n %{name}-samples
%license LICENSE
%{_datadir}/%{name}/
%files -n lib%{name}-devel
%license LICENSE
%{_includedir}/%{name}/
%{_libdir}/cmake/%{name}%{version}/
%{_libdir}/*.so
%{_libdir}/pkgconfig/openvino.pc
%changelog
* Thu Nov 21 2024 zhouyi <zhouyi198@h-partners.com> - 2024.3.0-2
- Add lfs config
Edit openvino.spec
* Mon Sep 2 2024 Anokhov,Artyom <artyom.anokhov@intel.com> - 2024.3.0-1
- Initial openvino spec file
此处可能存在不合适展示的内容,页面不予展示。您可通过相关编辑功能自查并修改。
如您确认内容无涉及 不当用语 / 纯广告导流 / 暴力 / 低俗色情 / 侵权 / 盗版 / 虚假 / 无价值内容或违法国家有关法律法规的内容,可点击提交进行申诉,我们将尽快为您处理。