-
Notifications
You must be signed in to change notification settings - Fork 3.1k
Description
OpenVINO Version
2026.0.0
Operating System
WSL2 for Windows
Hardware Architecture
x86 (64 bits)
Target Platform
- OS: Windows (WSL2 with Ubuntu 24.04)
- Device: Intel Integrated GPU (iGPU)
- OpenVINO version: 2026.0.0
Build issue description
I am trying to run a PyTorch model using OpenVINO by converting it to IR format and compiling it on an Intel GPU.
However, I encounter a single allocation memory limit issue during compilation.
Build script or step-by-step to reproduce
config = {"GPU_ENABLE_LARGE_ALLOCATIONS": "YES"}
compiled = core.compile_model(ov_model_for_compile, "GPU", config)
Expected Behavior
The 1GB single memory allocation limit should be removed, so that allocations larger than 1GB (e.g., 1.2GB) can be handled and compiled without issues.
Using the GPU_ENABLE_LARGE_ALLOCATIONS environment variable or config option should allow large models to be compiled and run on Intel GPU without errors.
Current Behavior
When allocating more than 1GB, the following error occurs and compilation fails.
The same issue occurs even when using the GPU_ENABLE_LARGE_ALLOCATIONS environment variable or config option.
Relevant log output
Exception from src/inference/src/dev/plugin.cpp:53:
Check '!exceed_allocatable_mem_size' failed at src/plugins/intel_gpu/src/runtime/ocl/ocl_engine.cpp:112:
[GPU] Exceeded max size of memory object allocation: requested 1244659712 bytes, but max alloc size supported by device is 1073741824 bytes.Please try to reduce batch size or use lower precision.Issue submission checklist
- I'm reporting an issue. It's not a question.
- I checked the problem with the documentation, FAQ, open issues, Stack Overflow, etc., and have not found a solution.