core.inference.unified_memory#
Module Contents#
Classes#
Enum to distinguish between unified memory (UVM) compilation states. |
Functions#
Context manager to timeout compilation. |
|
Attempt to compile UVM allocator. |
|
Create a unified memory mempool using CUDA managed memory. |
Data#
API#
- class core.inference.unified_memory.CompilationState(*args, **kwds)#
Bases:
enum.EnumEnum to distinguish between unified memory (UVM) compilation states.
Initialization
- UNATTEMPTED#
‘auto(…)’
- FAILURE#
‘auto(…)’
- SUCCESS#
‘auto(…)’
- exception core.inference.unified_memory.UnifiedMemoryUnsupportedError#
Bases:
ExceptionUnified memory is not supported on this system.
Initialization
Initialize self. See help(type(self)) for accurate signature.
- exception core.inference.unified_memory.UnifiedMemoryCompileTimeoutError#
Bases:
core.inference.unified_memory.UnifiedMemoryUnsupportedErrorUnified memory compilation timed out.
Initialization
Initialize self. See help(type(self)) for accurate signature.
- core.inference.unified_memory._compilation_state#
None
- core.inference.unified_memory._alloc#
None
- core.inference.unified_memory._mod#
None
- core.inference.unified_memory._compile_timeout(timeout_s: int)#
Context manager to timeout compilation.
- Parameters:
timeout_s (int) – Timeout in seconds.
- core.inference.unified_memory.compile_allocator()#
Attempt to compile UVM allocator.
- core.inference.unified_memory.create_unified_mempool() torch.cuda.memory.MemPool#
Create a unified memory mempool using CUDA managed memory.
- Returns:
(MemPool) Unified memory mempool.