open-gpu-share
mainly defines the data structure of a scheduler cache. It is used to extend the scheduler, under Kubernetes scheduling framework, to allocate pods that can share a GPU, i.e., apply for a portion of full GPU memory.
This project heavily relies on AliyunContainerService/gpushare-scheduler-extender, referring to Nvidia Docker2 and their GPU sharing design.