device.py 5.1 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151
  1. # copyright (c) 2024 PaddlePaddle Authors. All Rights Reserve.
  2. #
  3. # Licensed under the Apache License, Version 2.0 (the "License");
  4. # you may not use this file except in compliance with the License.
  5. # You may obtain a copy of the License at
  6. #
  7. # http://www.apache.org/licenses/LICENSE-2.0
  8. #
  9. # Unless required by applicable law or agreed to in writing, software
  10. # distributed under the License is distributed on an "AS IS" BASIS,
  11. # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  12. # See the License for the specific language governing permissions and
  13. # limitations under the License.
  14. import os
  15. import GPUtil
  16. import lazy_paddle as paddle
  17. from . import logging
  18. from .flags import DISABLE_DEV_MODEL_WL
  19. from .errors import raise_unsupported_device_error
  20. from .custom_device_whitelist import (
  21. DCU_WHITELIST,
  22. MLU_WHITELIST,
  23. NPU_WHITELIST,
  24. XPU_WHITELIST,
  25. GCU_WHITELIST,
  26. )
  27. SUPPORTED_DEVICE_TYPE = ["cpu", "gpu", "xpu", "npu", "mlu", "gcu", "dcu"]
  28. def _constr_device(device_type, device_ids):
  29. if device_ids:
  30. device_ids = ",".join(map(str, device_ids))
  31. return f"{device_type}:{device_ids}"
  32. else:
  33. return f"{device_type}"
  34. def get_default_device():
  35. avail_gpus = GPUtil.getAvailable()
  36. if not avail_gpus:
  37. # maybe edge devices like Jetson
  38. if os.path.exists("/etc/nv_tegra_release"):
  39. avail_gpus = [0]
  40. logging.info(
  41. "Detected that the current device is a Jetson edge device. The default behavior will be to use GPU: 0"
  42. )
  43. if not avail_gpus:
  44. return "cpu"
  45. else:
  46. return _constr_device("gpu", [avail_gpus[0]])
  47. def parse_device(device):
  48. """parse_device"""
  49. # According to https://www.paddlepaddle.org.cn/documentation/docs/zh/api/paddle/device/set_device_cn.html
  50. parts = device.split(":")
  51. if len(parts) > 2:
  52. raise ValueError(f"Invalid device: {device}")
  53. if len(parts) == 1:
  54. device_type, device_ids = parts[0], None
  55. else:
  56. device_type, device_ids = parts
  57. device_ids = device_ids.split(",")
  58. for device_id in device_ids:
  59. if not device_id.isdigit():
  60. raise ValueError(
  61. f"Device ID must be an integer. Invalid device ID: {device_id}"
  62. )
  63. device_ids = list(map(int, device_ids))
  64. device_type = device_type.lower()
  65. # raise_unsupported_device_error(device_type, SUPPORTED_DEVICE_TYPE)
  66. assert device_type.lower() in SUPPORTED_DEVICE_TYPE
  67. return device_type, device_ids
  68. def update_device_num(device, num):
  69. device_type, device_ids = parse_device(device)
  70. if device_ids:
  71. assert len(device_ids) >= num
  72. return _constr_device(device_type, device_ids[:num])
  73. else:
  74. return _constr_device(device_type, device_ids)
  75. def set_env_for_device(device):
  76. def _set(envs):
  77. for key, val in envs.items():
  78. os.environ[key] = val
  79. logging.debug(f"{key} has been set to {val}.")
  80. device_type, device_ids = parse_device(device)
  81. # XXX: is_compiled_with_rocm() must be True on dcu platform ?
  82. if device_type.lower() == "dcu" and paddle.is_compiled_with_rocm():
  83. envs = {"FLAGS_conv_workspace_size_limit": "2000"}
  84. _set(envs)
  85. if device_type.lower() == "npu":
  86. envs = {
  87. "FLAGS_npu_jit_compile": "0",
  88. "FLAGS_use_stride_kernel": "0",
  89. "FLAGS_allocator_strategy": "auto_growth",
  90. "CUSTOM_DEVICE_BLACK_LIST": "pad3d,pad3d_grad,set_value,set_value_with_tensor",
  91. "FLAGS_npu_scale_aclnn": "True",
  92. "FLAGS_npu_split_aclnn": "True",
  93. }
  94. _set(envs)
  95. if device_type.lower() == "xpu":
  96. envs = {
  97. "BKCL_FORCE_SYNC": "1",
  98. "BKCL_TIMEOUT": "1800",
  99. "FLAGS_use_stride_kernel": "0",
  100. "XPU_BLACK_LIST": "pad3d",
  101. }
  102. _set(envs)
  103. if device_type.lower() == "mlu":
  104. envs = {"FLAGS_use_stride_kernel": "0"}
  105. _set(envs)
  106. if device_type.lower() == "gcu":
  107. envs = {"FLAGS_use_stride_kernel": "0"}
  108. _set(envs)
  109. def check_supported_device(device, model_name):
  110. if DISABLE_DEV_MODEL_WL:
  111. logging.warning(
  112. "Skip checking if model is supported on device because the flag `PADDLE_PDX_DISABLE_DEV_MODEL_WL` has been set."
  113. )
  114. return
  115. device_type, device_ids = parse_device(device)
  116. if device_type == "dcu":
  117. assert (
  118. model_name in DCU_WHITELIST
  119. ), f"The DCU device does not yet support `{model_name}` model!"
  120. elif device_type == "mlu":
  121. assert (
  122. model_name in MLU_WHITELIST
  123. ), f"The MLU device does not yet support `{model_name}` model!"
  124. elif device_type == "npu":
  125. assert (
  126. model_name in NPU_WHITELIST
  127. ), f"The NPU device does not yet support `{model_name}` model!"
  128. elif device_type == "xpu":
  129. assert (
  130. model_name in XPU_WHITELIST
  131. ), f"The XPU device does not yet support `{model_name}` model!"
  132. elif device_type == "gcu":
  133. assert (
  134. model_name in GCU_WHITELIST
  135. ), f"The GCU device does not yet support `{model_name}` model!"