You can not select more than 25 topics Topics must start with a chinese character,a letter or number, can include dashes ('-') and can be up to 35 characters long.

test_tensor.py 11 kB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384
  1. # -*- coding: utf-8 -*-
  2. # MegEngine is Licensed under the Apache License, Version 2.0 (the "License")
  3. #
  4. # Copyright (c) 2014-2020 Megvii Inc. All rights reserved.
  5. #
  6. # Unless required by applicable law or agreed to in writing,
  7. # software distributed under the License is distributed on an
  8. # "AS IS" BASIS, WITHOUT ARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  9. import platform
  10. import numpy as np
  11. import pytest
  12. import megengine.functional as F
  13. from megengine import tensor
  14. from megengine.core._trace_option import use_tensor_shape
  15. from megengine.core.tensor.utils import astensor1d
  16. from megengine.distributed.helper import get_device_count_by_fork
  17. from megengine.test import assertTensorClose
  18. def _default_compare_fn(x, y):
  19. assertTensorClose(x.numpy(), y)
  20. def opr_test(cases, func, compare_fn=_default_compare_fn, ref_fn=None, **kwargs):
  21. """
  22. func: the function to run opr.
  23. compare_fn: the function to compare the result and expected, use assertTensorClose if None.
  24. ref_fn: the function to generate expected data, should assign output if None.
  25. cases: the list which have dict element, the list length should be 2 for dynamic shape test.
  26. and the dict should have input,
  27. and should have output if ref_fn is None.
  28. should use list for multiple inputs and outputs for each case.
  29. kwargs: The additional kwargs for opr func.
  30. simple examples:
  31. dtype = np.float32
  32. cases = [{"input": [10, 20]}, {"input": [20, 30]}]
  33. opr_test(cases,
  34. F.eye,
  35. ref_fn=lambda n, m: np.eye(n, m).astype(dtype),
  36. dtype=dtype)
  37. """
  38. def check_results(results, expected):
  39. if not isinstance(results, tuple):
  40. results = (results,)
  41. for r, e in zip(results, expected):
  42. compare_fn(r, e)
  43. def get_param(cases, idx):
  44. case = cases[idx]
  45. inp = case.get("input", None)
  46. outp = case.get("output", None)
  47. if inp is None:
  48. raise ValueError("the test case should have input")
  49. if not isinstance(inp, list):
  50. inp = (inp,)
  51. else:
  52. inp = tuple(inp)
  53. if ref_fn is not None and callable(ref_fn):
  54. outp = ref_fn(*inp)
  55. if outp is None:
  56. raise ValueError("the test case should have output or reference function")
  57. if not isinstance(outp, list):
  58. outp = (outp,)
  59. else:
  60. outp = tuple(outp)
  61. return inp, outp
  62. if len(cases) == 0:
  63. raise ValueError("should give one case at least")
  64. if not callable(func):
  65. raise ValueError("the input func should be callable")
  66. inp, outp = get_param(cases, 0)
  67. inp_tensor = [tensor(inpi) for inpi in inp]
  68. results = func(*inp_tensor, **kwargs)
  69. check_results(results, outp)
  70. def test_eye():
  71. dtype = np.float32
  72. cases = [{"input": [10, 20]}, {"input": [20, 30]}]
  73. for case in cases:
  74. assertTensorClose(
  75. F.eye(case["input"], dtype=dtype).numpy(),
  76. np.eye(*case["input"]).astype(dtype),
  77. )
  78. def test_concat():
  79. def get_data_shape(length: int):
  80. return (length, 2, 3)
  81. data1 = np.random.random(get_data_shape(5)).astype("float32")
  82. data2 = np.random.random(get_data_shape(6)).astype("float32")
  83. data3 = np.random.random(get_data_shape(7)).astype("float32")
  84. def run(data1, data2):
  85. return F.concat([data1, data2])
  86. cases = [{"input": [data1, data2]}, {"input": [data1, data3]}]
  87. opr_test(cases, run, ref_fn=lambda x, y: np.concatenate([x, y]))
  88. def test_concat_device():
  89. data1 = tensor(np.random.random((3, 2, 2)).astype("float32"), device="cpu0")
  90. data2 = tensor(np.random.random((2, 2, 2)).astype("float32"), device="cpu1")
  91. out = F.concat([data1, data2], device="cpu0")
  92. assert str(out.device).split(":")[0] == "cpu0"
  93. def test_stack():
  94. data1 = np.random.random((3, 2, 2)).astype("float32")
  95. data2 = np.random.random((3, 2, 2)).astype("float32")
  96. data3 = np.random.random((3, 2, 2)).astype("float32")
  97. cases = [{"input": [data1, data2]}, {"input": [data1, data3]}]
  98. for ai in range(3):
  99. def run(data1, data2):
  100. return F.stack([data1, data2], axis=ai)
  101. opr_test(cases, run, ref_fn=lambda x, y: np.stack([x, y], axis=ai))
  102. def test_split():
  103. data = np.random.random((2, 3, 4, 5)).astype(np.float32)
  104. mge_out1 = F.split(tensor(data), 2, axis=3)
  105. mge_out2 = F.split(tensor(data), [3, 5], axis=3)
  106. np_out = np.split(data, [3, 5], axis=3)
  107. np.testing.assert_equal(mge_out1[0].numpy(), mge_out2[0].numpy())
  108. np.testing.assert_equal(mge_out1[0].numpy(), np_out[0])
  109. def test_reshape():
  110. x = np.arange(6, dtype="float32")
  111. xx = tensor(x)
  112. y = x.reshape(1, 2, 3)
  113. for shape in [
  114. (1, 2, 3),
  115. (1, -1, 3),
  116. (1, tensor(-1), 3),
  117. np.array([1, -1, 3], dtype="int32"),
  118. tensor([1, -1, 3]),
  119. ]:
  120. yy = F.reshape(xx, shape)
  121. np.testing.assert_equal(yy.numpy(), y)
  122. def test_squeeze():
  123. x = np.arange(6, dtype="float32").reshape(1, 2, 3, 1)
  124. xx = tensor(x)
  125. for axis in [None, 3, -4, (3, -4)]:
  126. y = np.squeeze(x, axis)
  127. yy = F.squeeze(xx, axis)
  128. np.testing.assert_equal(y, yy.numpy())
  129. def test_expand_dims():
  130. x = np.arange(6, dtype="float32").reshape(2, 3)
  131. xx = tensor(x)
  132. for axis in [2, -3, (3, -4), (1, -4)]:
  133. y = np.expand_dims(x, axis)
  134. yy = F.expand_dims(xx, axis)
  135. np.testing.assert_equal(y, yy.numpy())
  136. def test_elemwise_dtype_promotion():
  137. x = np.random.rand(2, 3).astype("float32")
  138. y = np.random.rand(1, 3).astype("float16")
  139. xx = tensor(x)
  140. yy = tensor(y)
  141. z = xx * yy
  142. np.testing.assert_equal(z.numpy(), x * y)
  143. z = xx + y
  144. np.testing.assert_equal(z.numpy(), x + y)
  145. z = x - yy
  146. np.testing.assert_equal(z.numpy(), x - y)
  147. def test_linspace():
  148. cases = [
  149. {"input": [1, 9, 9]},
  150. {"input": [3, 10, 8]},
  151. ]
  152. opr_test(
  153. cases,
  154. F.linspace,
  155. ref_fn=lambda start, end, step: np.linspace(start, end, step, dtype=np.float32),
  156. )
  157. cases = [
  158. {"input": [9, 1, 9]},
  159. {"input": [10, 3, 8]},
  160. ]
  161. opr_test(
  162. cases,
  163. F.linspace,
  164. ref_fn=lambda start, end, step: np.linspace(start, end, step, dtype=np.float32),
  165. )
  166. def test_arange():
  167. cases = [
  168. {"input": [1, 9, 1]},
  169. {"input": [2, 10, 2]},
  170. ]
  171. opr_test(
  172. cases,
  173. F.arange,
  174. ref_fn=lambda start, end, step: np.arange(start, end, step, dtype=np.float32),
  175. )
  176. cases = [
  177. {"input": [9, 1, -1]},
  178. {"input": [10, 2, -2]},
  179. ]
  180. opr_test(
  181. cases,
  182. F.arange,
  183. ref_fn=lambda start, end, step: np.arange(start, end, step, dtype=np.float32),
  184. )
  185. cases = [
  186. {"input": [9.3, 1.2, -0.5]},
  187. {"input": [10.3, 2.1, -1.7]},
  188. ]
  189. opr_test(
  190. cases,
  191. F.arange,
  192. ref_fn=lambda start, end, step: np.arange(start, end, step, dtype=np.float32),
  193. )
  194. def test_round():
  195. data1_shape = (15,)
  196. data2_shape = (25,)
  197. data1 = np.random.random(data1_shape).astype(np.float32)
  198. data2 = np.random.random(data2_shape).astype(np.float32)
  199. cases = [{"input": data1}, {"input": data2}]
  200. opr_test(cases, F.round, ref_fn=np.round)
  201. def test_broadcast():
  202. input1_shape = (20, 30)
  203. output1_shape = (30, 20, 30)
  204. data1 = np.random.random(input1_shape).astype(np.float32)
  205. input2_shape = (10, 20)
  206. output2_shape = (20, 10, 20)
  207. data2 = np.random.random(input2_shape).astype(np.float32)
  208. def compare_fn(x, y):
  209. assert x.numpy().shape == y
  210. cases = [
  211. {"input": [data1, output1_shape], "output": output1_shape},
  212. {"input": [data2, output2_shape], "output": output2_shape},
  213. ]
  214. opr_test(cases, F.broadcast, compare_fn=compare_fn)
  215. def test_utils_astensor1d():
  216. reference = tensor(0)
  217. # literal
  218. x = [1, 2, 3]
  219. for dtype in [None, "float32"]:
  220. xx = astensor1d(x, reference, dtype=dtype)
  221. assert type(xx) is tensor
  222. np.testing.assert_equal(xx.numpy(), x)
  223. # numpy array
  224. x = np.asarray([1, 2, 3], dtype="int32")
  225. for dtype in [None, "float32"]:
  226. xx = astensor1d(x, reference, dtype=dtype)
  227. assert type(xx) is tensor
  228. np.testing.assert_equal(xx.numpy(), x.astype(dtype) if dtype else x)
  229. # tensor
  230. x = tensor([1, 2, 3], dtype="int32")
  231. for dtype in [None, "float32"]:
  232. xx = astensor1d(x, reference, dtype=dtype)
  233. assert type(xx) is tensor
  234. np.testing.assert_equal(xx.numpy(), x.numpy())
  235. # mixed
  236. x = [1, tensor(2), 3]
  237. for dtype in [None, "float32"]:
  238. xx = astensor1d(x, reference, dtype=dtype)
  239. assert type(xx) is tensor
  240. np.testing.assert_equal(xx.numpy(), [1, 2, 3])
  241. def test_device():
  242. x = tensor([1, 2, 3], dtype="float32")
  243. y1 = F.eye(x.shape, dtype="float32")
  244. y2 = F.eye(x.shape, dtype="float32", device=None)
  245. np.testing.assert_almost_equal(y1.numpy(), y2.numpy())
  246. y3 = F.eye(x.shape, dtype="float32", device="xpux")
  247. y4 = F.eye(x.shape, dtype="float32", device=x.device.to_c())
  248. np.testing.assert_almost_equal(y3.numpy(), y4.numpy())
  249. y5 = F.full((3, 2), 4, device=x.device)
  250. y6 = F.full((3, 2), 4, device="xpux")
  251. np.testing.assert_almost_equal(y5.numpy(), y6.numpy())
  252. def copy_test(dst, src):
  253. data = np.random.random((2, 3)).astype(np.float32)
  254. x = tensor(data, device=src)
  255. y = F.copy(x, dst)
  256. assert np.allclose(data, y.numpy())
  257. z = x.to(dst)
  258. assert np.allclose(data, z.numpy())
  259. @pytest.mark.skipif(
  260. platform.system() == "Darwin", reason="do not imp GPU mode at macos now"
  261. )
  262. @pytest.mark.skipif(
  263. platform.system() == "Windows", reason="do not imp GPU mode at Windows now"
  264. )
  265. @pytest.mark.skipif(get_device_count_by_fork("gpu") == 0, reason="CUDA is disabled")
  266. def test_copy_h2d():
  267. copy_test("cpu0", "gpu0")
  268. @pytest.mark.skipif(
  269. platform.system() == "Darwin", reason="do not imp GPU mode at macos now"
  270. )
  271. @pytest.mark.skipif(
  272. platform.system() == "Windows", reason="do not imp GPU mode at Windows now"
  273. )
  274. @pytest.mark.skipif(get_device_count_by_fork("gpu") == 0, reason="CUDA is disabled")
  275. def test_copy_d2h():
  276. copy_test("gpu0", "cpu0")
  277. @pytest.mark.skipif(
  278. platform.system() == "Darwin", reason="do not imp GPU mode at macos now"
  279. )
  280. @pytest.mark.skipif(
  281. platform.system() == "Windows", reason="do not imp GPU mode at Windows now"
  282. )
  283. @pytest.mark.skipif(get_device_count_by_fork("gpu") < 2, reason="need more gpu device")
  284. def test_copy_d2d():
  285. copy_test("gpu0", "gpu1")
  286. copy_test("gpu0:0", "gpu0:1")
  287. def test_param_pack_split():
  288. a = tensor(np.ones((10,), np.int32))
  289. b, c = F.param_pack_split(a, [0, 1, 1, 10], [(1,), (3, 3)])
  290. assert np.allclose(b.numpy(), a.numpy()[1])
  291. assert np.allclose(c.numpy(), a.numpy()[1:].reshape(3, 3))
  292. def test_param_pack_concat():
  293. a = tensor(np.ones((1,), np.int32))
  294. b = tensor(np.ones((3, 3), np.int32))
  295. offsets_val = [0, 1, 1, 10]
  296. offsets = tensor(offsets_val, np.int32)
  297. c = F.param_pack_concat([a, b], offsets, offsets_val)
  298. assert np.allclose(np.concatenate([a.numpy(), b.numpy().flatten()]), c.numpy())

MegEngine 安装包中集成了使用 GPU 运行代码所需的 CUDA 环境,不用区分 CPU 和 GPU 版。 如果想要运行 GPU 程序,请确保机器本身配有 GPU 硬件设备并安装好驱动。 如果你想体验在云端 GPU 算力平台进行深度学习开发的感觉,欢迎访问 MegStudio 平台