| @@ -67,7 +67,7 @@ namespace Tensorflow | |||
| /// <param name="status">TF_Status*</param> | |||
| /// <returns>TFE_TensorHandle*</returns> | |||
| [DllImport(TensorFlowLibName)] | |||
| public static extern IntPtr TFE_TensorHandleCopyToDevice(IntPtr h, SafeContextHandle ctx, string device_name, SafeStatusHandle status); | |||
| public static extern SafeTensorHandleHandle TFE_TensorHandleCopyToDevice(SafeTensorHandleHandle h, SafeContextHandle ctx, string device_name, SafeStatusHandle status); | |||
| /// <summary> | |||
| /// Retrieves the full name of the device (e.g. /job:worker/replica:0/...) | |||
| @@ -33,7 +33,7 @@ namespace Tensorflow.Eager | |||
| { | |||
| for (int i = 0; i < inputs.Length; ++i) | |||
| { | |||
| IntPtr tensor_handle; | |||
| SafeTensorHandleHandle tensor_handle; | |||
| switch (inputs[i]) | |||
| { | |||
| case EagerTensor et: | |||
| @@ -50,10 +50,10 @@ namespace Tensorflow.Eager | |||
| if (status.ok() && attrs != null) | |||
| SetOpAttrs(op, attrs); | |||
| var outputs = new IntPtr[num_outputs]; | |||
| var outputs = new SafeTensorHandleHandle[num_outputs]; | |||
| if (status.ok()) | |||
| { | |||
| c_api.TFE_Execute(op, outputs, ref num_outputs, status.Handle); | |||
| c_api.TFE_Execute(op, outputs, out num_outputs, status.Handle); | |||
| status.Check(true); | |||
| } | |||
| return outputs.Select(x => new EagerTensor(x)).ToArray(); | |||
| @@ -154,8 +154,8 @@ namespace Tensorflow.Eager | |||
| num_retvals += (int)delta; | |||
| } | |||
| var retVals = new IntPtr[num_retvals]; | |||
| c_api.TFE_Execute(op, retVals, ref num_retvals, status.Handle); | |||
| var retVals = new SafeTensorHandleHandle[num_retvals]; | |||
| c_api.TFE_Execute(op, retVals, out num_retvals, status.Handle); | |||
| status.Check(true); | |||
| var flat_result = retVals.Select(x => new EagerTensor(x)).ToArray(); | |||
| @@ -220,7 +220,7 @@ namespace Tensorflow.Eager | |||
| SafeOpHandle op, | |||
| Status status) | |||
| { | |||
| IntPtr input_handle; | |||
| SafeTensorHandleHandle input_handle; | |||
| // ConvertToTensor(); | |||
| switch (inputs) | |||
| @@ -14,7 +14,7 @@ namespace Tensorflow.Eager | |||
| } | |||
| public EagerTensor(IntPtr handle) : base(IntPtr.Zero) | |||
| public EagerTensor(SafeTensorHandleHandle handle) : base(IntPtr.Zero) | |||
| { | |||
| EagerTensorHandle = handle; | |||
| Resolve(); | |||
| @@ -58,14 +58,20 @@ namespace Tensorflow.Eager | |||
| } | |||
| public override IntPtr ToPointer() | |||
| => EagerTensorHandle; | |||
| => EagerTensorHandle?.DangerousGetHandle() ?? IntPtr.Zero; | |||
| protected override void DisposeManagedResources() | |||
| { | |||
| base.DisposeManagedResources(); | |||
| //print($"deleting DeleteTensorHandle {Id} {EagerTensorHandle.ToString("x16")}"); | |||
| EagerTensorHandle.Dispose(); | |||
| } | |||
| protected override void DisposeUnmanagedResources(IntPtr handle) | |||
| { | |||
| //print($"deleting DeleteTensorHandle {Id} {_handle.ToString("x16")}"); | |||
| c_api.TF_DeleteTensor(_handle); | |||
| //print($"deleting DeleteTensorHandle {Id} {EagerTensorHandle.ToString("x16")}"); | |||
| c_api.TFE_DeleteTensorHandle(EagerTensorHandle); | |||
| } | |||
| } | |||
| } | |||
| @@ -8,7 +8,8 @@ namespace Tensorflow.Eager | |||
| { | |||
| public partial class EagerTensor | |||
| { | |||
| [Obsolete("Implicit conversion of EagerTensor to IntPtr is not supported.", error: true)] | |||
| public static implicit operator IntPtr(EagerTensor tensor) | |||
| => tensor.EagerTensorHandle; | |||
| => throw new NotSupportedException(); | |||
| } | |||
| } | |||
| @@ -0,0 +1,40 @@ | |||
| /***************************************************************************** | |||
| Copyright 2018 The TensorFlow.NET Authors. All Rights Reserved. | |||
| Licensed under the Apache License, Version 2.0 (the "License"); | |||
| you may not use this file except in compliance with the License. | |||
| You may obtain a copy of the License at | |||
| http://www.apache.org/licenses/LICENSE-2.0 | |||
| Unless required by applicable law or agreed to in writing, software | |||
| distributed under the License is distributed on an "AS IS" BASIS, | |||
| WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
| See the License for the specific language governing permissions and | |||
| limitations under the License. | |||
| ******************************************************************************/ | |||
| using System; | |||
| using Tensorflow.Util; | |||
| namespace Tensorflow.Eager | |||
| { | |||
| public sealed class SafeTensorHandleHandle : SafeTensorflowHandle | |||
| { | |||
| private SafeTensorHandleHandle() | |||
| { | |||
| } | |||
| public SafeTensorHandleHandle(IntPtr handle) | |||
| : base(handle) | |||
| { | |||
| } | |||
| protected override bool ReleaseHandle() | |||
| { | |||
| c_api.TFE_DeleteTensorHandle(handle); | |||
| SetHandle(IntPtr.Zero); | |||
| return true; | |||
| } | |||
| } | |||
| } | |||
| @@ -1,19 +0,0 @@ | |||
| using System; | |||
| using System.Collections.Generic; | |||
| using System.Runtime.InteropServices; | |||
| using System.Text; | |||
| namespace Tensorflow.Eager | |||
| { | |||
| [StructLayout(LayoutKind.Sequential)] | |||
| public struct TFE_TensorHandle | |||
| { | |||
| IntPtr _handle; | |||
| public static implicit operator IntPtr(TFE_TensorHandle tensor) | |||
| => tensor._handle; | |||
| public override string ToString() | |||
| => $"TFE_TensorHandle 0x{_handle.ToString("x16")}"; | |||
| } | |||
| } | |||
| @@ -3,6 +3,7 @@ using System; | |||
| using System.Runtime.InteropServices; | |||
| using Tensorflow.Device; | |||
| using Tensorflow.Eager; | |||
| using Tensorflow.Util; | |||
| namespace Tensorflow | |||
| { | |||
| @@ -66,7 +67,7 @@ namespace Tensorflow | |||
| /// <param name="status">TF_Status*</param> | |||
| /// <returns></returns> | |||
| [DllImport(TensorFlowLibName)] | |||
| public static extern int TFE_OpAddInputList(SafeOpHandle op, IntPtr[] inputs, int num_inputs, SafeStatusHandle status); | |||
| public static extern int TFE_OpAddInputList(SafeOpHandle op, [In, MarshalAs(UnmanagedType.CustomMarshaler, MarshalTypeRef = typeof(SafeHandleArrayMarshaler))] SafeTensorHandleHandle[] inputs, int num_inputs, SafeStatusHandle status); | |||
| /// <summary> | |||
| /// | |||
| @@ -90,6 +91,20 @@ namespace Tensorflow | |||
| [DllImport(TensorFlowLibName)] | |||
| public static extern void TFE_DeleteContext(IntPtr ctx); | |||
| public static void TFE_Execute(SafeOpHandle op, SafeTensorHandleHandle[] retvals, out int num_retvals, SafeStatusHandle status) | |||
| { | |||
| unsafe | |||
| { | |||
| num_retvals = retvals?.Length ?? 0; | |||
| var rawReturns = stackalloc IntPtr[num_retvals]; | |||
| TFE_Execute(op, rawReturns, ref num_retvals, status); | |||
| for (var i = 0; i < num_retvals; i++) | |||
| { | |||
| retvals[i] = new SafeTensorHandleHandle(rawReturns[i]); | |||
| } | |||
| } | |||
| } | |||
| /// <summary> | |||
| /// Execute the operation defined by 'op' and return handles to computed | |||
| /// tensors in `retvals`. | |||
| @@ -99,7 +114,7 @@ namespace Tensorflow | |||
| /// <param name="num_retvals">int*</param> | |||
| /// <param name="status">TF_Status*</param> | |||
| [DllImport(TensorFlowLibName)] | |||
| public static extern void TFE_Execute(SafeOpHandle op, IntPtr[] retvals, ref int num_retvals, SafeStatusHandle status); | |||
| private static unsafe extern void TFE_Execute(SafeOpHandle op, IntPtr* retvals, ref int num_retvals, SafeStatusHandle status); | |||
| /// <summary> | |||
| /// | |||
| @@ -198,7 +213,7 @@ namespace Tensorflow | |||
| /// <param name="h">TFE_TensorHandle*</param> | |||
| /// <param name="status">TF_Status*</param> | |||
| [DllImport(TensorFlowLibName)] | |||
| public static extern void TFE_OpAddInput(SafeOpHandle op, IntPtr h, SafeStatusHandle status); | |||
| public static extern void TFE_OpAddInput(SafeOpHandle op, SafeTensorHandleHandle h, SafeStatusHandle status); | |||
| /// <summary> | |||
| /// | |||
| @@ -206,10 +221,10 @@ namespace Tensorflow | |||
| /// <param name="t">const tensorflow::Tensor&</param> | |||
| /// <returns>TFE_TensorHandle*</returns> | |||
| [DllImport(TensorFlowLibName)] | |||
| public static extern TFE_TensorHandle TFE_NewTensorHandle(IntPtr t, SafeStatusHandle status); | |||
| public static extern SafeTensorHandleHandle TFE_NewTensorHandle(IntPtr t, SafeStatusHandle status); | |||
| [DllImport(TensorFlowLibName)] | |||
| public static extern IntPtr TFE_EagerTensorHandle(IntPtr t); | |||
| public static extern SafeTensorHandleHandle TFE_EagerTensorHandle(IntPtr t); | |||
| /// <summary> | |||
| /// Sets the default execution mode (sync/async). Note that this can be | |||
| @@ -226,7 +241,7 @@ namespace Tensorflow | |||
| /// <param name="h">TFE_TensorHandle*</param> | |||
| /// <returns></returns> | |||
| [DllImport(TensorFlowLibName)] | |||
| public static extern TF_DataType TFE_TensorHandleDataType(IntPtr h); | |||
| public static extern TF_DataType TFE_TensorHandleDataType(SafeTensorHandleHandle h); | |||
| /// <summary> | |||
| /// This function will block till the operation that produces `h` has | |||
| @@ -237,7 +252,7 @@ namespace Tensorflow | |||
| /// <param name="status">TF_Status*</param> | |||
| /// <returns></returns> | |||
| [DllImport(TensorFlowLibName)] | |||
| public static extern IntPtr TFE_TensorHandleResolve(IntPtr h, SafeStatusHandle status); | |||
| public static extern IntPtr TFE_TensorHandleResolve(SafeTensorHandleHandle h, SafeStatusHandle status); | |||
| /// <summary> | |||
| @@ -247,10 +262,10 @@ namespace Tensorflow | |||
| /// <param name="status">TF_Status*</param> | |||
| /// <returns></returns> | |||
| [DllImport(TensorFlowLibName)] | |||
| public static extern int TFE_TensorHandleNumDims(IntPtr h, SafeStatusHandle status); | |||
| public static extern int TFE_TensorHandleNumDims(SafeTensorHandleHandle h, SafeStatusHandle status); | |||
| [DllImport(TensorFlowLibName)] | |||
| public static extern int TFE_TensorHandleDim(IntPtr h, int dim, SafeStatusHandle status); | |||
| public static extern int TFE_TensorHandleDim(SafeTensorHandleHandle h, int dim, SafeStatusHandle status); | |||
| /// <summary> | |||
| /// Returns the device of the operation that produced `h`. If `h` was produced by | |||
| @@ -263,7 +278,7 @@ namespace Tensorflow | |||
| /// <param name="status">TF_Status*</param> | |||
| /// <returns></returns> | |||
| [DllImport(TensorFlowLibName)] | |||
| public static extern IntPtr TFE_TensorHandleDeviceName(IntPtr h, SafeStatusHandle status); | |||
| public static extern IntPtr TFE_TensorHandleDeviceName(SafeTensorHandleHandle h, SafeStatusHandle status); | |||
| /// <summary> | |||
| /// Returns the name of the device in whose memory `h` resides. | |||
| @@ -272,7 +287,7 @@ namespace Tensorflow | |||
| /// <param name="status">TF_Status*</param> | |||
| /// <returns></returns> | |||
| [DllImport(TensorFlowLibName)] | |||
| public static extern IntPtr TFE_TensorHandleBackingDeviceName(IntPtr h, SafeStatusHandle status); | |||
| public static extern IntPtr TFE_TensorHandleBackingDeviceName(SafeTensorHandleHandle h, SafeStatusHandle status); | |||
| /// <summary> | |||
| /// | |||
| @@ -12,7 +12,7 @@ namespace Tensorflow | |||
| { | |||
| public class EagerTensorV2 : DisposableObject, ITensor | |||
| { | |||
| IntPtr EagerTensorHandle; | |||
| SafeTensorHandleHandle EagerTensorHandle; | |||
| public string Device => c_api.StringPiece(c_api.TFE_TensorHandleDeviceName(EagerTensorHandle, tf.status.Handle)); | |||
| public EagerTensorV2(IntPtr handle) | |||
| @@ -64,10 +64,14 @@ namespace Tensorflow | |||
| } | |||
| }*/ | |||
| protected override void DisposeManagedResources() | |||
| { | |||
| EagerTensorHandle.Dispose(); | |||
| } | |||
| protected override void DisposeUnmanagedResources(IntPtr handle) | |||
| { | |||
| c_api.TF_DeleteTensor(_handle); | |||
| c_api.TFE_DeleteTensorHandle(EagerTensorHandle); | |||
| } | |||
| } | |||
| } | |||
| @@ -23,6 +23,7 @@ using System.Linq; | |||
| using System.Runtime.InteropServices; | |||
| using System.Text; | |||
| using static Tensorflow.Binding; | |||
| using Tensorflow.Eager; | |||
| using Tensorflow.Framework; | |||
| namespace Tensorflow | |||
| @@ -94,7 +95,7 @@ namespace Tensorflow | |||
| /// <summary> | |||
| /// TFE_TensorHandle | |||
| /// </summary> | |||
| public IntPtr EagerTensorHandle { get; set; } | |||
| public SafeTensorHandleHandle EagerTensorHandle { get; set; } | |||
| /// <summary> | |||
| /// Returns the shape of a tensor. | |||
| @@ -0,0 +1,132 @@ | |||
| /***************************************************************************** | |||
| Copyright 2018 The TensorFlow.NET Authors. All Rights Reserved. | |||
| Licensed under the Apache License, Version 2.0 (the "License"); | |||
| you may not use this file except in compliance with the License. | |||
| You may obtain a copy of the License at | |||
| http://www.apache.org/licenses/LICENSE-2.0 | |||
| Unless required by applicable law or agreed to in writing, software | |||
| distributed under the License is distributed on an "AS IS" BASIS, | |||
| WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
| See the License for the specific language governing permissions and | |||
| limitations under the License. | |||
| ******************************************************************************/ | |||
| using System; | |||
| using System.Runtime.ExceptionServices; | |||
| using System.Runtime.InteropServices; | |||
| namespace Tensorflow.Util | |||
| { | |||
| internal sealed class SafeHandleArrayMarshaler : ICustomMarshaler | |||
| { | |||
| private static readonly SafeHandleArrayMarshaler Instance = new SafeHandleArrayMarshaler(); | |||
| private SafeHandleArrayMarshaler() | |||
| { | |||
| } | |||
| #pragma warning disable IDE0060 // Remove unused parameter (method is used implicitly) | |||
| public static ICustomMarshaler GetInstance(string cookie) | |||
| #pragma warning restore IDE0060 // Remove unused parameter | |||
| { | |||
| return Instance; | |||
| } | |||
| public int GetNativeDataSize() | |||
| { | |||
| return IntPtr.Size; | |||
| } | |||
| [HandleProcessCorruptedStateExceptions] | |||
| public IntPtr MarshalManagedToNative(object ManagedObj) | |||
| { | |||
| if (ManagedObj is null) | |||
| return IntPtr.Zero; | |||
| var array = (SafeHandle[])ManagedObj; | |||
| var native = IntPtr.Zero; | |||
| var marshaledArrayHandle = false; | |||
| try | |||
| { | |||
| native = Marshal.AllocHGlobal((array.Length + 1) * IntPtr.Size); | |||
| Marshal.WriteIntPtr(native, GCHandle.ToIntPtr(GCHandle.Alloc(array))); | |||
| marshaledArrayHandle = true; | |||
| var i = 0; | |||
| var success = false; | |||
| try | |||
| { | |||
| for (i = 0; i < array.Length; i++) | |||
| { | |||
| success = false; | |||
| var current = array[i]; | |||
| var currentHandle = IntPtr.Zero; | |||
| if (current is object) | |||
| { | |||
| current.DangerousAddRef(ref success); | |||
| currentHandle = current.DangerousGetHandle(); | |||
| } | |||
| Marshal.WriteIntPtr(native, ofs: (i + 1) * IntPtr.Size, currentHandle); | |||
| } | |||
| return IntPtr.Add(native, IntPtr.Size); | |||
| } | |||
| catch | |||
| { | |||
| // Clean up any handles which were leased prior to the exception | |||
| var total = success ? i + 1 : i; | |||
| for (var j = 0; j < total; j++) | |||
| { | |||
| var current = array[i]; | |||
| if (current is object) | |||
| current.DangerousRelease(); | |||
| } | |||
| throw; | |||
| } | |||
| } | |||
| catch | |||
| { | |||
| if (native != IntPtr.Zero) | |||
| { | |||
| if (marshaledArrayHandle) | |||
| GCHandle.FromIntPtr(Marshal.ReadIntPtr(native)).Free(); | |||
| Marshal.FreeHGlobal(native); | |||
| } | |||
| throw; | |||
| } | |||
| } | |||
| public void CleanUpNativeData(IntPtr pNativeData) | |||
| { | |||
| if (pNativeData == IntPtr.Zero) | |||
| return; | |||
| var managedHandle = GCHandle.FromIntPtr(Marshal.ReadIntPtr(pNativeData, -IntPtr.Size)); | |||
| var array = (SafeHandle[])managedHandle.Target; | |||
| managedHandle.Free(); | |||
| for (var i = 0; i < array.Length; i++) | |||
| { | |||
| if (array[i] is object && !array[i].IsClosed) | |||
| array[i].DangerousRelease(); | |||
| } | |||
| } | |||
| public object MarshalNativeToManaged(IntPtr pNativeData) | |||
| { | |||
| throw new NotSupportedException(); | |||
| } | |||
| public void CleanUpManagedData(object ManagedObj) | |||
| { | |||
| throw new NotSupportedException(); | |||
| } | |||
| } | |||
| } | |||
| @@ -56,10 +56,10 @@ namespace TensorFlowNET.UnitTest | |||
| protected void TF_SetAttrBool(OperationDescription desc, string attrName, bool value) | |||
| => c_api.TF_SetAttrBool(desc, attrName, value); | |||
| protected TF_DataType TFE_TensorHandleDataType(IntPtr h) | |||
| protected TF_DataType TFE_TensorHandleDataType(SafeTensorHandleHandle h) | |||
| => c_api.TFE_TensorHandleDataType(h); | |||
| protected int TFE_TensorHandleNumDims(IntPtr h, SafeStatusHandle status) | |||
| protected int TFE_TensorHandleNumDims(SafeTensorHandleHandle h, SafeStatusHandle status) | |||
| => c_api.TFE_TensorHandleNumDims(h, status); | |||
| protected TF_Code TF_GetCode(Status s) | |||
| @@ -83,7 +83,7 @@ namespace TensorFlowNET.UnitTest | |||
| protected ulong TF_TensorByteSize(IntPtr t) | |||
| => c_api.TF_TensorByteSize(t); | |||
| protected void TFE_OpAddInput(SafeOpHandle op, IntPtr h, SafeStatusHandle status) | |||
| protected void TFE_OpAddInput(SafeOpHandle op, SafeTensorHandleHandle h, SafeStatusHandle status) | |||
| => c_api.TFE_OpAddInput(op, h, status); | |||
| protected void TFE_OpSetAttrType(SafeOpHandle op, string attr_name, TF_DataType value) | |||
| @@ -98,11 +98,11 @@ namespace TensorFlowNET.UnitTest | |||
| protected SafeOpHandle TFE_NewOp(SafeContextHandle ctx, string op_or_function_name, SafeStatusHandle status) | |||
| => c_api.TFE_NewOp(ctx, op_or_function_name, status); | |||
| protected IntPtr TFE_NewTensorHandle(IntPtr t, SafeStatusHandle status) | |||
| protected SafeTensorHandleHandle TFE_NewTensorHandle(IntPtr t, SafeStatusHandle status) | |||
| => c_api.TFE_NewTensorHandle(t, status); | |||
| protected void TFE_Execute(SafeOpHandle op, IntPtr[] retvals, ref int num_retvals, SafeStatusHandle status) | |||
| => c_api.TFE_Execute(op, retvals, ref num_retvals, status); | |||
| protected void TFE_Execute(SafeOpHandle op, SafeTensorHandleHandle[] retvals, out int num_retvals, SafeStatusHandle status) | |||
| => c_api.TFE_Execute(op, retvals, out num_retvals, status); | |||
| protected SafeContextOptionsHandle TFE_NewContextOptions() | |||
| => c_api.TFE_NewContextOptions(); | |||
| @@ -113,7 +113,7 @@ namespace TensorFlowNET.UnitTest | |||
| protected int TFE_OpGetInputLength(SafeOpHandle op, string input_name, SafeStatusHandle status) | |||
| => c_api.TFE_OpGetInputLength(op, input_name, status); | |||
| protected int TFE_OpAddInputList(SafeOpHandle op, IntPtr[] inputs, int num_inputs, SafeStatusHandle status) | |||
| protected int TFE_OpAddInputList(SafeOpHandle op, SafeTensorHandleHandle[] inputs, int num_inputs, SafeStatusHandle status) | |||
| => c_api.TFE_OpAddInputList(op, inputs, num_inputs, status); | |||
| protected int TFE_OpGetOutputLength(SafeOpHandle op, string input_name, SafeStatusHandle status) | |||
| @@ -128,13 +128,13 @@ namespace TensorFlowNET.UnitTest | |||
| protected void TFE_ExecutorWaitForAllPendingNodes(SafeExecutorHandle executor, SafeStatusHandle status) | |||
| => c_api.TFE_ExecutorWaitForAllPendingNodes(executor, status); | |||
| protected IntPtr TFE_TensorHandleResolve(IntPtr h, SafeStatusHandle status) | |||
| protected IntPtr TFE_TensorHandleResolve(SafeTensorHandleHandle h, SafeStatusHandle status) | |||
| => c_api.TFE_TensorHandleResolve(h, status); | |||
| protected string TFE_TensorHandleDeviceName(IntPtr h, SafeStatusHandle status) | |||
| protected string TFE_TensorHandleDeviceName(SafeTensorHandleHandle h, SafeStatusHandle status) | |||
| => c_api.StringPiece(c_api.TFE_TensorHandleDeviceName(h, status)); | |||
| protected string TFE_TensorHandleBackingDeviceName(IntPtr h, SafeStatusHandle status) | |||
| protected string TFE_TensorHandleBackingDeviceName(SafeTensorHandleHandle h, SafeStatusHandle status) | |||
| => c_api.StringPiece(c_api.TFE_TensorHandleBackingDeviceName(h, status)); | |||
| protected SafeDeviceListHandle TFE_ContextListDevices(SafeContextHandle ctx, SafeStatusHandle status) | |||
| @@ -149,7 +149,7 @@ namespace TensorFlowNET.UnitTest | |||
| protected string TF_DeviceListName(SafeDeviceListHandle list, int index, SafeStatusHandle status) | |||
| => c_api.TF_DeviceListName(list, index, status); | |||
| protected IntPtr TFE_TensorHandleCopyToDevice(IntPtr h, SafeContextHandle ctx, string device_name, SafeStatusHandle status) | |||
| protected SafeTensorHandleHandle TFE_TensorHandleCopyToDevice(SafeTensorHandleHandle h, SafeContextHandle ctx, string device_name, SafeStatusHandle status) | |||
| => c_api.TFE_TensorHandleCopyToDevice(h, ctx, device_name, status); | |||
| protected void TFE_OpSetDevice(SafeOpHandle op, string device_name, SafeStatusHandle status) | |||
| @@ -33,21 +33,25 @@ namespace TensorFlowNET.UnitTest.NativeAPI | |||
| { | |||
| CHECK_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| var m = TestMatrixTensorHandle(); | |||
| var retvals = new IntPtr[] { IntPtr.Zero, IntPtr.Zero }; | |||
| using (var matmul = MatMulOp(ctx, m, m)) | |||
| var retvals = new SafeTensorHandleHandle[2]; | |||
| try | |||
| { | |||
| int num_retvals = 2; | |||
| c_api.TFE_Execute(matmul, retvals, ref num_retvals, status); | |||
| EXPECT_EQ(1, num_retvals); | |||
| EXPECT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| } | |||
| TFE_DeleteTensorHandle(m); | |||
| using (var m = TestMatrixTensorHandle()) | |||
| using (var matmul = MatMulOp(ctx, m, m)) | |||
| { | |||
| int num_retvals; | |||
| c_api.TFE_Execute(matmul, retvals, out num_retvals, status); | |||
| EXPECT_EQ(1, num_retvals); | |||
| EXPECT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| } | |||
| t = TFE_TensorHandleResolve(retvals[0], status); | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| TFE_DeleteTensorHandle(retvals[0]); | |||
| t = TFE_TensorHandleResolve(retvals[0], status); | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| } | |||
| finally | |||
| { | |||
| retvals[0]?.Dispose(); | |||
| } | |||
| } | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| @@ -24,9 +24,10 @@ namespace TensorFlowNET.UnitTest.NativeAPI | |||
| using var ctx = NewContext(status); | |||
| CHECK_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| var input1 = TestMatrixTensorHandle(); | |||
| var input2 = TestMatrixTensorHandle(); | |||
| var retvals = new IntPtr[2]; | |||
| using var input1 = TestMatrixTensorHandle(); | |||
| using var input2 = TestMatrixTensorHandle(); | |||
| var retvals = new SafeTensorHandleHandle[2]; | |||
| using (var identityOp = TFE_NewOp(ctx, "IdentityN", status)) | |||
| { | |||
| CHECK_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| @@ -37,7 +38,7 @@ namespace TensorFlowNET.UnitTest.NativeAPI | |||
| EXPECT_EQ(-1, TFE_OpGetOutputLength(identityOp, "output", status)); | |||
| CHECK_NE(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| var inputs = new IntPtr[] { input1, input2 }; | |||
| var inputs = new SafeTensorHandleHandle[] { input1, input2 }; | |||
| TFE_OpAddInputList(identityOp, inputs, 2, status); | |||
| CHECK_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| @@ -47,21 +48,24 @@ namespace TensorFlowNET.UnitTest.NativeAPI | |||
| EXPECT_EQ(2, TFE_OpGetOutputLength(identityOp, "output", status)); | |||
| CHECK_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| int num_retvals = 2; | |||
| TFE_Execute(identityOp, retvals, ref num_retvals, status); | |||
| int num_retvals; | |||
| TFE_Execute(identityOp, retvals, out num_retvals, status); | |||
| EXPECT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| // Try to retrieve lengths after executing the op (should work) | |||
| EXPECT_EQ(2, TFE_OpGetInputLength(identityOp, "input", status)); | |||
| CHECK_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| EXPECT_EQ(2, TFE_OpGetOutputLength(identityOp, "output", status)); | |||
| CHECK_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| try | |||
| { | |||
| // Try to retrieve lengths after executing the op (should work) | |||
| EXPECT_EQ(2, TFE_OpGetInputLength(identityOp, "input", status)); | |||
| CHECK_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| EXPECT_EQ(2, TFE_OpGetOutputLength(identityOp, "output", status)); | |||
| CHECK_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| } | |||
| finally | |||
| { | |||
| retvals[0]?.Dispose(); | |||
| retvals[1]?.Dispose(); | |||
| } | |||
| } | |||
| TFE_DeleteTensorHandle(input1); | |||
| TFE_DeleteTensorHandle(input2); | |||
| TFE_DeleteTensorHandle(retvals[0]); | |||
| TFE_DeleteTensorHandle(retvals[1]); | |||
| } | |||
| } | |||
| } | |||
| @@ -1,5 +1,4 @@ | |||
| using Microsoft.VisualStudio.TestTools.UnitTesting; | |||
| using System; | |||
| using Tensorflow; | |||
| using Tensorflow.Eager; | |||
| @@ -24,10 +23,9 @@ namespace TensorFlowNET.UnitTest.NativeAPI | |||
| using var ctx = NewContext(status); | |||
| CHECK_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| var condition = TestScalarTensorHandle(true); | |||
| var t1 = TestMatrixTensorHandle(); | |||
| var t2 = TestAxisTensorHandle(); | |||
| var retvals = new IntPtr[1]; | |||
| using var condition = TestScalarTensorHandle(true); | |||
| using var t1 = TestMatrixTensorHandle(); | |||
| using var t2 = TestAxisTensorHandle(); | |||
| using (var assertOp = TFE_NewOp(ctx, "Assert", status)) | |||
| { | |||
| CHECK_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| @@ -44,15 +42,13 @@ namespace TensorFlowNET.UnitTest.NativeAPI | |||
| //EXPECT_EQ(attr_found->second.list().type(1), tensorflow::DataType::DT_FLOAT); | |||
| //EXPECT_EQ(attr_found->second.list().type(2), tensorflow::DataType::DT_INT32); | |||
| int num_retvals = 1; | |||
| TFE_Execute(assertOp, retvals, ref num_retvals, status); | |||
| var retvals = new SafeTensorHandleHandle[1]; | |||
| int num_retvals; | |||
| TFE_Execute(assertOp, retvals, out num_retvals, status); | |||
| EXPECT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| } | |||
| TFE_DeleteTensorHandle(condition); | |||
| TFE_DeleteTensorHandle(t1); | |||
| TFE_DeleteTensorHandle(t2); | |||
| TFE_DeleteTensorHandle(retvals[0]); | |||
| retvals[0]?.Dispose(); | |||
| } | |||
| } | |||
| } | |||
| } | |||
| @@ -13,7 +13,7 @@ namespace TensorFlowNET.UnitTest.NativeAPI | |||
| [TestMethod] | |||
| public unsafe void TensorHandle() | |||
| { | |||
| var h = TestMatrixTensorHandle(); | |||
| using var h = TestMatrixTensorHandle(); | |||
| EXPECT_EQ(TF_FLOAT, c_api.TFE_TensorHandleDataType(h)); | |||
| var status = c_api.TF_NewStatus(); | |||
| @@ -28,7 +28,6 @@ namespace TensorFlowNET.UnitTest.NativeAPI | |||
| EXPECT_EQ(3.0f, data[2]); | |||
| EXPECT_EQ(4.0f, data[3]); | |||
| c_api.TF_DeleteTensor(t); | |||
| c_api.TFE_DeleteTensorHandle(h); | |||
| } | |||
| } | |||
| } | |||
| @@ -24,47 +24,52 @@ namespace TensorFlowNET.UnitTest.NativeAPI | |||
| using var ctx = NewContext(status); | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| var hcpu = TestMatrixTensorHandle(); | |||
| var device_name = TFE_TensorHandleDeviceName(hcpu, status); | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| ASSERT_TRUE(device_name.Contains("CPU:0")); | |||
| var backing_device_name = TFE_TensorHandleBackingDeviceName(hcpu, status); | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| ASSERT_TRUE(backing_device_name.Contains("CPU:0")); | |||
| // Disable the test if no GPU is present. | |||
| string gpu_device_name = ""; | |||
| if(GetDeviceName(ctx, ref gpu_device_name, "GPU")) | |||
| using (var hcpu = TestMatrixTensorHandle()) | |||
| { | |||
| var hgpu = TFE_TensorHandleCopyToDevice(hcpu, ctx, gpu_device_name, status); | |||
| ASSERT_TRUE(TF_GetCode(status) == TF_OK, TF_Message(status)); | |||
| var device_name = TFE_TensorHandleDeviceName(hcpu, status); | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| ASSERT_TRUE(device_name.Contains("CPU:0")); | |||
| var retvals = new IntPtr[1]; | |||
| using (var shape_op = ShapeOp(ctx, hgpu)) | |||
| var backing_device_name = TFE_TensorHandleBackingDeviceName(hcpu, status); | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| ASSERT_TRUE(backing_device_name.Contains("CPU:0")); | |||
| // Disable the test if no GPU is present. | |||
| string gpu_device_name = ""; | |||
| if (GetDeviceName(ctx, ref gpu_device_name, "GPU")) | |||
| { | |||
| TFE_OpSetDevice(shape_op, gpu_device_name, status); | |||
| ASSERT_TRUE(TF_GetCode(status) == TF_OK, TF_Message(status)); | |||
| int num_retvals = 1; | |||
| c_api.TFE_Execute(shape_op, retvals, ref num_retvals, status); | |||
| using var hgpu = TFE_TensorHandleCopyToDevice(hcpu, ctx, gpu_device_name, status); | |||
| ASSERT_TRUE(TF_GetCode(status) == TF_OK, TF_Message(status)); | |||
| // .device of shape is GPU since the op is executed on GPU | |||
| device_name = TFE_TensorHandleDeviceName(retvals[0], status); | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| ASSERT_TRUE(device_name.Contains("GPU:0")); | |||
| var retvals = new SafeTensorHandleHandle[1]; | |||
| using (var shape_op = ShapeOp(ctx, hgpu)) | |||
| { | |||
| TFE_OpSetDevice(shape_op, gpu_device_name, status); | |||
| ASSERT_TRUE(TF_GetCode(status) == TF_OK, TF_Message(status)); | |||
| int num_retvals; | |||
| c_api.TFE_Execute(shape_op, retvals, out num_retvals, status); | |||
| ASSERT_TRUE(TF_GetCode(status) == TF_OK, TF_Message(status)); | |||
| // .backing_device of shape is CPU since the tensor is backed by CPU | |||
| backing_device_name = TFE_TensorHandleBackingDeviceName(retvals[0], status); | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| ASSERT_TRUE(backing_device_name.Contains("CPU:0")); | |||
| } | |||
| try | |||
| { | |||
| // .device of shape is GPU since the op is executed on GPU | |||
| device_name = TFE_TensorHandleDeviceName(retvals[0], status); | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| ASSERT_TRUE(device_name.Contains("GPU:0")); | |||
| TFE_DeleteTensorHandle(retvals[0]); | |||
| TFE_DeleteTensorHandle(hgpu); | |||
| // .backing_device of shape is CPU since the tensor is backed by CPU | |||
| backing_device_name = TFE_TensorHandleBackingDeviceName(retvals[0], status); | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| ASSERT_TRUE(backing_device_name.Contains("CPU:0")); | |||
| } | |||
| finally | |||
| { | |||
| retvals[0]?.Dispose(); | |||
| } | |||
| } | |||
| } | |||
| } | |||
| TFE_DeleteTensorHandle(hcpu); | |||
| // not export api | |||
| using var executor = TFE_ContextGetExecutorForThread(ctx); | |||
| TFE_ExecutorWaitForAllPendingNodes(executor, status); | |||
| @@ -25,35 +25,42 @@ namespace TensorFlowNET.UnitTest.NativeAPI | |||
| using var ctx = NewContext(status); | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| var var_handle = CreateVariable(ctx, 12.0f, status); | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| int num_retvals = 1; | |||
| var value_handle = new[] { IntPtr.Zero }; | |||
| using (var op = TFE_NewOp(ctx, "ReadVariableOp", status)) | |||
| using (var var_handle = CreateVariable(ctx, 12.0f, status)) | |||
| { | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| TFE_OpSetAttrType(op, "dtype", TF_FLOAT); | |||
| TFE_OpAddInput(op, var_handle, status); | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| TFE_Execute(op, value_handle, ref num_retvals, status); | |||
| } | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| ASSERT_EQ(1, num_retvals); | |||
| EXPECT_EQ(TF_FLOAT, TFE_TensorHandleDataType(value_handle[0])); | |||
| EXPECT_EQ(0, TFE_TensorHandleNumDims(value_handle[0], status)); | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| var value = 0f; // new float[1]; | |||
| var t = TFE_TensorHandleResolve(value_handle[0], status); | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| ASSERT_EQ(sizeof(float), (int)TF_TensorByteSize(t)); | |||
| tf.memcpy(&value, TF_TensorData(t).ToPointer(), sizeof(float)); | |||
| c_api.TF_DeleteTensor(t); | |||
| EXPECT_EQ(12.0f, value); | |||
| int num_retvals = 1; | |||
| var value_handle = new SafeTensorHandleHandle[1]; | |||
| using (var op = TFE_NewOp(ctx, "ReadVariableOp", status)) | |||
| { | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| TFE_OpSetAttrType(op, "dtype", TF_FLOAT); | |||
| TFE_OpAddInput(op, var_handle, status); | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| TFE_Execute(op, value_handle, out num_retvals, status); | |||
| } | |||
| try | |||
| { | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| ASSERT_EQ(1, num_retvals); | |||
| EXPECT_EQ(TF_FLOAT, TFE_TensorHandleDataType(value_handle[0])); | |||
| EXPECT_EQ(0, TFE_TensorHandleNumDims(value_handle[0], status)); | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| var value = 0f; // new float[1]; | |||
| var t = TFE_TensorHandleResolve(value_handle[0], status); | |||
| ASSERT_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| ASSERT_EQ(sizeof(float), (int)TF_TensorByteSize(t)); | |||
| tf.memcpy(&value, TF_TensorData(t).ToPointer(), sizeof(float)); | |||
| c_api.TF_DeleteTensor(t); | |||
| EXPECT_EQ(12.0f, value); | |||
| } | |||
| finally | |||
| { | |||
| value_handle[0]?.Dispose(); | |||
| } | |||
| } | |||
| TFE_DeleteTensorHandle(var_handle); | |||
| TFE_DeleteTensorHandle(value_handle[0]); | |||
| CHECK_EQ(TF_OK, TF_GetCode(status), TF_Message(status)); | |||
| } | |||
| } | |||
| @@ -12,7 +12,7 @@ namespace TensorFlowNET.UnitTest.NativeAPI | |||
| [TestClass] | |||
| public partial class CApiEagerTest : CApiTest | |||
| { | |||
| IntPtr TestMatrixTensorHandle() | |||
| SafeTensorHandleHandle TestMatrixTensorHandle() | |||
| { | |||
| var dims = new long[] { 2, 2 }; | |||
| var data = new float[] { 1.0f, 2.0f, 3.0f, 4.0f }; | |||
| @@ -26,7 +26,7 @@ namespace TensorFlowNET.UnitTest.NativeAPI | |||
| return th; | |||
| } | |||
| SafeOpHandle MatMulOp(SafeContextHandle ctx, IntPtr a, IntPtr b) | |||
| SafeOpHandle MatMulOp(SafeContextHandle ctx, SafeTensorHandleHandle a, SafeTensorHandleHandle b) | |||
| { | |||
| using var status = TF_NewStatus(); | |||
| @@ -64,7 +64,7 @@ namespace TensorFlowNET.UnitTest.NativeAPI | |||
| return false; | |||
| } | |||
| SafeOpHandle ShapeOp(SafeContextHandle ctx, IntPtr a) | |||
| SafeOpHandle ShapeOp(SafeContextHandle ctx, SafeTensorHandleHandle a) | |||
| { | |||
| using var status = TF_NewStatus(); | |||
| @@ -77,28 +77,28 @@ namespace TensorFlowNET.UnitTest.NativeAPI | |||
| return op; | |||
| } | |||
| unsafe IntPtr CreateVariable(SafeContextHandle ctx, float value, SafeStatusHandle status) | |||
| unsafe SafeTensorHandleHandle CreateVariable(SafeContextHandle ctx, float value, SafeStatusHandle status) | |||
| { | |||
| var var_handle = new IntPtr[1]; | |||
| int num_retvals = 1; | |||
| var var_handle = new SafeTensorHandleHandle[1]; | |||
| int num_retvals; | |||
| using (var op = TFE_NewOp(ctx, "VarHandleOp", status)) | |||
| { | |||
| if (TF_GetCode(status) != TF_OK) return IntPtr.Zero; | |||
| if (TF_GetCode(status) != TF_OK) return new SafeTensorHandleHandle(IntPtr.Zero); | |||
| TFE_OpSetAttrType(op, "dtype", TF_FLOAT); | |||
| TFE_OpSetAttrShape(op, "shape", new long[0], 0, status); | |||
| TFE_OpSetAttrString(op, "container", "", 0); | |||
| TFE_OpSetAttrString(op, "shared_name", "", 0); | |||
| if (TF_GetCode(status) != TF_OK) return IntPtr.Zero; | |||
| TFE_Execute(op, var_handle, ref num_retvals, status); | |||
| if (TF_GetCode(status) != TF_OK) return new SafeTensorHandleHandle(IntPtr.Zero); | |||
| TFE_Execute(op, var_handle, out num_retvals, status); | |||
| } | |||
| if (TF_GetCode(status) != TF_OK) return IntPtr.Zero; | |||
| if (TF_GetCode(status) != TF_OK) return new SafeTensorHandleHandle(IntPtr.Zero); | |||
| CHECK_EQ(1, num_retvals); | |||
| // Assign 'value' to it. | |||
| using (var op = TFE_NewOp(ctx, "AssignVariableOp", status)) | |||
| { | |||
| if (TF_GetCode(status) != TF_OK) return IntPtr.Zero; | |||
| if (TF_GetCode(status) != TF_OK) return new SafeTensorHandleHandle(IntPtr.Zero); | |||
| TFE_OpSetAttrType(op, "dtype", TF_FLOAT); | |||
| TFE_OpAddInput(op, var_handle[0], status); | |||
| @@ -107,22 +107,22 @@ namespace TensorFlowNET.UnitTest.NativeAPI | |||
| tf.memcpy(TF_TensorData(t).ToPointer(), &value, TF_TensorByteSize(t)); | |||
| var value_handle = c_api.TFE_NewTensorHandle(t, status); | |||
| if (TF_GetCode(status) != TF_OK) return IntPtr.Zero; | |||
| if (TF_GetCode(status) != TF_OK) return new SafeTensorHandleHandle(IntPtr.Zero); | |||
| TFE_OpAddInput(op, value_handle, status); | |||
| if (TF_GetCode(status) != TF_OK) return IntPtr.Zero; | |||
| if (TF_GetCode(status) != TF_OK) return new SafeTensorHandleHandle(IntPtr.Zero); | |||
| num_retvals = 0; | |||
| c_api.TFE_Execute(op, null, ref num_retvals, status); | |||
| c_api.TFE_Execute(op, null, out num_retvals, status); | |||
| } | |||
| if (TF_GetCode(status) != TF_OK) return IntPtr.Zero; | |||
| if (TF_GetCode(status) != TF_OK) return new SafeTensorHandleHandle(IntPtr.Zero); | |||
| CHECK_EQ(0, num_retvals); | |||
| return var_handle[0]; | |||
| } | |||
| IntPtr TestAxisTensorHandle() | |||
| SafeTensorHandleHandle TestAxisTensorHandle() | |||
| { | |||
| var dims = new long[] { 1 }; | |||
| var data = new int[] { 1 }; | |||
| @@ -135,7 +135,7 @@ namespace TensorFlowNET.UnitTest.NativeAPI | |||
| return th; | |||
| } | |||
| IntPtr TestScalarTensorHandle(bool value) | |||
| SafeTensorHandleHandle TestScalarTensorHandle(bool value) | |||
| { | |||
| var data = new[] { value }; | |||
| var t = c_api.TF_AllocateTensor(TF_BOOL, null, 0, sizeof(bool)); | |||
| @@ -147,7 +147,7 @@ namespace TensorFlowNET.UnitTest.NativeAPI | |||
| return th; | |||
| } | |||
| IntPtr TestScalarTensorHandle(float value) | |||
| SafeTensorHandleHandle TestScalarTensorHandle(float value) | |||
| { | |||
| var data = new [] { value }; | |||
| var t = c_api.TF_AllocateTensor(TF_FLOAT, null, 0, sizeof(float)); | |||