You can not select more than 25 topics Topics must start with a chinese character,a letter or number, can include dashes ('-') and can be up to 35 characters long.

README.md 11 kB

6 years ago
6 years ago
7 years ago
7 years ago
5 years ago
6 years ago
5 years ago
6 years ago
6 years ago
5 years ago
6 years ago
7 years ago
6 years ago
6 years ago
6 years ago
5 years ago
6 years ago
7 years ago
6 years ago
5 years ago
6 years ago
5 years ago
6 years ago
6 years ago
6 years ago
6 years ago
6 years ago
6 years ago
5 years ago
123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266
  1. ![logo](docs/assets/tf.net.logo.png)
  2. **TensorFlow.NET** (TF.NET) provides a .NET Standard binding for [TensorFlow](https://www.tensorflow.org/). It aims to implement the complete Tensorflow API in C# which allows .NET developers to develop, train and deploy Machine Learning models with the cross-platform .NET Standard framework. TensorFlow.NET has built-in Keras high-level interface and is released as an independent package [TensorFlow.Keras](https://www.nuget.org/packages/TensorFlow.Keras/).
  3. [![Join the chat at https://gitter.im/publiclab/publiclab](https://badges.gitter.im/Join%20Chat.svg)](https://gitter.im/sci-sharp/community)
  4. [![Tensorflow.NET](https://ci.appveyor.com/api/projects/status/wx4td43v2d3f2xj6?svg=true)](https://ci.appveyor.com/project/Haiping-Chen/tensorflow-net)
  5. [![NuGet](https://img.shields.io/nuget/dt/TensorFlow.NET.svg)](https://www.nuget.org/packages/TensorFlow.NET)
  6. [![Documentation Status](https://readthedocs.org/projects/tensorflownet/badge/?version=latest)](https://tensorflownet.readthedocs.io/en/latest/?badge=latest)
  7. [![Badge](https://img.shields.io/badge/link-996.icu-red.svg)](https://996.icu/#/en_US)
  8. [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/javiercp/BinderTF.NET/master?urlpath=lab)
  9. *master branch is based on tensorflow 2.3 now, v0.15-tensorflow1.15 is from tensorflow1.15.*
  10. ![tensors_flowing](docs/assets/tensors_flowing.gif)
  11. ### Why TensorFlow in C# and F# ?
  12. `SciSharp STACK`'s mission is to bring popular data science technology into the .NET world and to provide .NET developers with a powerful Machine Learning tool set without reinventing the wheel. Since the APIs are kept as similar as possible you can immediately adapt any existing Tensorflow code in C# or F# with a zero learning curve. Take a look at a comparison picture and see how comfortably a Tensorflow/Python script translates into a C# program with TensorFlow.NET.
  13. ![pythn vs csharp](docs/assets/syntax-comparision.png)
  14. SciSharp's philosophy allows a large number of machine learning code written in Python to be quickly migrated to .NET, enabling .NET developers to use cutting edge machine learning models and access a vast number of Tensorflow resources which would not be possible without this project.
  15. In comparison to other projects, like for instance [TensorFlowSharp](https://www.nuget.org/packages/TensorFlowSharp/) which only provide Tensorflow's low-level C++ API and can only run models that were built using Python, Tensorflow.NET also implements Tensorflow's high level API where all the magic happens. This computation graph building layer is still under active development. Once it is completely implemented you can build new Machine Learning models in C# or F#.
  16. ### How to use
  17. | TensorFlow | tf native1.14 | tf native 1.15 | tf native 2.3 |
  18. | ------------------------- | ------------- | -------------- | ------------- |
  19. | tf.net 0.30, tf.keras 0.1 | | | x |
  20. | tf.net 0.20 | | x | x |
  21. | tf.net 0.15 | x | x | |
  22. | tf.net 0.14 | x | | |
  23. Read the docs & book [The Definitive Guide to Tensorflow.NET](https://tensorflownet.readthedocs.io/en/latest/FrontCover.html).
  24. There are many examples reside at [TensorFlow.NET Examples](https://github.com/SciSharp/TensorFlow.NET-Examples).
  25. Troubleshooting of running example or installation, please refer [here](tensorflowlib/README.md).
  26. #### C# Example
  27. Install TF.NET and TensorFlow binary through NuGet.
  28. ```sh
  29. ### install tensorflow C#/F# binding
  30. PM> Install-Package TensorFlow.NET
  31. ### install keras for tensorflow
  32. PM> Install-Package TensorFlow.Keras
  33. ### Install tensorflow binary
  34. ### For CPU version
  35. PM> Install-Package SciSharp.TensorFlow.Redist
  36. ### For GPU version (CUDA and cuDNN are required)
  37. PM> Install-Package SciSharp.TensorFlow.Redist-Windows-GPU
  38. ```
  39. Import TF.NET and Keras API in your project.
  40. ```cs
  41. using static Tensorflow.Binding;
  42. using static Tensorflow.KerasApi;
  43. ```
  44. Linear Regression in `Eager` mode:
  45. ```c#
  46. // Parameters
  47. int training_steps = 1000;
  48. float learning_rate = 0.01f;
  49. int display_step = 100;
  50. // Sample data
  51. train_X = np.array(3.3f, 4.4f, 5.5f, 6.71f, 6.93f, 4.168f, 9.779f, 6.182f, 7.59f, 2.167f,
  52. 7.042f, 10.791f, 5.313f, 7.997f, 5.654f, 9.27f, 3.1f);
  53. train_Y = np.array(1.7f, 2.76f, 2.09f, 3.19f, 1.694f, 1.573f, 3.366f, 2.596f, 2.53f, 1.221f,
  54. 2.827f, 3.465f, 1.65f, 2.904f, 2.42f, 2.94f, 1.3f);
  55. n_samples = train_X.shape[0];
  56. // We can set a fixed init value in order to demo
  57. var W = tf.Variable(-0.06f, name: "weight");
  58. var b = tf.Variable(-0.73f, name: "bias");
  59. var optimizer = tf.optimizers.SGD(learning_rate);
  60. // Run training for the given number of steps.
  61. foreach (var step in range(1, training_steps + 1))
  62. {
  63. // Run the optimization to update W and b values.
  64. // Wrap computation inside a GradientTape for automatic differentiation.
  65. using var g = tf.GradientTape();
  66. // Linear regression (Wx + b).
  67. var pred = W * X + b;
  68. // Mean square error.
  69. var loss = tf.reduce_sum(tf.pow(pred - Y, 2)) / (2 * n_samples);
  70. // should stop recording
  71. // Compute gradients.
  72. var gradients = g.gradient(loss, (W, b));
  73. // Update W and b following gradients.
  74. optimizer.apply_gradients(zip(gradients, (W, b)));
  75. if (step % display_step == 0)
  76. {
  77. pred = W * X + b;
  78. loss = tf.reduce_sum(tf.pow(pred - Y, 2)) / (2 * n_samples);
  79. print($"step: {step}, loss: {loss.numpy()}, W: {W.numpy()}, b: {b.numpy()}");
  80. }
  81. }
  82. ```
  83. Run this example in [Jupyter Notebook](https://github.com/SciSharp/SciSharpCube).
  84. Toy version of `ResNet` in `Keras` functional API:
  85. ```csharp
  86. // input layer
  87. var inputs = keras.Input(shape: (32, 32, 3), name: "img");
  88. // convolutional layer
  89. var x = layers.Conv2D(32, 3, activation: "relu").Apply(inputs);
  90. x = layers.Conv2D(64, 3, activation: "relu").Apply(x);
  91. var block_1_output = layers.MaxPooling2D(3).Apply(x);
  92. x = layers.Conv2D(64, 3, activation: "relu", padding: "same").Apply(block_1_output);
  93. x = layers.Conv2D(64, 3, activation: "relu", padding: "same").Apply(x);
  94. var block_2_output = layers.add(x, block_1_output);
  95. x = layers.Conv2D(64, 3, activation: "relu", padding: "same").Apply(block_2_output);
  96. x = layers.Conv2D(64, 3, activation: "relu", padding: "same").Apply(x);
  97. var block_3_output = layers.add(x, block_2_output);
  98. x = layers.Conv2D(64, 3, activation: "relu").Apply(block_3_output);
  99. x = layers.GlobalAveragePooling2D().Apply(x);
  100. x = layers.Dense(256, activation: "relu").Apply(x);
  101. x = layers.Dropout(0.5f).Apply(x);
  102. // output layer
  103. var outputs = layers.Dense(10).Apply(x);
  104. // build keras model
  105. model = keras.Model(inputs, outputs, name: "toy_resnet");
  106. model.summary();
  107. // compile keras model in tensorflow static graph
  108. model.compile(optimizer: keras.optimizers.RMSprop(1e-3f),
  109. loss: keras.losses.CategoricalCrossentropy(from_logits: true),
  110. metrics: new[] { "acc" });
  111. // prepare dataset
  112. var ((x_train, y_train), (x_test, y_test)) = keras.datasets.cifar10.load_data();
  113. // training
  114. model.fit(x_train[new Slice(0, 1000)], y_train[new Slice(0, 1000)],
  115. batch_size: 64,
  116. epochs: 10,
  117. validation_split: 0.2f);
  118. ```
  119. #### F# Example
  120. Linear Regression in `Eager` mode:
  121. ```fsharp
  122. #r "nuget: TensorFlow.Net"
  123. #r "nuget: TensorFlow.Keras"
  124. #r "nuget: SciSharp.TensorFlow.Redist"
  125. #r "nuget: NumSharp"
  126. open System
  127. open NumSharp
  128. open Tensorflow
  129. open Tensorflow.Keras
  130. let tf = Binding.New<tensorflow>()
  131. tf.enable_eager_execution()
  132. // Parameters
  133. let training_steps = 1000
  134. let learning_rate = 0.01f
  135. let display_step = 100
  136. // Sample data
  137. let train_X =
  138. np.array(3.3f, 4.4f, 5.5f, 6.71f, 6.93f, 4.168f, 9.779f, 6.182f, 7.59f, 2.167f,
  139. 7.042f, 10.791f, 5.313f, 7.997f, 5.654f, 9.27f, 3.1f)
  140. let train_Y =
  141. np.array(1.7f, 2.76f, 2.09f, 3.19f, 1.694f, 1.573f, 3.366f, 2.596f, 2.53f, 1.221f,
  142. 2.827f, 3.465f, 1.65f, 2.904f, 2.42f, 2.94f, 1.3f)
  143. let n_samples = train_X.shape.[0]
  144. // We can set a fixed init value in order to demo
  145. let W = tf.Variable(-0.06f,name = "weight")
  146. let b = tf.Variable(-0.73f, name = "bias")
  147. let optimizer = KerasApi.keras.optimizers.SGD(learning_rate)
  148. // Run training for the given number of steps.
  149. for step = 1 to (training_steps + 1) do
  150. // Run the optimization to update W and b values.
  151. // Wrap computation inside a GradientTape for automatic differentiation.
  152. use g = tf.GradientTape()
  153. // Linear regressoin (Wx + b).
  154. let pred = W * train_X + b
  155. // Mean square error.
  156. let loss = tf.reduce_sum(tf.pow(pred - train_Y,2)) / (2 * n_samples)
  157. // should stop recording
  158. // compute gradients
  159. let gradients = g.gradient(loss,struct (W,b))
  160. // Update W and b following gradients.
  161. optimizer.apply_gradients(Binding.zip(gradients, struct (W,b)))
  162. if (step % display_step) = 0 then
  163. let pred = W * train_X + b
  164. let loss = tf.reduce_sum(tf.pow(pred-train_Y,2)) / (2 * n_samples)
  165. printfn $"step: {step}, loss: {loss.numpy()}, W: {W.numpy()}, b: {b.numpy()}"
  166. ```
  167. ### Contribute:
  168. Feel like contributing to one of the hottest projects in the Machine Learning field? Want to know how Tensorflow magically creates the computational graph? We appreciate every contribution however small. There are tasks for novices to experts alike, if everyone tackles only a small task the sum of contributions will be huge.
  169. You can:
  170. * Let everyone know about this project
  171. * Port Tensorflow unit tests from Python to C# or F#
  172. * Port missing Tensorflow code from Python to C# or F#
  173. * Port Tensorflow examples to C# or F# and raise issues if you come accross missing parts of the API
  174. * Debug one of the unit tests that is marked as Ignored to get it to work
  175. * Debug one of the not yet working examples and get it to work
  176. ### How to debug unit tests:
  177. The best way to find out why a unit test is failing is to single step it in C# or F# and its corresponding Python at the same time to see where the flow of execution digresses or where variables exhibit different values. Good Python IDEs like PyCharm let you single step into the tensorflow library code.
  178. ### Git Knowhow for Contributors
  179. Add SciSharp/TensorFlow.NET as upstream to your local repo ...
  180. ```git
  181. git remote add upstream git@github.com:SciSharp/TensorFlow.NET.git
  182. ```
  183. Please make sure you keep your fork up to date by regularly pulling from upstream.
  184. ```git
  185. git pull upstream master
  186. ```
  187. ### Contact
  188. Feel free to star or raise issue on [Github](https://github.com/SciSharp/TensorFlow.NET).
  189. Follow us on [Medium](https://medium.com/scisharp).
  190. Join our chat on [Gitter](https://gitter.im/sci-sharp/community).
  191. Scan QR code to join Tencent TIM group:
  192. ![SciSharp STACK](docs/TIM.jpg)
  193. WeChat Sponsor 微信打赏:
  194. ![SciSharp STACK](docs/assets/WeChatCollection.jpg)
  195. TensorFlow.NET is a part of [SciSharp STACK](https://scisharp.github.io/SciSharp/)
  196. <br>
  197. <a href="http://scisharpstack.org"><img src="https://github.com/SciSharp/SciSharp/blob/master/art/scisharp-stack.png" width="391" height="100" /></a>