Jump to content

Leaderboard

Popular Content

Showing content with the highest reputation on 03/30/2022 in all areas

  1. I don't have a use case for this but I love that the API uses plymorphic VI's-a vastly underused feature IMO. If you want to organise the functions instead of having a huge linear list, you can group them by separating the menu item with a colon ":".
    1 point
  2. Dear Community, The HAIBAL project is structured in the same way as Keras. The project consists of more than 3000 VIs including, all is coded in LabVIEW native:😱😱😱 16 activations (ELU, Exponential, GELU, HardSigmoid, LeakyReLU, Linear, PReLU, ReLU, SELU, Sigmoid, SoftMax, SoftPlus, SoftSign, Swish, TanH, ThresholdedReLU), nonlinear mathematical function generally placed after each layer having weights. 84 functional layers/layers (Dense, Conv, MaxPool, RNN, Dropout, etc…). 14 loss functions (BinaryCrossentropy, BinaryCrossentropyWithLogits, Crossentropy, CrossentropyWithLogits, Hinge, Huber, KLDivergence, LogCosH, MeanAbsoluteError, MeanAbsolutePercentage, MeanSquare, MeanSquareLog, Poisson, SquaredHinge), function evaluating the prediction in relation to the target. 15 initialization functions (Constant, GlorotNormal, GlorotUniform, HeNormal, HeUniform, Identity, LeCunNormal, LeCunUniform, Ones, Orthogonal, RandomNormal, Random,Uniform, TruncatedNormal, VarianceScaling, Zeros), function initializing the weights. 7 Optimizers (Adagrad, Adam, Inertia, Nadam, Nesterov, RMSProp, SGD), function to update the weights. Currently, we are working on the full integration of Keras in compatibility HDF5 file and will start soon same job for PyTorch. (we are able to load model from and will able to save model to in the future – this part is important for us). Well obviously, Cuda is already working if you use Nvidia board and NI FPGA board will also be – not done yet. We also working on the full integration on all Xilinx Alveo system for acceleration. User will be able to do all the models he wants to do; the only limitation will be his hardware. (we will offer the same liberty as Keras or Pytorch) and in the future our company could propose Harware (Linux server with Xilinx Alveo card for exemple --> https://www.xilinx.com/products/boards-and-kits/alveo.html All full compatible Haibal !!!) About the project communication: The website will be completely redone, a Youtube channel will be set up with many tutorials and a set of known examples will be offered within the library (Yolo, Mnist, etc.). For now, we didn’t define release date, but we thought in the next July (it’s not official – we do our best to finish our product but as we are a small passionate team (we are 3 working on it) we do our best to release it soon). This work is titanic and believe me it makes us happy that you encourage us in it. (it boosts us). In short, we are doing our best to release this library as soon as possible. Still a little patience … Youtube Video : This exemple is a template state machine using HAIBAL library. It show a signal (here it's Cos) and the neural network during his training has to learn to predict this signal (here we choose 40 neurones by layers, 5 layers, layer choose is dense). This template will be proposed as basic example to understood how we initialize, train and use neural network model. This kind of "visualisation exemple" is inspired from https://playground.tensorflow.org/ help who want to start to learn deep learning.
    1 point
  3. The HAIBAL project sounds very interesting. So far there is only the Deep Learning Toolkit from the company Ngene with which it is possible to program and train artificial neural networks in native LabVIEW code. I am very interested in HAIBAL. Is there a release date yet?
    1 point
  4. The examples are working well. I don't know anything about Machine Learning, but this will be the perfect opportunity to get started. Thank you so much ! 👍 Marc.
    1 point
×
×
  • Create New...

Important Information

By using this site, you agree to our Terms of Use.