Jump to content


  • Content Count

  • Joined

  • Last visited

Community Reputation


About volatile

  • Rank
    LAVA groupie

LabVIEW Information

  • Version
    LabVIEW 2014
  • Since
  1. I attached a minimal VI showing what Im doing now and what Id like to have faster. HW is a cDAQ-9172 with NI 9206 and 9264 for IO. The PC its running on is a normal laptop, Thinkpad x230i. Like I said, Id be happy if the vi reached 50ms loop timing. @hooovahh: Ill try continous modes @ShaunR: Yes, in the real application Im using custom scales. Pipelining the short demo vi made it even a little slower. Edit: I checked which task needs the most time, reading took 70ms. Setting the read task to continuous chopped those 70ms off of the loop time, Im reaching 25ms loop timing for a single ch
  2. Looks like I should have been clearer with my post, sorry. In the past Ive worked with LV on Windows, RT Desktop PCs, myRIOs and different cRIOs. Ive chosen Windows for this application because it should be fast enough and because of the development overhead needed for the FPGA Dev and so on. Also my application is more complex than the solar panels, but those explain my central problem (read - calc - write, no chance of writing or reading multiple data points instead of single DBLs because only the next value is known). Getting more performance out of my code is my goal now, again Im sor
  3. Lets pretend I wanted to simulate a solar panel using a cDAQ, a programmable power supply and a light sensor. Id have to measure the voltage from the light sensor and the PSUs current, use those values to find the respective operating point on my light/voltage/current-curve and update the PSUs settings. Lets also say I wanted to do this for multiple systems in parallel, all connected to 1 analog input and 1 analog output module in 1 cDAQ. What would be the best way to achieve the quickest response time? Simply reading and writing single samples seams to be pretty slow (though I can encapsulat
  • Create New...

Important Information

By using this site, you agree to our Terms of Use.