Jump to content

volatile

Members
  • Content Count

    3
  • Joined

  • Last visited

Community Reputation

0

About volatile

  • Rank
    LAVA groupie

LabVIEW Information

  • Version
    LabVIEW 2014
  • Since
    2012
  1. I attached a minimal VI showing what Im doing now and what Id like to have faster. HW is a cDAQ-9172 with NI 9206 and 9264 for IO. The PC its running on is a normal laptop, Thinkpad x230i. Like I said, Id be happy if the vi reached 50ms loop timing. @hooovahh: Ill try continous modes @ShaunR: Yes, in the real application Im using custom scales. Pipelining the short demo vi made it even a little slower. Edit: I checked which task needs the most time, reading took 70ms. Setting the read task to continuous chopped those 70ms off of the loop time, Im reaching 25ms loop timing for a single channel or 50ms for 16 channel r/w - THANK YOU! 2slow.vi
  2. Looks like I should have been clearer with my post, sorry. In the past Ive worked with LV on Windows, RT Desktop PCs, myRIOs and different cRIOs. Ive chosen Windows for this application because it should be fast enough and because of the development overhead needed for the FPGA Dev and so on. Also my application is more complex than the solar panels, but those explain my central problem (read - calc - write, no chance of writing or reading multiple data points instead of single DBLs because only the next value is known). Getting more performance out of my code is my goal now, again Im sorry I didnt make that clear in the first post. I know the PSUs arent really quick, thats okay. The data sheet states rise and fall times for the outputs in the range of 2ms to 50ms (depending on the load). Id be perfectly happy when the code reading the voltages and feeding the PSUs is as fast as that. Microcontrollers certainly would work but again that would take to much time now (and the simulation is only the first step in the project, later its the 'real deal' when I know I wont break something). So, btt: Ill try to build a minimal VI that shows the problem tomorrow to post here, then we can be more specific. For now Id like to know what the general approach would be here. Would the code be significantly faster if I read all Channels in parallel (with 1 daqmx read), calculate everything and write the values with 1 write instead of single DBL reads and writes encapsulated in multiple VIs (Overhead)? Could the USB connection (cable quality, length, USB hubs) make a big influence? Would an Ethernet Chassis be faster? Is there a way to speed up reads and writes in the cDAQ?
  3. Lets pretend I wanted to simulate a solar panel using a cDAQ, a programmable power supply and a light sensor. Id have to measure the voltage from the light sensor and the PSUs current, use those values to find the respective operating point on my light/voltage/current-curve and update the PSUs settings. Lets also say I wanted to do this for multiple systems in parallel, all connected to 1 analog input and 1 analog output module in 1 cDAQ. What would be the best way to achieve the quickest response time? Simply reading and writing single samples seams to be pretty slow (though I can encapsulate everything neatly this way). Is there a better way?
×
×
  • Create New...

Important Information

By using this site, you agree to our Terms of Use.