Jump to content

Search the Community

Showing results for tags 'dmm'.



More search options

  • Search By Tags

    Type tags separated by commas.
  • Search By Author

Content Type


Forums

  • Software & Hardware Discussions
    • LabVIEW Community Edition
    • LabVIEW General
    • LabVIEW (By Category)
    • Hardware
  • Resources
    • LabVIEW Getting Started
    • GCentral
    • Code Repository (Certified)
    • LAVA Code on LabVIEW Tools Network
    • Code In-Development
    • OpenG
  • Community
    • LAVA Lounge
    • LabVIEW Feedback for NI
    • LabVIEW Ecosystem
  • LAVA Site Related
    • Site Feedback & Support
    • Wiki Help

Categories

  • *Uncertified*
  • LabVIEW Tools Network Certified
  • LabVIEW API
    • VI Scripting
    • JKI Right-Click Framework Plugins
    • Quick Drop Plugins
    • XNodes
  • General
  • User Interface
    • X-Controls
  • LabVIEW IDE
    • Custom Probes
  • LabVIEW OOP
  • Database & File IO
  • Machine Vision & Imaging
  • Remote Control, Monitoring and the Internet
  • Hardware

Find results in...

Find results that contain...


Date Created

  • Start

    End


Last Updated

  • Start

    End


Filter by number of...

Joined

  • Start

    End


Group


Personal Website


Company Website


Twitter Name


LinkedIn Profile


Facebook Page


Location


Interests

Found 3 results

  1. Hello, I am trying to get my Keithley3706A up and running with LabVIEW but there are still some problems I am facing. I want to measure x channels with usage of filter. So I want to measure each channel x times and filter the results. I have found some functions like: dmm.filter.enable = True dmm.filter.count = 100 dmm.filter.type = dmm.FILTER_REPEAT_AVG dmm.filter.window =0,1 but it seems it still wont work. I am using the vi's Keithley itself provides on their homepage. Another problem I have is that I cant get more than 9 values into my buffer, for example if i want to create a scanlist with 1001:1010 I only get one value. If i create my scanlist with 1001:1009 it works just fine. I could not find any reference to this problem by now. thanks in advance Sagi
  2. I'm reading a sine wave generated by a AO DAQ and reading this same signal with another AI DAQ and a DMM at the same time. AO is set at 0.65V AI RMS reads 0.5V±0.0056 DMM reads 0.46053V ± 0.00001 0.65/1.414=0.46 so the DMM read is accurate. What makes the RMS LabVIEW function not reporting the accurate information? Could it be the Glitch Energy affecting the read? On a side note, I also need an accuracy of 0.2%. AO generates a 3KHz signal at 0.65V (450KS/s @ 150S/Buffer). AI reads at 90KS/s at 15KS/channel. DMM a 34401A is set VAC. Thanks
  3. Hello, The 34401A is not longer available in Canada. We need to change and use the 34461A as a replacement and we are having trouble with the peak-to-peak measurements (We are using LabVIEW). We use the peak-to-peak option while reading a communication line in order to get the min and max values. The code: Visa Write 34410A CONF:VOLT:AC VOLT:AC:PEAK:STAT ON INIT FETC? Visa Read Visa Write FETC:VOLT:AC:PTP? Visa Read This code no longer works with the 34461A. Reading the AC will only give the RMS values. We dont really care how to get the min-max of the communication channel, as long as it is in a respectable time (10s at most...). Reading the signal with DC is not fast enough. The communication signal is ±100mv. We verify the min-max range to be around 200mV. Thanks
×
×
  • Create New...

Important Information

By using this site, you agree to our Terms of Use.