Jump to content

ShaunR

Members
  • Posts

    4,997
  • Joined

  • Days Won

    311

Posts posted by ShaunR

  1. I'm with Ton on this one

    Your example screen shot is actually starting at an "amplitude" of 60, not 60°. 90° would start at the peak of the waveform (cosine) so to start at 60° it would be much further up.

    For the other question. Just use the "change sign" function on the data and start at 180-degrees or -degrees depending on that your trying to achieve..

  2. I prefer being certain the data can't get corrupted (the main reason I'm using SQLite). I'm not convinced having a journaled file system permits me to avoid syncs.

    It's more to do with data loss than corruption. Don't forget, its not turning off journalling in SQLite. Its just returning as soon as the OS has the info (and therefore is present in the OSs journal). The worst that can happen (I believe) is that during a crash, changes to SQLites journal aren't transferred to the OSs journal therefore some piece of data might not be written to disk when restarted. On restart, the OS will ensure that incomplete transactions (file system transactions) are resolved. And when SQlite is started, it will ensure incomplete SQL transations are resolved. Additionally. I open and close the file on every query which automagically causes commits on every query which (in my mind) is safer. But I have made it an option so it's up to the user to decide.

    With this benchmark, turning off sync makes very little difference (maybe 5 ms) on my system (are you using something exotic for a hard drive or filesystem?). If I test a large amount of transactions the difference is enormous.

    Nope.I'm using NTFS (write cacheing enabled). But something is different since (as you can see from the images) the insert time of the get_table is more in tune with inserts of my implementation when Synch is FULL (~200 ms). The only way I can get the same results as your benchmark is to use in-memory temp tables then I'm at the same insert times. What are the compilation options for your DLL?

    Just to note it turns out if you use a blank path the database performs like you had sync off (since it's temporary it doesn't need durability).

    So my insert time is really 74ms not 68ms (path now has a default value in my test code, since forget to set it the last time).

    Yes. This I'm not sure about. Since I can also find little difference between an in-memory DB and a "temporary" DB. It doesn't state it, but what could be happening is that the journal and temporary tables are created in-memory when the db name is blank giving rise to similar performance to an in-memory DB.

  3. Here's info on frameworks

    If I don't handle strings containing null mine takes 60ms to dump, and If I don't try to concat results from multiple statements I can get down to 50ms (I think I'll remove this feature, the only real use I can think of is checking pragma settings). So I'm just as fast at the same functionality (at least in LV2010). With handling strings containing null but not concatenating I'm at 57ms. Apparently SQLiteVIEW is handling select slightly better than me, since they're still faster. At least I'm winning the insert benchmark by a good margin:D.

    That's cheating laugh.gif Thats like me getting 60 ms on LV2009 biggrin.gif

    I want it all; The extra functionality and the speed (eventually)

    I haven't checked but the GetValueByPointer.xnode may call the UI thread, if it does it won't parallelise too well.

    If you're talking about parallelising in terms of for loops across multiple processors. Then there's not much in it. A good choice of non-subroutine execution systems and subroutines yields better results.

    I'm not too happy about using the xnode (not keen on Xnode technology in its current form anyway). I will probably knock up a more raw version using moveblock since I don't need the polymorphism and who knows what other stuff locked away inside.

    I surprised that the dump is that much faster in 2009, my best guess is that it the GetValueByPointer is different somehow (maybe the old one doesn't call the UI thread).

    How are you turning off synchronization, and which SQLite library get's down to an insert of 45?

    I'm just using the Pragma command to switch synch. I've made it so a simple change in a project conditional statement means you can switch between them all.

    I don't think synchronisation is necessary on already journeled systems (e.g NTFS, and ext3). I think its more appropriate for FAT32 and other less robust file systems.So the shipped setting on the next release will be OFF.

    Here's some results from my latest incarnation showing the effect of the turning off synch.. I've switched to testing by averaging over 100 iterations since there is a bit of jitter due to file access at these sorts of times. You'll probably notice the difference between the average insert time and insert time from the last iteration. With Synch OFF they are much more in agreement.

    Synch=FULL. LV 2009 x64 on Win 7 X64

    Synch=OFF. LV 2009 x64 on Win 7 X64

    Your "get_table", SYNCH=FULL LV2009 x64 on Win 7 x64

  4. I tend to update when only at the beginning of new projects AND only when a service pack has been released. But I don't see an update as being mutually exclusive to using previous versions.. Therefore I may have the latest version installed. Just not use it for production code. Quite often I will take a completed project and mass compile offline for the new version to see what issues pop-up (there are always some).

    I take it your not on an SSP? If not, then I would suggest pushing through for the budget and acquire a 2010 license. By the time it comes through, a service pack should be available (we're about due I think).

    So sorry wacko.gif I can't really answer your question fully since I've only "played" with 2010.I would suggest you obtain permission to install the evaluation of LV2010 on a machine and try compiling some known working projects so you can run them for a week or so to see what happens. That will shake the bushes enough for you to make your own mind up since we all use little tricks and workarounds that may or may not work between versions.

  5. I haven't tried porting mine to mac yet, but I figured out how to get the calls to work. If I remember all this right (don't have access to mac right now and it's been a couple of weeks) you need to create a "sqlite3.framework" folder and copy a sqlite3.dylib file into it (you can get one from the user/lib or extract it from firefox). You set the CLN in LabVIEW to point to the framework directory, and you have to type the functions names in (the function name ring didn't work for me).

    I hacked together (barely tested on LV2010f2 WinXP and not documented) a gettable implementation to see if there was any speed improvement. There wasn't much since get table is just a wrapper around all the other functions.

    GetTable

    Insert 141

    Dump 49

    Sweet. Nice work on that.

    I think perhaps blobs my be an issue with this as they are with straight text. But still, its definitely worth exploring further. You've shown me the way forward, so I will definitely be looking at this later.

    I'm not sure why you think its not much of an improvement. For select queries its. ~60% increase on mine and ~40% increase on yours (using your benchmark results) . I wouldn't expect much on inserts since no data is returned, therefore you don't have to iterate over the result rows.

    Incidentally. These VIs run faster on LV 2009 (for some reason). Your "get_table" example on LV64/32 2009 inserts at ~220ms and dumps at ~32 ms (averaged over 100 executions). On LV2010 I get roughly the same as you. Similarly, my 1.1 version inserts at 220 ms and dumps at 77 ms (averaged over 100 executions). Again. I get similar results to you in 2010. Of course. Dramatic insert improvement can be obtained by turning off synchronization. Then you are down to an insert speed of ~45ms.

    My next goal is to get it working on the Mac. I have a another release lined up and am just holding off so that I can include Mac also. So I will have a go with your suggestion, but it looks like a lot of work if you cannot simply select the lib. Do you know of a web-page (for newbies) that details the "Framework" directory details (e.g, why it needs to be like this. What its purpose is etc)?

  6. Hi all.

    Not really a "Labview" question....but related.

    I've "zero" experience with Macs so apart from a huge learning curve, I'm getting bogged down with multiple tool chains and a severe lack of understanding of MAC (which I think is BSD based).

    I've released an API which currently supports windows (x32 and x64) and has been reported to work with Linux x32. These work because I'm able to compile dll and .so for those targets. I'd really like to include Mac in the list of support but am having difficulty compiling a shared library (dynlib?) that Labviewwill accept.

    I have set up a Mac OSX Leopard 1.5 virtual machine with labview and Code:Blocks.It all runs (very slowly) but I'm able to compile a shared library in both x32 and x64 (Well. I think I can at least.. The gcc compiler is using -m32 for the 32 bit and won't compile if I have the wrong targets). The trial version of Labview I downloaded I know is 32 bit (got that from conditional disable structure) and I think the Mac is 64 bit.

    So I have the tool-chains set up and can produce outputs which I name with a "dynlib" extension. However. No matter what compiler options I try, whenever I try to load a successful build using the labview library function (i.e. select it in the dialogue) it always says it's "Not a valid library".

    Does anyone know what the build requirements (compiler options, architecture, switches etc) are for a Mac shared library? There are a plethora of them and I'm really in the dark as to what is required (i386?, Intel?, AMD? all of the above?, -m32?, BUILD_DLL?, shared?)

    Any help would be appreciated.

  7. I suggest you create a thread in the mac section to discuss this issue.

    On a side note, I think there are some mac maniacs on info-LabVIEW, dunno if you use it...

    I didn't even know LAVA had a Mac section oops.gif

    Thanks for that. I'll give it a whirl.

    Never mind; I was doing it wrong. Thanks for the snippet.

    I don't see that polymorphic Insert VI in the palettes. What am I doing wrong?

    post-7534-032862500 1288868129_thumb.jpg

    Your not doing anything wrong. You'll just have to wait for the next release. All that's happened is I've wrapped the original in a polymorphic VI since there is another insert which should make it easier to insert tables of data without the "clunky" addition of fields (which is fine for single row iterative inserts).

  8. The difference between a patch and a stability and performance release is largely in the integration between features and how deep into the corners we sweep. A patch fixes very targeted bugs, bugs that crash, bugs that have been specifically noted by a large number of customers, or bugs affecting high profile features which have no workarounds available. This release is going after a lot of non-critical issues.

    A fair comment. Although I do subscribe to the premiss that the corners are swept away on every release and a product up-issue is an extension of a rugged base. But then again, I'm more involved with mission critical software where even "minor" annoyances are unacceptable..

    Lets hope the "Tabs Panel" resizing is fixed finally biggrin.gif

  9. I finished up the last (hopefully) pass through my library a week or two ago. I got permission from my employer to post my SQLite library (I need to decide on a license and mention the funding used to create it). and when I went to try to figure out how to make a openg library (which I still haven't done yet) I saw the SQLiteVIEW library in VIPM. Which is similar to mine. But mine has a few advantages.

    Gives meaningful errors

    handles strings with /00 in them

    can store any type of of LabVIEW value (LVOOP classes, variant with attributes),

    can handle multistatement querys.

    has a caching system to help manage prepared querys.

    As for benchmarks I modified your speed example to use my code (using the string based api I added to mine since to the last comparison), and the SQLiteVIEW demo from VIPM. This is on LV2010f2 winxp

    Yours

    Insert 255 Dump 107

    This no longer available library that I've been using up until my new one (I'm surprised how fast this is, I think it's from having a wrapper dll handling the memory allocations instead of LabVIEW)

    Insert 158 Dump 45

    SQLiteVIEW

    Insert 153 Dump 43

    Mine

    Insert 67 Dump 73

    Splendid.

    The wrapper (if its the one I'm thinking of) uses sqlite_get_table. I wanted to use this to save all the fetches in labview, but it requires a char *** and I don't know how to reference that. The advantage of the wrapper is that you can easily use the "Helper" functions" (e.g exec, get_table) which would vastly improve performance. But I'm happy with those results for now. I'll review the performance in detail in a later release.

    Did you manage to compile SQLite for the Mac?

    I managed to compile sqlite on a virtual machine, but whatever I tried, labview always said it was an invalid library.wacko.gif Even the Mac library I downloaded from the sqlite site wouldn't load into labview.angry.gif I think it probably has something to do with the "bitness",

    Any Mac gurus out there?

  10. other SQLite tools.

    Such as? LV tools for SQLite are few and far between. hence the reason for publishing this API.

    If you look back, Matt W has done some benchmarking against his implementation. There is a "Speed" demo which means anyone can try it for themselves against their tools (post them here folks yes.gif).

    There are a few tweaks in the next release, but the improvements are nowhere near the same as between versions 1.0 and 1.1. Now its more to do with what hard disk you are using, what queries you are performing, whether you can tolerate less robust settings of the SQLite dll.

  11. Hi Shaun,

    I can find my USB camera in MAX, But if i try to stop the grab in max after caputuring the image the total MAX gets closed. the same thing is happening in LabView also( whenever i call the USB close.vi function). whether this is camera issue or i have to do some other thing. please give your ideas. (I am using Mightex USB camera).

    Regards,

    Kalanga

    MAX uses NI-IMAQdx so you should be using the IMAQdx functions instead of the USB driver.llb functions (see the link I posted earlier). These are notoriously problematic. Make sure they are uninstalled / removed from your system.

    Also, make sure you have the latest IMAQdx drivers installed.

  12. Interpolation (As the name suggests) is for creating intermediate values where none exist. I don't think this is what you are trying to do (maybe wrong rolleyes.gif ).

    From the VI implementation and description it looks like you are trying to find a motor position that is a close match to the desired frequency since the motor position cannot be fractions (discrete steps). I.e you have a look-up table.

    So based on that I have modified your VI to do this.

  13. You may wish to look at this - http://lavag.org/fil...ol-inheritance/

    I haven't used it myself, so I can't comment on it.

    Thanks Yair. Unfortunately uses OGlibs so cannot try it. It talks about inheriting "daughters'" properties which might be the reverse of what I was seeking.But should we really use a tool to do this? I did briefly look at scripting to generate the VIs for the properties and methods, but couldn't get a list of them from the original control. But still. That's a lot of VIs to do what the control already does blink.gif

  14. I have been playing with them more and more lately and getting a good feel for the style thats required to use them.

    I have a few gripes too that make me stay away from them, e.g. I can't include it's reference as a member of a Class (works in src not in build) etc...

    But some really powerful stuff there - they just need a little make-over IMHO.

    Indeed. My gripes are a bit more fundamental than that though.

    All I wanted to do was add a "Clear" method,alternating row colours and make it accept arrays of strings instead of using the "Item Strings" property to a multicolumn list box. But the amount of effort in re-implementing all the standard stuff just doesn't make it worth it. Only took me 15 minutes to create the control and make the changes I wanted, then it looked like another 2 days to re-implement the standard stuff.

×
×
  • Create New...

Important Information

By using this site, you agree to our Terms of Use.