[Vm-dev] Re: Re: How does Time>millisecondClockValue get a resolution of 1 millisecond?

Louis LaBrunda Lou at Keystone-Software.com
Wed Aug 8 22:22:29 UTC 2012

Hi Eliot,

>Oh, and importantly, the VM ups the resolution of timeGetTime() via timeBeginPeriod to 1 millisecond if possible. 

Thanks for the replies.  One more question, what is used for Linux or UNIX
systems?  I would like to recommend to Instantiations (VA Smalltalk) that
they change to functions that give a finer resolution than GetTickCount
(which is what I think they or IBM use).

The current VA Smalltalk code in this area asks for a timer interrupt every
100 milliseconds.  It then checks delays and callback that have been posted
to see if any need to expire.  So, you can't really do a delay for less
than 100 milliseconds.  Even though there are places in the base code that
sets delays at less than 100 milliseconds.

I have a few programs where this is a problem.  There is a method where I
can drop the interrupt period and I have used it to set the interrupt
period to 10 milliseconds and that helps my programs greatly.  But it
really only drops the resolution to 15 milliseconds.

I would like to point Instantiations to the functions that will give a 1
millisecond resolution on all the systems they support.

Many thanks.

Louis LaBrunda
Keystone Software Corp.
SkypeMe callto://PhotonDemon
mailto:Lou at Keystone-Software.com http://www.Keystone-Software.com

More information about the Vm-dev mailing list