Hi,
The page http://www.postgresql.org/docs/7.2/static/datatype-datetime.html mentions that the resolution of all time and timestamp data types is 1 microsecond. I have an application that runs on both a Windows (XP with SP2) machine and a Linux (SUSE 10.2) machine. I saw that on postgres enterprisedb 8.3 installed on both these machines, the default timestamp precision on the former is upto a millisecond and on the latter it is 1 microsecond.
My curiosity is : is this a universal phenomenon ie a basic issue with Windows? Or could there be some hardware or architectural differences or something else...
And my problem is: is there any way to enforce a higher precision in Windows? Because my application badly needs it.
Please help / guide.
Thanks a million,
Shruthi