John Woodhouse posted on Tue, 31 May 2011 03:49:49 -0700 as excerpted: > I'm on a number of forums that top post normally but I can try and > remember to switch on here. Back in the MSIE beta days (which that long rambling mentioned), their newsgroups didn't really care too much, as long as people didn't post a humongous unbroken block of reply at the top so people had to scroll down too far. The long quote block is bad in any case, tho (with the exception of where whole-quote-at-bottom is accepted). In general, if someone's posting more than a reasonably-sized page of worth of quoted text in a block, and replying below, they're doing it wrong. The idea is to edit to the bit replied to and reply to it, with inling quote/reply/quote/reply if there's several elements replied to. With some exceptions for technical output quoting, etc, then, if in a reasonably sized page one can't see a bit of their reply either top or bottom, they're quoting too much and need to either trim or summarize (with appropriate [] demarking so it's known to be a summary not a direct quote) the point to which they are replying. But I've not seen a *ix related list where top posting is really considered great, tho it's often tolerated. And personally, I don't really object to it either, certainly not anything even CLOSE to the extent to which I object to HTML posts. My only frustration with top posting is in trying to reply, keeping appropriate context, etc. Sometimes I reformat it for top quoting instead of top posting. Sometimes (as here) I reply to another post instead. Sometimes I reply to only one select-quoted point instead of the more thorough reply I'd have otherwise posted, and sometimes I just blow it off and don't reply at all, as by the point in the thread at which it becomes a real problem, any further contribution I might make isn't as important as it might be were I the first reply (which by definition can't have that problem). So I'm not really going to object to top posting, as long as you're not using HTML. I might gripe a bit in replies, etc, but that's different than objecting strenuously... > I've avoided dual monitors so far. Many use one as a sort of launch > point and another for actually working within but it can be useful to > mix that a little in the interests of real estate. Task bars are likely > to always fit in nicely for people who use just one monitor. Make that /most/ people and I'm fine with it. =:^) By the time I left MS, I didn't really need them so much any more, and believe I dumped them shortly after switching to Linux, as Linux had so much else of interest to stick on the panels. =:^) All that other stuff kind of crowded out the taskbar, which I didn't use enough to be worth it anyway. OTOH, the new taskbar/launcher combos, where many app buttons remain in the same place whether the apps running or not, and clicking it either starts the app or switches to it, is a very interesting innovation. It follows intuitively from the way things were going, but it's still rather creative. I hope it doesn't end up being one of those infamous patent problems... Anyway, if I hadn't switched to primarily hotkey launching, which if you think about it, is similar positional memory at work, only on the keyboard, not the task bar, I expect I'd find that rather useful. And I expect it'll keep the taskbar around for some a little longer than it might have stayed around otherwise, were it just a traditional running taskbar. But there's nothing wrong with that. Customize it to the way you work and be happy! =:^) > I run a 22ins monitor at 1680x1050 that's been with me for many years. > About 9 I think. Now hd tv's are available I am wondering about moving > in that direction but there really isn't much of a gain in resolution so > it's a question of just how big one can go before the dot size/spacing > becomes a problem. I run the monitor at 100dpi which produces nice tight > clear text. that is a little small for ageing eyes without some aid. At > some point I intend to play with this and x scaling to see just how > large I can go before things get objectionable. This is a rather practical consideration I too am facing. My current setup is 96 dpi standard. But while neither the 1080p monitors nor X appeared to have a built-in half-size/quarter-square resolution, I created one, 960x540, and for some time (until I switched to kwin's openGL based zooming effect, which I use most of the time now, tho the xrandr scripts I used before are still in my hotkey config), I regularly used the 960x540 modeline I had created. Somewhat to my surprise, I didn't really have a problem with that, at least at the comfortable distance of about four feet from the monitors I like to sit. (You'd not want to be closer than about 3 feet to dual-42- inch anyway, you'd have to move too much to see the whole thing!) Being half the normal resolution with 96 dpi, it amounted to 48 dpi. I've decided I don't have the problems with aliasing that some people have -- my mind must compensate automatically as at size, I can see the blocks if I deliberately look for them, but I don't see them and they don't normally bother me if I'm not looking for them. Part of it may also be that I *STRONGLY* (to the point I can start feeling physically sick if stuck too long using a standard scheme, check back in the archives either here or on the Linux list for my complaints about color handling back with kde 4.2 when I first tried to switch from 3.5, if you doubt) prefer the traditional light text and foregrounds on a dark background, reversing the more common dark text and foregrounds on a for me WAY too bright background. I remember back in the CRT era when I used to run my monitors close to their upper resolution tolerances, thus, close to the lower vertical refresh-rate tolerances. I experimented quite a bit trying to eek out every last bit of resolution I could, and realized in the process that the darker a picture was the more tolerate I was to low refresh rates. I eventually decided that part of my "reverse color- scheme" preference was because I tended to run high resolutions and thus low framerates, and long before I realized why, I instinctively didn't like the harsh blinking and glare of the brighter color schemes at the refresh rates the monitors I had could deliver at the resolutions I ran them at. Anyway, based on the half-resolution experiments I ran recently, and this is on higher quality LED based monitors (not florescent-backed LCDs) so it wasn't simply blur masking the problem, I don't anticipate that I'll have problems with upto 42" monitors, almost double-size what I have now. Were I to go higher than that, 47 would be borderline and 50-inch-plus would likely start to pixelate for me, but 42-inch should be fine as long as it's full 1080p or better resolution. Meanwhile, one of the things TVs do (or used to do at least) to look good at the larger sizes but smaller resolutions of even down to standard-TV, where monitors traditionally do just the OPPOSITE, is increase the blur. At the full-HD 1080p level, both TVs and monitors run comparable or often the exact same resolution and refresh rates, but monitors are at least in theory designed for sharp text display even up into the 30+ inch range, while TVs, particularly the larger ones, may blur a bit to avoid the pixelation effect. But, due to the mass production effect, TVs tend to be as cheap or cheaper even with the required tuner that monitors can leave out, so they're tempting. Additionally, I strongly suspect that at least in the 19-to-37- inch ranges where the commonly overlap, despite the theory above, the mass- production economics are strong enough that the exact same panels are being used for both monitors and TVs. Meanwhile, one thing I'd STRONGLY recommend to anyone purchasing for monitor use, is that you spend the few extra bucks for LED quality. I didn't realize how big a difference it made originally, and purchased the cheaper standard florescent-backed LCDs. But an accident cracked the one so I sold the good one to a friend and threw in the cracked one (which was still in good enough shape he used it as a second monitor for awhile himself), and upgraded to slightly smaller but *ASTOUNDINGLY* better color- quality LEDs. Had I not actually run the LCDs for a few months, I'd have not noticed the difference, but having done so, and having actually run the kgamma adjustment sequence for both the LCDs and the LEDs... put it this way, if I were satisfied with LCDs any longer, they're enough cheaper I'd already have my dual 37-42-inch monitors/TVs! But having seen the difference, I'm holding out until I scrape the money together for the LEDs. Basically, with the LCDs, I couldn't get the kgamma settings to distinguish between all colors blocks on all the tests no matter what I did, while with the LEDs, it was easy to tell the difference. And it was me in both cases so the difference isn't just my eyes vs someone elses'. That's the difference in color rendering quality! Plus the LEDs are MUCH lighter and thinner, and run MUCH cooler and greener. Here in Phoenix, AZ especially, where it's not uncommon to run the AC for an hour or so mid-day even in January, we pay for every bit of electricity a device uses twice, once powering the device and ultimately converting the energy to waste heat, then again to pump that heat outside! So the electricity savings alone should pay for the difference over the lifetime of the device, several times over. And being lighter and easier to handle, there's less chance of "accidents" like the one that cracked the LCD triggering my buying the LEDs, too. > There is even another > possible gain from using a TV. My graphics card has an hdmi output which > means it's possible to gain a decent sound system as well all in the > same unit.. Just how well it will work from a distance of under a couple > of feet though is questionable. Meh. I already have my workstation hooked up to my 5.1 stereo using the sound system's digital out. The TV's sound might well be worse, even if used for just the front channels. But my card doesn't have HDMI out anyway, only DVI. Of course that'll probably change when I next upgrade, but I'm shooting for a decade on this dual socket Opteron system from 2003, so that'll be a couple years yet anyway... But this economy is really hurting ATM and at this rate it could be that before I get the big monitors, pushing back the full computer upgrade a couple years beyond that! This tyan's been a good board tho, and could well make it a dozen years... > Back to the topic - nearly. There is one person kicking about that > refers to the thing in the top right corner as a hemroid. LOL! I hadn't seen /that/ label before! =:^) > It's interesting to note that for all I know a cashew might be something > entirely different and not the hemroid which just goes to show how > useful a name it really is. Yeah... -- Duncan - List replies preferred. No HTML msgs. "Every nonfree program has a lord, a master -- and if you use the program, he is your master." Richard Stallman ___________________________________________________ This message is from the kde mailing list. Account management: https://mail.kde.org/mailman/listinfo/kde. Archives: http://lists.kde.org/. More info: http://www.kde.org/faq.html.