Interesting and not so interesting update to RTINGS burn-in test. Quick summary:
- Most of the OLEDs in their test are holding up pretty well, much better than I was expecting given the early hiccups.
- For LG WOLEDs its business as usual with no severe burn-in.
- First generation QD-OLEDs are not doing as well as WOLEDs under RTINGS test condition. This is in part due to software problems such as Samsung requiring you to manually run long compensation cycles until a recent firmware update.
- Samsung continues their tradition of reducing monitor brightness after release by limiting SDR brightness on their OLED G8 monitor.
Keep in mind that this test is inherently unrealistic as they are stressing these displays way more than 99% of people would ever do. They are continually displaying static content with extremely limited variation in pixel colour for extended periods of time. However they are also only displaying SDR content so this doesnt really represent what continual HDR use might do.
Samsung continues their tradition of reducing monitor brightness after release by limiting SDR brightness on their OLED G8 monitor.
Typical marketing BS. Make monitor good in review, then tank the performance to avoid burn in warranty claims.
Samsung continues their tradition of reducing monitor brightness after release by limiting SDR brightness on their OLED G8 monitor.
Samsung just doesn’t want to have people’s trust do they? Huge shame that a company that can build an awesome hardware is consistently this scummy and makes weird software choices.
Keep in mind that this test is inherently unrealistic
I wouldn’t say it’s necessarily unrealistic.
They found burn-in with 700 hours of 16:9.
But how long should the monitor go before burning in?
3 years? In 1.5 years, 700 hours of 16:9 is only about 1.2 hours a day of 16:9 content which is not unrealistic in the slightest. Even double that is not unrealistic and burn-in in less than 1 year…
That’s certainly problematic.
This is exactly what happened to my monitor. Visible burn-in in 10 months, and 1.5 years on now it’s even worse.
Did you miss the part where they explicitly tested in on the same content the entire time, which makes burn in much worse?
It’s pretty much confirmed that they’re safe to use for normal day-to-day use now. Under these conditions you can expect 5+ years of use without serious problems.
Keep in mind that this test is inherently unrealistic as they are stressing these displays way more than 99% of people would ever do.
It is an accelerated test. Sure under normal usage you won’t see the effects in the time span rtings have but if you have the TV for 5 years you can build up to a similar amount of wear.
Samsung requiring you to manually run long compensation cycles until a recent firmware update.
Any idea how to manually run the long compensation cycle on one of the Alienware oled’s?
It should prompt you for it automatically every 1500hours use unless you disabled it the first time it asks you to.
Otherwise it does the shorter 5-15min cycle after 6hours of continuous use or whenever it goes to standby/powered off after being used for a few hours.
The short one is called Pixel Refresh, the longer one is called Panel Refresh and takes about an hour.
Well, it is purposefully done to accelerate aging on the monitor. Anyway, did someone say “displaying static content with extremely limited variation in pixel colour for extended periods of time.”? Well, game UI, and crosshair is a good example of that. Plus we do HDR in games too. So for people who game 10 hours a day, watch this I guess…
Only if you play the exact same game every single time and always on the same scene.
You can easily make 400 hours in game like witcher 3.
So, witcher 3, CP2077, one other long game and bam, you have burn in.
Not to mention no lifes that can do 3k hours CSGO in 1 year or other multiplayer games.
Keep in mind that this test is inherently unrealistic as they are stressing these displays way more than 99% of people would ever do.
It is an accelerated test. Sure under normal usage you won’t see the effects in the time span rtings have but if you have the TV for 5 years you can build up to a similar amount of wear.
I wonder if anyone has tested “underclocking” an OLED monitor; for example, if it can do 1000 cd/m^2 nits, run it at 600 cd/m^2. I would guess this would provide a much longer lifespan, and basically be the equivalent of paying more for a longer-lasting monitor?
If it’s anything like my experience with Plasma I expect signs of permanent panel degradation during normal use will show up at around 8-10k hours, and mine is mostly cause of subtitles and logos (now at almost 12k). Right now I think they said they are 5k so I’m curious to see how it’s going to go for the next update. I do need to change my TV eventually but probably going to keep it till it dies at this point.
I have a Panasonic ST30 with about 15k hours that saw plenty of PC desktop usage due to having an Htpc connected the entire time I’ve owned it, and I still have no burn in. OLED burn-in is cumulative. Plasma burn-in is completely avoidable.
So I guess AW3423DW wont work as a gaming/general monitor. Just when they got cheaper this video update drops…
Can confirm, it’s absolutely perfect for gaming/general use.
I’m over 4500 hours of use in HDR 1000 80%+ brightness. Lots of browsing/reading/gaming.
Not even a hint of burn in
My AW3423DW has almost 10K hours and has no burn in.
Thread from 6 months ago: https://www.reddit.com/r/OLED\_Gaming/comments/144smd5/aw3423dw\_with\_over\_8700\_hours\_and\_no\_burnin\_so/
My AW3423DW has almost 10K hours and has no burn in.
Thread from 6 months ago: https://www.reddit.com/r/OLED\_Gaming/comments/144smd5/aw3423dw\_with\_over\_8700\_hours\_and\_no\_burnin\_so/
Quick question abought brightness and HDR 80%+ brightness. Do you run it all the time, even desktop?
How is eye strain? Does it not hurt?
I have Lg 42C2 and run in SDR mode and 20% brightness absolute majority of time, since brightness is way too much for me to be comfortable.
I also run it this low to avoid disruptive ABL auto adjustments, but thats the price for not spending enough to get oled monitor, just oled tv.
Yes, I run it that way all the time.
I never touch my brightness and only ever turn HDR off if I’m remoting into the desktop as the HDR image gets blown out.
I put everything that I can into dark mode, and even use Dark Reader to dark mode everything I can in browser.
Once in a while ABL can be annoying on very large white apps/websites, but it’s pretty rare for me.
Watching nothing but CNN for 24/7 doesn’t seem like a particularly realistic test. Also basing conclusions on a sample of 1 is also rather misleading.
Since we are doing anecdotal “evidence”, here is mine tonally scientific test: I bought my Alienware aw3423dw at release and have been using it as my work and gaming monitor since than. I use dark mode and don’t watch CNN on my monitor, but it has over 6000 hours of use so far and no burn in…
Over 4500 hours now here, absolutely no burn in and I’ve run it in HDR 1000 high brightness the whole time.
It’s a very realistic test for monitor usage, especially when you are buying for longevity. My taskbar gets 5000 hours/year of display time on my main monitor and closer to 7000 hours/year on my secondary monitors.
My oldest monitor is at 9 years old now running 20 hours per day continously and it has no loss in quality or performance at all - and I expect to keep it for at least another 4-5 years.
It is not meant to be realistic. It is meant to be accelerated. If the TV is fine after watching thousands of hours of CNN with vibrant fixed elements then dark mode desktop users or gamers who play a lot of games with static UI elements should also be absolutely fine.
It is also great for indicating after how many hours you will start to see issues. For some that might happen in less than a year and for others it might be 4/5 years.
Sample size of 1 is an issue but a lot of the panels are used in multiple TVs, the QD OLED is used in both the S95B and A95K and the 2nd gen is in both the S95C and A95L. The WOLED panels LG make are used in a lot of TVs so really the sample size is quite a lot broader than a single panel.
They say multiple times their test isn’t the real world. It’s an extreme test with the absolute worst case scenario.
Does anyone know at what level of brightness are there tvs and monitors run?
It would be amazing if they woudl do testing of same model at the following:
- brightness level 100 nits effective
- brightness level 200 nits effective
- manual brightness level 20/100
- manual brightness level 50/100
- manual brightness level 70/100
- manual brightness level max
- auto economy/power saver
- auto vivid
Pretty hilarious how the s95b over brightens the wornout pixels making it appear like reverse burnin. I do feel like this is something that would be easily fixed in a firmware update so really curious if us regular users will have the same issue.
Pretty hilarious how the s95b over brightens the wornout pixels making it appear like reverse burnin. I do feel like this is something that would be easily fixed in a firmware update so really curious if us regular users will have the same issue.