I'm not sure where to begin. I have no idea what they were going for but boy, did they miss in every aspect. We had to stop watching. It didn't capture the essence, it was not a good homage and it was just plain bad. We just can't imagine who would come up with such an awful idea like remaking such a perfect classic. I use the term "remake" ever so loosely. Our expectations were low, considering how good the original was and we were expecting some new voices honoring some old songs and music, but again...it didn't happen. If you are even slightly, the tiniest fan of the original, this rendition will make you nauseous. It is that bad.
Let's take a look at how much data YouTube uses, how to measure its data consumption, and consider tips to reduce your YouTube data usage. You'll never have to guess how much data YouTube is using again.
Estimates for how much data each of these settings uses vary quite a bit, so let's run our own calculation to figure this out. Keep in mind that this is not exact and your results may vary.
Note that for 720p quality and above, YouTube also supports videos at 60FPS (frames per second) instead of the standard 30FPS. A higher FPS results in smoother video, but also more data usage, as you'd expect. YouTube also supports HDR video, which uses additional data, but as those videos aren't common at the time of writing, we haven't considered them here.
YouTube doesn't explain what these options actually mean, which is frustrating. Presumably, Higher picture quality plays the video in 720p or above, depending on how strong your connection is. Data saver likely caps the video at 480p.
To change this option, tap your profile picture at the top-right and go to Settings again. Visit General and tap Playback in feeds. Change this to Off or Wi-Fi only to avoid wasting data.
As you move up the LCD size chain, your 720p options become more limited because vendors are going with 1080p displays in most LCDs larger than 37 inches. When it comes to plasma, Panasonic's entry-level 42-inch TH-42PX8A carries a price of around AU$1,699, while the step-up 1080p version, the TH-42PZ80A, comes in at AU$2,549. Move up to 50-inch 1080p models and you're looking at AU$3,649.
3. Why is 1080p theoretically better than 1080i?1080i, the former king of the HDTV hill, actually boasts an identical 1920x1080 resolution, but conveys the images in an interlaced format (the "i" in 1080i). In a CRT, 1080i sources get rendered on-screen sequentially: the odd-numbered lines of the image appear first followed by even lines, all within 1/25 of a second. Progressive-scan formats such as 480p, 720p and 1080p convey all the lines sequentially in a single pass, which makes for smoother, cleaner visuals, especially with sports and other motion-intensive content.
4. What content is available in 1080p?Today's high-def broadcasts are done in either 1080i or 720p, and there's little or no chance they'll jump to 1080p anytime soon because of bandwidth issues. As for HD gaming, Xbox 360 and PlayStation 3 games are available in both 720p and 1080p resolutions. (Also, the 720p titles can be upscaled to 1080i or 1080p in the user settings of those consoles).
5. What kinds of TV technologies offer 1080p resolution?Aside from CRT, which has basically been discontinued, every technology on the market comes in 1080p versions. That means you can find 1080p-capable versions utilising all fixed-pixel technologies, including DLP, LCoS and LCD projectors, and flat panels (plasma and LCD). Of course, as specified above, more affordable entry-level models are still limited to 720p resolution. But whatever the resolution, all fixed-pixel TVs are essentially progressive-scan technologies. So when the incoming source is interlaced (1080i or even good old-fashioned 480i standard definition), they convert it to progressive-scan for display.
7. What happens when you feed a 1080p signal to 720p TV?Assuming the TV can accept a 1080p signal, it will be scaled to 720p. But the caveat is that many older 720p and even some 1080p models cannot handle 1080p signals at all. In which case, you'll get a blank screen. Thankfully, most newer HDTVs can accept 1080p signals.
Whether you're dealing with 1080p24 or video-based 1080p50 doesn't alter our overall views about 1080p TVs. We still believe that when you're dealing with TVs 50 inches and smaller, the added resolution has only a very minor impact on picture quality. In our tests, we put 720p next to 1080p sets, then feed them both the same source material from high-end Blu-ray players. We typically watch both sets for a while, with eyes darting back and forth between the two to look for differences in the most-detailed sections such as hair, textures of fabric, and grassy plains.
10. Should I save some dough and opt for a 720p TV?If you're just making the leap to HDTV and find the higher-end sets out of your price range, you shouldn't feel bad about going with an entry-level 720p model (just getting HD programming is going to make a huge difference). Also, in a lot of cases, folks are looking at 720p TVs as second sets for bedrooms or playrooms, and in a tough economy, a few hundred bucks make a big difference.
Finally, it's a good idea to go with 1080p instead of 720p if you plan to use your TV a lot as a big computer monitor. That said, if you set your computer to output at 1920x1080, you might find that the icons and text on the screen are too small to view from far away (as a result, you may end up zooming the desktop or even changing to a lower resolution). But a 1080p set does give you some added flexibility (and sharpness) when it comes to computer connectivity.
Second, be sure to enable HD. In theory, checking this option would allow you to broadcast a 720p video. However, when we tried recording the video stream with this option both activated and deactivated, we saw very little difference. We still think it's a good idea to enable HD, but the video recording quality might not improve much because of it.
i think this is a verified firmware bug with many lg tv. setting hdmi resolution manually works, but after an ac power loss the wd live forget the resolution settings and boot at 720p@50hz. can someone can move this thread to firmware section and accept this as a firmware bug ? thanks for support
The downside to that is some (most?) 720p sets, which are capable of downsizing 1080p to 720p, could potentially get degraded video from the Live upsampling to 1080p, only for the TV to then downsample again.
This may be true, and it explains why when it auto negotiates, it sets the TV to 720p instead of 1080i. However, the bigger problem is that the WD boxes do not remember the video resolution/frequency or colorspace values when they are manually set by the user.
No one expects HDMI handshaking to work with every TV, However, I do expect that once I manually set the video output resolution/frequency and colorspace, that the WD box should remember those settings the next time that I turn it on. In my case, even though I have manually set the video output to 1080i 60Hz, and colorspace to RGB low, when I power the unit off/on, about 50% of the time it comes up set at 720p (which my TV does not support) and 98% of the time the colorspace comes up set at RGB high.
What it needs is to do is compare the manual setting against the reported maximum supported resolution, not the native resolution. That could perhaps be the best compromise for those manually setting their resolution.
The way Group HD video works in Zoom is simple. When this feature is enabled, HD video which is 1280 x 720 or 720p will be activated for the active speaker in the video layout. At this time, full HD or 1080p video is limited to Business and Enterprise plans.
The way Group HD video works in Zoom is simple. When this feature is enabled, HD video which is 1280×720 or 720p will be activated for the active speaker in the video layout. At this time, full HD or 1080p video is limited to Business and Enterprise plans.
When choosing High Definition cameras, one of the most fundamental questions is what of the two common resolution levels one should choose: Do you use 720p or 1080p cameras? The fundamental alternatives are familiar to any user who has bought a TV in the last few years. However, how do you choose when buying network cameras?
The main case for 1080p cameras over 720p ones is that a 1080p camera has over twice the resolution (i.e., pixels) than a 720p one. A 1080p camera has a resolution of 1920 x 1080 (2.07 MP) while a 720p camera resolution is 'only' 1280 x 720 (.92 MP). Because of that, marketing people often conclude that a 1080p is equivalent to (2) 720p cameras.
In this test, we choose two very similar cameras from the same manufacturer - one that was 1080p (Sony CH210) and the other that was 720p (Sony CH110). As you can see in the sample image below, from the outside the cameras even look identical:
[Note: this is just the latest of our comparison tests of cameras with different resolution. We routinely test SD vs 720p vs 2MP vs 5MP etc. (for other comparisons, see our indoor camera shootout, our parking lot shoot, WDR camera shootout, etc.)]
Within Sony's line, all the 720p X series (entry level) cameras (like the CH110) use the generic CMOS sensor. However, all the 1080p X series and all other HD mid level and premium level camears (like the E and V series) do use EXMOR. Finally, while the EXMOR-R [link no longer available], the latest version of EXMOR has gained attention, none of the production Sony network cameras support that.
Now, let's drop the lighting to 7 lux. Looking at the image below, all of a sudden, the CH110 actually looks brighter and perhaps better than the 110. What's going on here is a 'trick' or, at least, a subtle variance in default camera settings. The CH110 with a default max exposure of 1/8s is taking in nearly 3x the light as the CH210 with a default max exposure of 1/30s. 2b1af7f3a8