Decades ago, the TV took five minutes to warm the tubes up before one could watch the news.
Today, the TV takes five minutes to boot, install updates, and mangle some configuration before one (eventually) can watch the news - if the TV has not lost it’s list of stations again.
By the mid 80s and 90s, CRTVs took just seconds to show output on the screen. Even the really old tube TV my grandma had would warm up within seconds.
I once got gifted a TV from a nice elderly guy. The TV had been edge of technology when it was built: It had a wireless remote! Although the remote worked with ultrasound instead of infrared…
This beast took several minutes before it actually showed a picture.
Must’ve been a REALLY old one. I’m old as dirt, and they’ve taken mere seconds all my life. Even fast TVs now take longer to show a picture than the console ones we had when I was a kid, although I did see some from the 50s and 60s that took quite some time.
The original “clicker” remotes were really neat tech! The way it worked unfortunately limited the number of buttons you could have, but still ingenious.
https://www.theverge.com/23810061/zenith-space-command-remote-control-button-of-the-month
I used to have one of those black plastic (or was it Bakelite?) Space-Commander 400 remotes, pictured in the black and white ad.
I was walking home from grade school. Somebody was getting rid of their ancient TV, and had left it on the curb. The boxy, awkwardly shaped remote was in its “holster” on the TV, so I grabbed it and took it home. Before then, I had assumed that only infrared wireless remotes existed.
The idea that a remote could work by ultrasound fascinated me, and the fact that it didn’t even need batteries absolutely blew my little mind.
Feels like everything is much more a faff to set up, then one bit updates & something or other is longer compatible.
Don’t even want to think about the waste it must generate, both of devices & of the hours trying to get things to work whether at the development end or in the home.
at this point i don’t understand why people bother with TVs rather than just hooking up an actual normal computer to a big screen and just watching youtube or torrenting media
Because you can get much bigger for cheaper compared to a monitor.
deleted by creator
even the monitors are “smart” now. have you seen Samsung’s latest computer monitors?
I even hate how Windows tries to determine which monitors are in use, so if I turn one off it sends everything to the other screen. Just literally be a dumb thing that displays output, please, don’t try to think for me.
Because size.
Tbh was referring to devices generally. I don’t go near televisions today. Too big, too expensive & way too much faff.
I tell my laptop to put the video in the vga port. It does. That’s it. There’s nothing plugged in, but it’s there.
I plug a vga cable in. There’s video in there now. With enough paperclips, I could get it out the other end. My laptop does not care. It wiggles the electrons regardless.
I plug the other end of the cable in. The shielding was eaten by mice and two pins are dead. But alas, lo and behold, purple tho it may be - the video comes out and is displayed.
Meanwhile, hdmi protocol negotiation wants to know if you’d like to set your screen as the default sound device. Not that teams would use it anyway. Actually nevermind, the receiving end doesn’t support the correct copyright protection suite. Get fucked, no video for you.
This is stressing me out :P
Flatscreens: Set it gently on the table, or you might break the screen.
CRTs: Set it gently on the table, or you might break the table.
4K HDR 120 Hertz, that we can easily put up surround sound is great. But fuck, you sneeze wrong and it gets a weird scanline issue, like mine.
Even old flat screens are ridiculously heavy compared to new ones. I replaced an old Sony 720p screen that weight probably 20 pounds with a 1080p smart TV of the same size that I could lift one-handed. And the new one cost less than $200.
I somewhat miss my old Sony 720p screen… it came with a full electronics diagram in case you wanted to repair it.
I tried using the “smart” of my tv once, it was so slow and laggy i plugged in my 7+ year old Roku and never touched the smart again.
deleted by creator
deleted by creator
All of those issues are covered by other devices that most people will already have. An XBox360/PS3 or any newer gaming system can output 4k and make the smart features in the TV unnecessary. The same is true of a cable box, Roku plugin, fire stick, or any other streaming device. All the TV really needs is to display the 4k signal it receives. TVs don’t even really need receivers anymore just a USB hub, a processor for video and audio output, and a screen.
deleted by creator
Or grab a mini media focused computer to hook up to it.
I was fine with the quality of old TVs ¯\_(ツ)_/¯
And no thank you, I’m not going to do all that. I don’t care enough about any shows to go through all that hassle. I just want my TV to work without extra expense, and I will complain when it doesn’t because I hate big corporations and I want them to fail.
I was happy with the quality, and don’t get more enjoyment from all the advancements since, but only ever remember plugging it into the wall, plugging an aerial into the back of it & pressing one button to get the tuner to pick up channels. Batteries into the remote once that became a thing. Plug in a VCR or DVD player once they appeared.
No need for a phone line or internet or updates.
I would prefer analog TV, honestly. At least then a partial signal is something.
Descrambler box + skinamax?
625i@50fps blobby
The comparison between CRT and digital is not as simple as “625 vs 4k”. Those analog signals were intended for a triangular subpixel shadow mask with no concept of horizontal pixel count, making them effectively into ∞×625i@50fps signals (1), compared to the digital fixed 3840×2160@60fps square pixels regardless of subpixel arrangement.
It takes about a 6x6 square pixel block to correctly represent a triangular subpixel mask, making 4K LCDs about the minimum required to properly view those 625i@50fps signals.
(1) I’m aware of optics limitations, CoC, quantum effects, and ground TV carrier band limitations, but still.
deleted by creator
I grew up with CRTs and VCRs, hard pass. There’s a certain nostalgia to it all: the bum-DOOON sound as its electron gun warmed up, the smell of ozone and tingly sensation that got exponentially stronger the closer you were, crusty visuals… But they were objectively orders of magnitude worse than what we have now, if nothing else than because they don’t weigh 150 pounds or make you wonder if watching Rugrats in Paris for the 30th time on this monster is giving you cancer. Maybe it’s because I’m techie, I’ve never really had much issue with “smart” TVs. Sure, apps will slow down or crash because of memory leaks and it’s not as customizable as I’d like, but I might be satiated just knowing that if push comes to shove I can plug in a spare computer and use it like a monitor for a media system.
I’m rooting it if it starts serving me out-of-band ads, though.
deleted by creator
I feel like I’ve missed something. I don’t dispute any of the horrible experiences people have had, however I’ve had nothing but good luck. The only thing about our current television that bothers me is the promotional wallpapers that get applied every-fucking-time a new Disney property needs advertising. We buy relatively modestly priced units in the $300-$500, so maybe we just have different expectations than someone buying a much more high end unit. It is also possible that it has been pure luck and I’ll reply to this message one day soon to recant everything.
The TVs you buy force you to see ads even when you’re not watching a program and you’re like “I’ve had great luck”?
Some people don’t base their entire personality around hating the existence of ads and jumping through outrageous amounts of steps to avoid them.
So yeah, count me in for a TV that always works how I want it to but had a background ad that I can completely ignore and has no actual bearing on my life.
There are so many more important things for me to spend my time and energy worrying about.
Ah yes, the only two options: complete apathy, or basing your entire personality around hating ads. Did I like hit a nerve or something, or why did you get so defensive over a joke?
deleted by creator
promotional wallpapers
What’s that? I mean, I know what a wallpaper is, but when would a TV display one of those?
Just don’t give your TV the wifi password.
Unfortunately streaming has become the norm, and cable’s no longer affordable
I’d rather go outside if that’s how it’s gonna be
You don’t have to let your TV do the streaming though, it just needs to play it. Lots of other devices can connect to streaming services, eg gaming consoles and DVD players. Me, I have a media PC that does all the fun stuff (and lets me stream my library while I’m away), but you could easily use an old laptop.
This is how I do it too. I bought a mini office PC and HDMI’d it to the TV. It’s nothing fancy, just an i5, a wifi card and IHD GPU. I threw Kodi onto it, cancelled every streaming service I was using, and then returned to the high seas to fill up the 10TB external HDD I connected. When a new episode of something drops, I just download it and then Kodi nicely organizes everything.
I’m a Jellyfin guy, but same.
There is a Jellyfin addon for Kodi, just saying.
deleted by creator
Absolutely, I didn’t mean to imply you were wrong for calling them out.
deleted by creator
You know, you don’t have to engage with everyone who comments on your post. Sounds like you’re taking on a little bit too much. Sorry for my part in that, but I hope you find a way to ease any stress you’re under.
How the hell did you get -1 downvotes OP?
Idk thankfully beehaw disables downvotes, so I never have to see that side of things
I liked how you could still watch a station even if it was barely coming in.
This is less an issue of “smartness” and moreso because analog signals degrade gracefully whereas digital signals are all or nothing unless specific mitigations are put in place. HDMI hits kind of a weird spot because it’s a digital protocol based on analog scanlines; if the signal gets disrupted for 0.02 ms, it might only affect the upper half and maybe shift the bits for the lower half. Digital is more contextual and it will resynchronize at least every frame, so this kind of degradation is also unstable.
analog signals degrade gracefully whereas digital signals are all or nothing unless specific mitigations are put in place
Not really. Digital signals come over analog mediums, and it’s up to the receiver to decide how much degradation is too much. Mitigations like error correction are intended to reduce the final errors to zero, but it’s up to the device to decide whether it shows/plays something with some errors, and how many of them, or if it switches to a “signal lost” mode.
For example, compressed digital video has a relatively high level of graceful degradation: full frames come every Nth frame and they are further subdivided into blocks, each block can fail to be decoded on its own without impacting the rest, then intermediate frames only encode block changes, so as long as the decoder manages to locate the header of a key frame, it can show a partial image that gets progressively more garbled until the next key frame. Even if it misses a key frame, it can freeze the output until it manages to locate another one.
Digital audio is more sensitive to non-corrected errors, that can cause high frequency and high volume screeches. Those need more mitigations like filtering to a normalized volume and frequency distribution based on the preceding blocks, but still allow a level of graceful degradation.
deleted by creator