This is less an issue of “smartness” and moreso because analog signals degrade gracefully whereas digital signals are all or nothing unless specific mitigations are put in place. HDMI hits kind of a weird spot because it’s a digital protocol based on analog scanlines; if the signal gets disrupted for 0.02 ms, it might only affect the upper half and maybe shift the bits for the lower half. Digital is more contextual and it will resynchronize at least every frame, so this kind of degradation is also unstable.
I grew up with CRTs and VCRs, hard pass. There’s a certain nostalgia to it all: the bum-DOOON sound as its electron gun warmed up, the smell of ozone and tingly sensation that got exponentially stronger the closer you were, crusty visuals… But they were objectively orders of magnitude worse than what we have now, if nothing else than because they don’t weigh 150 pounds or make you wonder if watching Rugrats in Paris for the 30th time on this monster is giving you cancer. Maybe it’s because I’m techie, I’ve never really had much issue with “smart” TVs. Sure, apps will slow down or crash because of memory leaks and it’s not as customizable as I’d like, but I might be satiated just knowing that if push comes to shove I can plug in a spare computer and use it like a monitor for a media system.
I’m rooting it if it starts serving me out-of-band ads, though.