The Super Nintendo’s interlaced video mode was basically never ever used. It could output 60Hz and more than often did.
Only some games had limited framerate for various reasons, such as Another World being limited by cartridge ram or Star Fox being limited by the power of the SuperFX. Yoshis Island also used the SuperFX and wasn’t limited like Star Fox was. Occasionally there was slowdown if a developer put too much on screen at once, but these were momentary and similar to today when a game hitches while trying to load a new area during gameplay.
I’m just being nitpicky because you are using CRT interchangeably with Television. CRT’s are used in TV’s but aren’t interlaced unless the circuitry around them sends interlaced. So no, interlacing is not native on CRT’s when receiving an interlaced signal. If I plugged a Nintendo into my old ViewSonic CRT, I wouldn’t get a signal because it didn’t support NTSC interlaced input.
It’s like saying interlacing is native on LCDs. LCD TVs are interlaced, not LCDs.
I’m just being nitpicky because you are using CRT interchangeably with Television.
That was intentional on my part because of the audience and good communication. You’re technically correct, but without a paragraph of tangential and irrelevant explanation your audience isn’t going to understand you. Modern parlance usage of “television” isn’t the CRT appliance, its any appliance that shows the moving pictures and sound content of television programming. If you walk into any store today and buy a TV, you’re going to get an LCD, AMOLED, or quantum dot display. None of those are CRTs, yet everyone born after about 2002 will associate a TV or Television with a flat panel non-CRT display.
So no, interlacing is not native on CRT’s when receiving an interlaced signal.
And in nobody’s mind was the vision of plugging a SNES into a computer monitor CRT. You introduced that idea only to show how its wrong. You win at pedantry, but lose at communication.
If someone says to you “I’m watching TV”, do you poke your head around the back of the unit to make sure it has a tuner in it and if it doesn’t you quip back to correct them “You’re not actually watching a TV, you’re watching a monitor. A TV requires a tuner, which this unit does not have, making it a monitor, not a TV”?
It is 59.94 fields per second, translating into 29.97 FPS. Interlaced video is fun. Reason why it’s not a round 60 or 30 FPS is due to maintaining compatibility with black and white sets.
240p uses each field as a frame, though, while still maintaining compatibility with NTSC. This is what most consoles pre-6th generation uses (same with PAL, but 288p at 50 FPS)
At 480i. SNES used 240p, which is technically not standard NTSC, but compatible. Nintendo called this “double strike”, since each field would display in the same location.
Sure they were technically 30 “fields” per second, but most games updated 60 times a second, even SMB on NES. You only saw one half of what the internal console rendered which is an output issue, not a rendering one.
Add on 480p and you get both 60 frames and 60 fields per second
If you don’t want the raw experience, the Viva New Vegas modlist does an amazing job of “vanilla plus”. Haven’t finished my latest playthrough, but nothing felt out of place, and it was less buggy than vanilla.
vivanewvegas.moddinglinked.comViva New Vegas is a community curated list of mods providing a near-Vanilla, yet expanded and bug-free gameplay. There aren’t any more recommended mods than what is on the list, anything beyond would be for your preference. It is SUPER important to follow the directions with scrutiny, mega mod lists are very fragile and finicky during the installation process.
Now, if you’re feeling particularly insane, you can actually splice Fallout 3 and Fallout: NV into what is known as The Tale of Two Wastelands. This essentially provides FNV’s updated game engine and modding support for FO3. Good luck: taleoftwowastelands.com/viewforum.php@f=51
I’ll second the recommendation for Tale of Two Wastelands. I think having the New Vegas engine makes Fallout 3 so much more playable. You can actually use the iron sights, there’s ammo types, etc. I play a heavily modded TTW and I’m having a blast blowing Swampfolk’s heads off with the Y86 gauss rifle.
I’d reflexively say “Yes” to New Vegas but it depends on what Fallouts you’ve played, and what you don’t like about them.
FO1 & 2 offer ZERO hand holding and expect you to know how to play an RPG, but offers a very open approach to the world and plot
FO3 and 4 are great games that primarily struggle with permanence of your actions in the world - it’s pretty on rails between and during setpieces, no secondary plot to really get lost in
I tried 2, 4, and 76. Only for like five hours each maybe? I can't even pîn point what it is I don't like. I don't mind on rails if the story is good, I don't mind open world or plot if it's rich, I've played other post apocalypse games and enjoyed them (Metro 2033 springs to mind).
Just something about the package that is Fallout I keep bouncing off of. I like Morrowind and Skyrim, so I really don't know.
I can’t comment on ‘76 but I have played the Metro series, which is 100% on rails but makes it work.
Fallout has a tongue in cheek goofy that permeates the IP and casts a thin layer of non-serious over everything. The brutalisism and commitment to tone is what I loved about Metro and STALKER, but Fallout is Disneyland in comparison
I think this is what makes Fallout a love it or hate it setting.
Fallout tells often whimsical stories against the horrific backdrop of nuclear annihilation, and that’s what gives it it’s charm IMO.
I actually feel like it’s more realistic in a sense than overly grimdark settings. People are goofy, and with over 200 years since the bombs fell it’s believable that people will have some laughs and some motivations other than pure survival.
Oh, absolutely and that’s why I love NV and 1 & 2. They have self awareness and embrace the whimsy, a character like Myron or ’Fisto’ the sexbot would NEVER feature in a grimdark lore like STALKER, but that’s the humanistic charm of it.
If there’s some kind of post-apocalyptic society, there’s gonna be weird people and freaks, just like now - but in an absurd context.
You are all heathens, blaspheming against the Sacred Game, get smited with all the fury of an angry nerd (with a downvote, the extent of my rage and power)
I remember playing this game in high school when it came out, and it was an amazing game at that time for that time. It had bugs, however the branching storylines and care put into that world is pretty evident when compared to fallout 3
Tbh I’d disagree. I played NV for the first time last year, and it was a great experience. It definitely shows its age a bit, especially needing to use the pipboy to swap to grenades. That being said, the architecture of the game is amazing. The map, story, and gamedesign is superb, and now that I’m playing fallout 4, I’m enjoying the new game aspects but I’m missing everything that’s great about new vegas.
I disagree. I’m not much of a Fallout fan, but I picked up New Vegas for the first time because of the show, installed the Viva New Vegas mods mentioned in another comment, and this game absolutely fucking rules.
Any true 2D game, because the console was designed for 2D games. The SuperFX chip used for Star Fox was also used for Yoshi’s Island, which did maintain 60Hz.
Because the moon landing was rendered by the TV station and your TV only showed the end result. You can do the same with GeForce NOW or other streaming service.
It’s almost like having double frame buffers for 720p or larger, 16 bit PCM audio, memory safe(ish) languages, streaming video, security sandboxes, rendering fully textured 3d objects with a million polygons in real time, etc. are all things that take up cpu and ram.
I will run any game at 60fps if it was designed for this exact machine that does nothing but play games designed for it and is also 16-bit with pixel graphics and also has low quality audio and also fits in the memory of the cartridge
I didn’t realize web browsing in Chrome required fully textured 3D objects. Not to mention playing 720p video with PCM audio in a separate app doesn’t grind everything to a halt.
well the gpu doesn’t care if it’s 2d or 3d, but you are rendering a whole bunch of textured triangles… (separated into tiles for fast partial or multithreaded re-rendering), and also just-in-time rasterizing fonts, running a complex constraint solver to lay out the ui, parsing 3 completely separate languages, communicating using multiple complex network protocols, doing a whole bunch interprocess communication in order to sandbox stuff
WebGL means the browser has access to the GPU. Also, the whole desktop tends to be rendered as a 3D space these days. It makes things like scaling and blur effects easier, among other benefits.
SNES resolution was 256 x 224 with a 15 bit colour. You’re using your chrome on a 3,840 x 2,160 screen with a 32 bit colour. That’s 308x more data per frame to render. You should be really impressed that in a span of just three decades we got 300x improvement in performance.
lemmy.world
Aktywne