Rise of the Tomb Raider

Rise of the Tomb Raider

View Stats:
Video Memory usage is insane and making very high textures unplayable. Is there a fix for this?
So I have a 2060 super with 8gb of vram. When on very high textures the game warns me that I need at least 4gb. No problem.

Except that this game gobbles all 8gb and then stutters constantly and becomes unplayable.

Anyone know a fix for this that isn't just drop the textures to low? I play in 4k and like the shiny textures.
< >
Showing 1-12 of 12 comments
Dan_The_Noob Jul 2, 2021 @ 7:54pm 
4K is never going to work on 2060 with max textures...
HelenaEX2 Jul 3, 2021 @ 12:00pm 
If you had the antialiassing on SSAO x4 try to reduce this one.

If you are playing on DX12 just go back to DX11. DX12 is unestable on this game.
Bowielicius Jul 5, 2021 @ 11:59am 
I had the same issue with the same card. Turn off antialiasing or however is spelled. Literally doubled my fps. I play in 1080 tho.
Dan_The_Noob Jul 5, 2021 @ 10:02pm 
I play with a GTX1070, max settings (except anti-aliasing off) and i force FXAA through ReShade (the ingame anti aliasing is busted) definitely not a solid 1080p/60, but it's good enough most the time for me.
Last edited by Dan_The_Noob; Jul 5, 2021 @ 10:02pm
LordOfTheBread Jul 5, 2021 @ 10:31pm 
Originally posted by Dan_The_Noob™:
I play with a GTX1070, max settings (except anti-aliasing off) and i force FXAA through ReShade (the ingame anti aliasing is busted) definitely not a solid 1080p/60, but it's good enough most the time for me.

You can't compare 4K with 180p as this is 4 times the amount of pixels.

As for OP, a 2060 won't do 4k with high textures due to the limited amount of memory the card has, a minimum of 8GB would be recommended.
HelenaEX2 Jul 5, 2021 @ 11:00pm 
Originally posted by LordOfTheDread:
Originally posted by Dan_The_Noob™:
I play with a GTX1070, max settings (except anti-aliasing off) and i force FXAA through ReShade (the ingame anti aliasing is busted) definitely not a solid 1080p/60, but it's good enough most the time for me.

You can't compare 4K with 180p as this is 4 times the amount of pixels.

As for OP, a 2060 won't do 4k with high textures due to the limited amount of memory the card has, a minimum of 8GB would be recommended.

Well the 2060 or it is a fake one or it has something creating some kinda of unestability (I belive is this last one bc D12 is awful and after years still unoptimized) bc ROTTR uses exactly 7785 MB of VRAM (except on open areas like the zone where Jacob leaves Lara after she fall to the river) those areas requires more VRAM bc the game renderize all the area to get a huge detail even on long distance, on those areas the VRAM sometimes goes up to 8-9 GB. In areas like that is normal the stuttering for the amount of textures the game is loading, but close areas like tombs or crypths shouldnt be a problem for a 2060.
LordOfTheBread Jul 6, 2021 @ 1:19am 
Originally posted by HelenaEX2:
Originally posted by LordOfTheDread:

You can't compare 4K with 180p as this is 4 times the amount of pixels.

As for OP, a 2060 won't do 4k with high textures due to the limited amount of memory the card has, a minimum of 8GB would be recommended.

Well the 2060 or it is a fake one or it has something creating some kinda of unestability (I belive is this last one bc D12 is awful and after years still unoptimized) bc ROTTR uses exactly 7785 MB of VRAM (except on open areas like the zone where Jacob leaves Lara after she fall to the river) those areas requires more VRAM bc the game renderize all the area to get a huge detail even on long distance, on those areas the VRAM sometimes goes up to 8-9 GB. In areas like that is normal the stuttering for the amount of textures the game is loading, but close areas like tombs or crypths shouldnt be a problem for a 2060.

6GB is not enough memory to play games in 4K, the RTX 2060 is not a 4K card unless you are using DLSS in performance mode.
HelenaEX2 Jul 6, 2021 @ 1:37pm 
Originally posted by LordOfTheDread:
Originally posted by HelenaEX2:

Well the 2060 or it is a fake one or it has something creating some kinda of unestability (I belive is this last one bc D12 is awful and after years still unoptimized) bc ROTTR uses exactly 7785 MB of VRAM (except on open areas like the zone where Jacob leaves Lara after she fall to the river) those areas requires more VRAM bc the game renderize all the area to get a huge detail even on long distance, on those areas the VRAM sometimes goes up to 8-9 GB. In areas like that is normal the stuttering for the amount of textures the game is loading, but close areas like tombs or crypths shouldnt be a problem for a 2060.

6GB is not enough memory to play games in 4K, the RTX 2060 is not a 4K card unless you are using DLSS in performance mode.
Dude i play a few game on a 1660Ti on 4k and you are telling me that the 2060 is not for 4k?

Is not for playing on 4k in ultra quality i agree. But play with high setting it must be. Probably no antialiassing or not on a huge quality of AA.

I even test a 1650Ti on 4k and that GPU manage many games on 4k so well. Tomb Raider is exgent and the point that stays on 45 with drops to 30 in open areas with medium -high details is something good.

I insist a RTX 2060 on 4k with ultra settings is never going to run this game. But say "Oh is not a 4k card" is kinda pointless, you can run many games on 4k with high settings a few ones on ultra but definitly not games with so powerful details like the case of this game where all the map is renderized unless you go to indoor zones.

This game running everything in the max settings except texture, leaving texture on "high" the game uses 5.8GB on 4k and using "Ultra high" goes up to 7.9Gb of course you need to sacrifice this option on the settings to run it on 4k, far as i know a GTX 1660Ti has the closest performance to the RTX 2060. Why a 1660Ti is capable to run in 4k but not the RTX 2060?

Once more i agree not on ultra or max settings but sacrificing a few options those cards can run games in 4k with an amazing quality.
LordOfTheBread Jul 6, 2021 @ 6:07pm 
Originally posted by HelenaEX2:
Originally posted by LordOfTheDread:

6GB is not enough memory to play games in 4K, the RTX 2060 is not a 4K card unless you are using DLSS in performance mode.
Dude i play a few game on a 1660Ti on 4k and you are telling me that the 2060 is not for 4k?

Is not for playing on 4k in ultra quality i agree. But play with high setting it must be. Probably no antialiassing or not on a huge quality of AA.

I even test a 1650Ti on 4k and that GPU manage many games on 4k so well. Tomb Raider is exgent and the point that stays on 45 with drops to 30 in open areas with medium -high details is something good.

I insist a RTX 2060 on 4k with ultra settings is never going to run this game. But say "Oh is not a 4k card" is kinda pointless, you can run many games on 4k with high settings a few ones on ultra but definitly not games with so powerful details like the case of this game where all the map is renderized unless you go to indoor zones.

This game running everything in the max settings except texture, leaving texture on "high" the game uses 5.8GB on 4k and using "Ultra high" goes up to 7.9Gb of course you need to sacrifice this option on the settings to run it on 4k, far as i know a GTX 1660Ti has the closest performance to the RTX 2060. Why a 1660Ti is capable to run in 4k but not the RTX 2060?

Once more i agree not on ultra or max settings but sacrificing a few options those cards can run games in 4k with an amazing quality.

A 1660ti in 4K? At what framerate? 30? Or are they games from 10+ years ago? You will have a hard time to reach stable 60FPS in 4K in GTA 5 with a 1660Ti. A bloody 16 years old game.

Dude... a 2060 has only 6GB memory, modern games use more tham that for high res textures in 4K, of course you can "run" games at 4K dfoes not mean anyone should enjoy a sub 30FPS slideshow instead of playing a smooth 60+ fps game.

The xx60 series of cards never meant to be 4K cards period.
brokenbirthday Sep 5, 2021 @ 12:50pm 
Originally posted by LordOfTheBread:
Originally posted by HelenaEX2:
Dude i play a few game on a 1660Ti on 4k and you are telling me that the 2060 is not for 4k?

Is not for playing on 4k in ultra quality i agree. But play with high setting it must be. Probably no antialiassing or not on a huge quality of AA.

I even test a 1650Ti on 4k and that GPU manage many games on 4k so well. Tomb Raider is exgent and the point that stays on 45 with drops to 30 in open areas with medium -high details is something good.

I insist a RTX 2060 on 4k with ultra settings is never going to run this game. But say "Oh is not a 4k card" is kinda pointless, you can run many games on 4k with high settings a few ones on ultra but definitly not games with so powerful details like the case of this game where all the map is renderized unless you go to indoor zones.

This game running everything in the max settings except texture, leaving texture on "high" the game uses 5.8GB on 4k and using "Ultra high" goes up to 7.9Gb of course you need to sacrifice this option on the settings to run it on 4k, far as i know a GTX 1660Ti has the closest performance to the RTX 2060. Why a 1660Ti is capable to run in 4k but not the RTX 2060?

Once more i agree not on ultra or max settings but sacrificing a few options those cards can run games in 4k with an amazing quality.

A 1660ti in 4K? At what framerate? 30? Or are they games from 10+ years ago? You will have a hard time to reach stable 60FPS in 4K in GTA 5 with a 1660Ti. A bloody 16 years old game.

Dude... a 2060 has only 6GB memory, modern games use more tham that for high res textures in 4K, of course you can "run" games at 4K dfoes not mean anyone should enjoy a sub 30FPS slideshow instead of playing a smooth 60+ fps game.

The xx60 series of cards never meant to be 4K cards period.

I have a 2060 Super. 8GB of vram. That's in the original post. Also, I play all my games in 4k. Including the new Resident Evil games, Horizon Zero Dawn (which I can also max settings that game is insanely well-optimized), Control. Literally whatever. Maybe I have to lower some settings a bit and I never use RTX (I don't see a difference). Oh yeah and also Shadow of the Tomb Raider. I was asking specifically why this game gobbles so much vram and starts stuttering constantly like it has terrible memory management or something.
HelenaEX2 Sep 5, 2021 @ 6:00pm 
Originally posted by brokenbirthday:

The xx60 series of cards never meant to be 4K cards period.
Oh yeah and also Shadow of the Tomb Raider. I was asking specifically why this game gobbles so much vram and starts stuttering constantly like it has terrible memory management or something. [/quote]

This part is intresting bc i never had issues on Shadow Of The Tomb Raider.

Resident Evil, dude RE:Engine is so ligth. Y can put the resolution scale on 200% in 4k and get 50 stable FPS. Only RE:Resistance that is optimized with the ass is the only game i can only play on 1440p stable at 60. RE2R, RE3R RE7 and RE8 runs perfect at 4k at 200% with drops as much to 45 fps (30 as worst). RE Engine is ligth. You can run it perfectly on many GPU´s

Horizon never play it. Control i only saw a video a year ago about the raytraicing but is intriguing you dont notice the RTX effects.

So you had issues with Shadow Of The Tomb Raider and you dont notice the RTX on Control. Or theres something wrong on your PC or you modify something on your configs that are creating an inestability on certain games.

BTW i was about to say that Rise and Shadow of the tomb raider depending on the area you are the game will renderize more or less details on the distance for further areas. But if GTX can handle that and a RTX cant then i belive you create an inestability with some random config.
brokenbirthday Sep 6, 2021 @ 10:06am 
Originally posted by HelenaEX2:
Originally posted by brokenbirthday:

The xx60 series of cards never meant to be 4K cards period.
Oh yeah and also Shadow of the Tomb Raider. I was asking specifically why this game gobbles so much vram and starts stuttering constantly like it has terrible memory management or something.

This part is intresting bc i never had issues on Shadow Of The Tomb Raider.

Resident Evil, dude RE:Engine is so ligth. Y can put the resolution scale on 200% in 4k and get 50 stable FPS. Only RE:Resistance that is optimized with the ass is the only game i can only play on 1440p stable at 60. RE2R, RE3R RE7 and RE8 runs perfect at 4k at 200% with drops as much to 45 fps (30 as worst). RE Engine is ligth. You can run it perfectly on many GPU´s

Horizon never play it. Control i only saw a video a year ago about the raytraicing but is intriguing you dont notice the RTX effects.

So you had issues with Shadow Of The Tomb Raider and you dont notice the RTX on Control. Or theres something wrong on your PC or you modify something on your configs that are creating an inestability on certain games.

BTW i was about to say that Rise and Shadow of the tomb raider depending on the area you are the game will renderize more or less details on the distance for further areas. But if GTX can handle that and a RTX cant then i belive you create an inestability with some random config. [/quote]


Sorry, I wrote that poorly. I meant to convey that Shadow was also another game I played in 4k with no problems. It's just Rise.

And I do notice some RTX effects, like the difference in reflections which I do use if the game allows specific aspects of RTX to be managed independently.

I don't notice lighting differences unless I take screenshots with it on/off in the same spot and look at them side by side. Than I can see very small differences that just aren't worth the huge performance hit.
Last edited by brokenbirthday; Sep 6, 2021 @ 10:09am
< >
Showing 1-12 of 12 comments
Per page: 1530 50

Date Posted: Jul 1, 2021 @ 2:59pm
Posts: 12