Usage, probably not. You can get 99% usage in a GPU, but still have it run cool enough to avoid thermal throttling. Conversely, you could have a GPU that might only hit 40-50% usage, but because of bad thermal paste/bad fans/fans not spinning it hits its heat limit & throttles down.
Regardless, the HD 5450 is not really a gaming GPU. I should know; the first GPU on my current machine was the 6450 (next generation version, but pretty similar in performance). It was fine for my older games (Halo:CE, StarCraft/Brood War, SWAT 3/4, etc.), but didn't cut it for more modern games.
What you should replace it with is going to depend on a) what resolutions your monitor can handle & b) the quality/wattage of your PSU. If your PSU isn't good quality and at least 450W, & your monitor is lucky to hit 1600x900 resolutions, I wouldn't go much higher than a GT 1030 or GTX 1050 (not the TI). If your PSU is powerful/reliable enough, but your monitor is still limited on resolution, you could probably go up to a GTX 1050TI or 1060 (or even an RX 470/480/570/580).
If you really want to experiment a bit, you could always get a more powerful GPU (GTX 1070/1070TI/1080, RX Fury, or RX Vega 56/64), & use "supersampling". Basically, the higher the resolution, the more work is done by the GPU. Supersampling has the GPU render the game at a high resolution, then scale it back down to the monitor's resolution to eliminate/reduce the jagged edges ("aliasing") on the screen. AMD implements it on their HD 7790 & later GPUs, & even has a list of what settings you can use depending on your monitor's resolution (http://support.amd.com/en-us/kb-articles/Pages/VSR.aspx). NVidia does something similar, looks like it was implemented with the Maxwell GPUs (https://www.geforce.com/whats-new/articles/dynamic-super-resolution-instantly-improves-your-games-with-4k-quality-graphics).