SAM tests on the RX 6800. AMD tunes more than it reveals

Methodology: performance tests

The Smart Access Memory feature is far from perfect. And although it is a useful technology overall, it also proves to be very unstable. What was true today may not tomorrow, and an in-depth analysis shows all those patches. On the one hand, those that eliminate the biggest problems that no one boasts about, and on the other hand, those that raise average performance in small steps.

Gaming tests

The largest sample of tests is from games. This is quite natural given that GeForce and Radeons, i.e. cards primarily intended for gaming use, will mostly be tested.

We chose the test games primarily to ensure the balance between the titles better optimized for the GPU of one manufacturer (AMD) or the other one (Nvidia). But we also took into account the popularity of the titles so that you could find your own results in the charts. Emphasis was also placed on genre diversity. Games such as RTS, FPS, TPS, car racing as well as a flight simulator, traditional RPG and sports games are represented by the most played football game. You can find a list of test games in the library of chapters (9–32), with each game having its own chapter, sometimes even two (chapters) for the best possible clarity, but this has its good reason, which we will share with you in the following text.

Before we start the gaming tests, each graphics card will pass the tests in 3D Mark to warm up to operating temperature. That’s good synthetics to start with.

We’re testing performance in games across three resolutions with an aspect ratio of 16:9 – FHD (1920 × 1080 px), QHD (2560 × 1440 px) and UHD (3840 × 2160 px) and always with the highest graphic settings, which can be set the same on all current GeForce and Radeon graphics cards. We turned off proprietary settings for the objectivity of the conclusions, and the settings with ray-tracing graphics are tested separately, as lower class GPUs do not support them. You will find their results in the complementary chapters. In addition to native ray-tracing, also after deploying Nvidia DLSS (2.0) and AMD FidelityFX CAS.

If the game has a built-in benchmark, we use that one (the only exception is Forza Horizon 4, where due to its instability – it used to crash here and there – we drive on our track), in other cases the measurements take place on the games’ own scenes. From those we capture the times of consecutive frames in tables (CSV) via OCAT, which FLAT interprets into intelligible fps speech. Both of these applications are from the workshop of colleagues from the gpureport.cz magazine. In addition to the average frame rate, we also write the minimum in the graphs. That contributes significantly to the overall gaming experience. For the highest possible accuracy, all measurements are repeated three times and the final results form their average value.

Tests with active AMD Smart Access Memory will not be part of the standard methodology yet. Of course, we will focus on SAM, but for better orientation, we will include these tests in a separate article. But we’re doing this just temporarily, until GeForce graphics supports it as well. Then we switch to the opposite model and all cards will be tested with SAM turned on. Until then, however, SAM will be turned off in standard tests, and we will publish the performance increase under its influence separately. No one will be cut short by anything (neither those who have pure AMDs in their cases, nor the owners of Intel platforms) and the clarity of the results will be nicely preserved. Still, putting multiple modes of one card into the same chart (or having 500 charts per article instead of 300) would no longer do any good.

We plan to do one more thing – once a quarter to measure the impact of various updates (drivers, OS, games, BIOS) on performance. This will result in percentage increases or drops in performance that you can work with when studying older tests. It’s a bit of a compromise, but it’s definitely a better option than releasing new tests with out-of-date software. Of course, it would be ideal to test all previous cards before doing every new test, but this is unrealistic. But we believe that you will also appreciate the continuous measurement with one GeForce graphics card and one Radeon and the inclusion of the appropriate coefficient in the criteria of interactive graphs.

Computational tests

Testing the graphics card comprehensively, even in terms of computing power, is more difficult than drawing conclusions from the gaming environment. Just because such tests are usually associated with expensive software that you don’t just buy for the editorial office. On the other hand, we’ve found ways to bring the available computing performance to you. On the one hand, thanks to well-built benchmarks, on the other hand, there are also some freely available and at the same time relevant applications, and thirdly, we have invested something in the paid ones

The tests begin with ComputeBench, which computes various simulations (including game graphics). Then we move on to the popular SPECviewperf benchmark (2020), which integrates partial operations from popular 2D and 3D applications, including 3Ds max and SolidWorks. Details on this test package can be found at spec.org. From the same team also comes SPECworkstation 3, where GPU acceleration is in the Caffe and Folding@Home tests. You can also find the results of the LuxMark 3.1 3D render in the graphs, and the remarkable GPGPU theoretical test also includes AIDA64 with FLOPS, IOPS and memory speed measurements.

For obvious reasons, 3D rendering makes the largest portion of the tests. This is also the case, for example, in the Blender practical tests (2.91). In addition to Cycles, we will also test the cards in Eevee and radeon ProRender renderers (let AMD have a related test, as most are optimized for Nvidia cards with proprietary CUDA and OptiX frameworks). Of course, an add-on for V-ray would also be interesting, but at the moment the editorial office can’t afford it, we may manage to get a “press” license in time, though, we’ll see. We want to expand application tests in the future. Definitely with some advanced AI testing (we haven’t come up with a reasonable way yet), including noise reduction (there would be some ideas already, but we haven’t incorporated those due to time constraints).

Graphics cards can also be tested well in photo editing. To get an idea of the performance in the popular Photoshop, we’re using a script in PugetBench, which simulates real work with various filters. Among them are those that use GPU acceleration. A comprehensive benchmark suggesting the performance of raster and vector graphics is then also used in alternative Affinity Photo. In Lightroom, there are remarkable color corrections (Enhance Details) of raw uncompressed photos. We apply these in batches to a 1 GB archive. All of these tasks can be accelerated by both GeForce and Radeon.

From another perspective, there are decryption tests in Hashcat with a selection of AES, MD5, NTLMv2, SHA1, SHA2-256/512 and WPA-EAPOL-PBKDF2 ciphers. Finally, in the OBS and XSplit broadcast applications, we measure how much the game performance will be reduced while recording. It is no longer provided by shaders, but by coders (AMD VCE and Nvidia Nvenc). These tests show how much spare performance each card has for typical online streaming.

There are, of course, more hardware acceleration options, typically for video editing and conversion. However, this is purely in the hands of encoders, which are always the same within one generation of cards from one manufacturer, so there is no point in testing them on every graphics card. It is different across generations and tests of this type will sooner or later appear. Just fine-tuning the metric is left, where the output will always have the same bitrate and pixel match. This is important for objective comparisons, because the encoder of one company/card may be faster in a particular profile with the same settings, but at the expense of the lower quality that another encoder has (but may not have, it’s just an example).


  •  
  •  
  •  
Flattr this!

CUDIMM will work on AM5, MSI says, Ryzen 9000 or 8000 needed

Recently, we reported that the Intel LGA 1851 platform with Core Ultra 200 (“Arrow Lake”) processors will premiere certain memory innovation – CUDIMM modules that have an integrated clock driver (CKD) that enables operation at (hopefully) higher clock speeds. It was unclear for a while whether this would end up being an Intel exclusive feature, but now it looks like CUDIMMs will be usable on AM5 boards with Ryzen processors as well. Read more “CUDIMM will work on AM5, MSI says, Ryzen 9000 or 8000 needed” »

  •  
  •  
  •  

Zen 5 with V-Cache coming soon. Ryzen 7 9800X3D next month?

AMD’s launch of Zen 5 desktop CPUs was limited to regular models without 3D V-Cache, which is a differentiator of the so-called “X3D” models that significantly improves performance in games. Gamers have no successor for the past generation Ryzen 7 7800X3D, which remains the best gaming CPU. This could change soon though, it seems that AMD is starting talk about the X3D version of Zen 5 and the release is coming soon. Read more “Zen 5 with V-Cache coming soon. Ryzen 7 9800X3D next month?” »

  •  
  •  
  •  

How much does AVX-512 help Zen 5 in x265? And how to turn it on

You may not know this, but the x265 video encoder can use AVX-512, but they are unused by default for historical reasons. In the past we made a guide showing how to enable the optimizations and looked at their effect, first on Intel Rocket Lake processors and then on Zen 4. Due to the popularity of those articles, we’ve now repeated the same tests on the new Zen 5 architecture, for comparison with previous cores. Read more “How much does AVX-512 help Zen 5 in x265? And how to turn it on” »

  •  
  •  
  •  

Leave a Reply

Your email address will not be published. Required fields are marked *