Methodology: performance tests
It’s the fastest Core i3 yet, but it’s also the hungriest. The 14100F’s (Raptor Lake Refresh) biggest competition in its own ranks is in the form of older models (13100F and 12100F). These are a bit slower, but lower-power. The “better” choice depends on what holds more weight on your scales. Maybe it will be that record-breaking speed? In this class (Core i3), power consumption is always relatively low.
Gaming tests
We test performance in games in four resolutions with different graphics settings. To warm up, there is more or less a theoretical resolution of 1280 × 720 px. We had been tweaking graphics settings for this resolution for a long time. We finally decided to go for the lowest possible (Low, Lowest, Ultra Low, …) settings that a game allows.
One could argue that a processor does not calculate how many objects are drawn in such settings (so-called draw calls). However, with high detail at this very low resolution, there was not much difference in performance compared to FHD (which we also test). On the contrary, the GPU load was clearly higher, and this impractical setting should demonstrate the performance of a processor with the lowest possible participation of a graphics card.
At higher resolutions, high settings (for FHD and QHD) and highest (for UHD) are used. In Full HD it’s usually with Anti-Aliasing turned off, but overall, these are relatively practical settings that are commonly used.
The selection of games was made considering the diversity of genres, player popularity and processor performance requirements. For a complete list, see Chapters 7–16. A built-in benchmark is used when a game has one, otherwise we have created our own scenes, which we always repeat with each processor in the same way. We use OCAT to record fps, or the times of individual frames, from which fps are then calculated, and FLAT to analyze CSV. Both were developed by the author of articles (and videos) from GPUreport.cz. For the highest possible accuracy, all runs are repeated three times and the average values of average and minimum fps are drawn in the graphs. These multiple repetitions also apply to non-gaming tests.
Computing tests
Let’s start lightly with PCMark 10, which tests more than sixty sub-tasks in various applications as part of a complete set of “benchmarks for a modern office”. It then sorts them into fewer thematic categories and for the best possible overview we include the gained points from them in the graphs. Lighter test tasks are also represented by tests in a web browser – Speedometer and Octane. Other tests usually represent higher load or are aimed at advanced users.
We test the 3D rendering performance in Cinebench. In R20, where the results are more widespread, but mainly in R23. Rendering in this version takes longer with each processor, cycles of at least ten minutes. We also test 3D rendering in Blender, with the Cycles render in the BMW and Classroom projects. You can also compare the latter with the test results of graphics cards (contains the same number of tiles).
We test how processors perform in video editing in Adobe Premiere Pro and DaVinci Resolve Studio 17. We use a PugetBench plugin, which deals with all the tasks you may encounter when editing videos. We also use PugetBench services in Adobe After Effects, where the performance of creating graphic effects is tested. Some subtasks use GPU acceleration, but we never turn it off, as no one will do it in practice. Some things don’t even work without GPU acceleration, but on the contrary, it’s interesting to see that the performance in the tasks accelerated by the graphics card also varies as some operations are still serviced by the CPU.
We test video encoding under SVT-AV1, in HandBrake and benchmarks (x264 HD and HWBot x265). x264 HD benchmark works in 32-bit mode (we did not manage to run 64-bit consistently on W10 and in general on newer OS’s it may be unstable and show errors in video). In HandBrake we use the x264 processor encoder for AVC and x265 for HEVC. Detailed settings of individual profiles can be found in the corresponding chapter 25. In addition to video, we also encode audio, where all the details are also stated in the chapter of these tests. Gamers who record their gameplay on video can also have to do with the performance of processor encoders. Therefore, we also test the performance of “processor broadcasting” in two popular applications OBS Studio and Xsplit.
We also have two chapters dedicated to photo editing performance. Adobe has a separate one, where we test Photoshop via PugetBench. However, we do not use PugetBench in Lightroom, because it requires various OS modifications for stable operation, and overall we rather avoided it (due to the higher risk of complications) and create our own test scenes. Both are CPU intensive, whether it’s exporting RAW files to 16-bit TIFF with ProPhotoRGB color space or generating 1:1 thumbnails of 42 lossless CR2 photos.
However, we also have several alternative photo editing applications in which we test CPU performance. These include Affinity Photo, in which we use a built-in benchmark, or XnViewMP for batch photo editing or ZPS X. Of the truly modern ones, there are three Topaz Labz applications that use AI algorithms. DeNoise AI, Gigapixel AI and Sharpen AI. Topaz Labs often and happily compares its results with Adobe applications (Photoshop and Lightroom) and boasts of better results. So we’ll see, maybe we’ll get into it from the image point of view sometime. In processor tests, however, we are primarily focused on performance.
We test compression and decompression performance in WinRAR, 7-Zip and Aida64 (Zlib) benchmarks, decryption in TrueCrypt and Aida64, where in addition to AES there are also SHA3 tests. In Aida64, we also test FPU in the chapter of mathematical calculations. From this category you may also be interested in the results of Stockfish 13 and the number of chess combinations achieved per unit time. We perform many tests that can be included in the category of mathematics in SPECworkstation 3.1. It is a set of professional applications extending to various simulations, such as LAMMPS or NAMD, which are molecular simulators. A detailed description of the tests from SPECworkstation 3.1 can be found at spec.org. We do not test 7-zip, Blender and HandBrake from the list for redundancy, because we test performance in them separately in applications. A detailed listing of SPECWS results usually represents times or fps, but we graph “SPEC ratio”, which represents gained points—higher means better.
Processor settings…
We test processors in the default settings, without active PBO2 (AMD) or ABT (Intel) technologies, but naturally with active XMP 2.0.
… and app updates
The tests should also take into account that, over time, individual updates may affect performance comparisons. Some applications are used in portable versions, which are not updated or can be kept on a stable version, but this is not the case for some others. Typically, games update over time. On the other hand, even intentional obsolescence (and testing something out of date that already behaves differently) would not be entirely the way to go.
In short, just take into account that the accuracy of the results you are comparing decreases a bit over time. To make this analysis easier for you, we indicate when each processor was tested. You can find this in the dialog box, where there is information about the test date of each processor. This dialog box appears in interactive graphs, just hover the mouse cursor over any bar.
- Contents
- Intel Core i3-14100F in detail
- Methodology: performance tests
- Methodology: how we measure power draw
- Methodology: temperature and clock speed tests
- Test setup
- 3DMark
- Assassin’s Creed: Valhalla
- Borderlands 3
- Counter-Strike: GO
- Cyberpunk 2077
- DOOM Eternal
- F1 2020
- Metro Exodus
- Microsoft Flight Simulator
- Shadow of the Tomb Raider
- Total War Saga: Troy
- Overall gaming performance
- Gaming performance per euro
- PCMark and Geekbench
- Web performance
- 3D rendering: Cinebench, Blender, ...
- Video 1/2: Adobe Premiere Pro
- Video 2/2: DaVinci Resolve Studio
- Graphics effects: Adobe After Effects
- Video encoding
- Audio encoding
- Broadcasting (OBS and Xsplit)
- Photos 1/2: Adobe Photoshop and Lightroom
- Photos 2/2: Affinity Photo, Topaz Labs AI Apps, ZPS X, ...
- (De)compression
- (De)encryption
- Numerical computing
- Simulations
- Memory and cache tests
- Processor power draw curve
- Average processor power draw
- Performance per watt
- Achieved CPU clock speed
- CPU temperature
- Conclusion
Thank you for the article. I have been looking for 14th gen non-K cpu reviews.
Do you have an explanation why does 14100 take so much more power at idle, compared to 13100? In the other power graphs too, 13100 seems to be an outlier… I expected 14100 to be basically the same cpu, just being produced using a tad better refined process, and the clocks whipped up a bit.
More aggressive clock speed management. The Core i3-14100F does not go to 400 MHz like the Core i3-13100F (although the working range of the multiplier should be the same). I’m not saying it never does, but not at the level of our load corresponding to “idle”. And it won’t be on the edge either, nothing changed by terminating some processes (for example launchers) in the background, after which less load is put on the processor. Sometimes, under the same conditions, the Ci3-14100F doesn’t underclock as aggressively (as the Ci3-13100F). I can’t note what this is related to, but it might have something to do with a more aggressive TB (2.0), which makes the processor run at higher clock speeds even at very low load.
Sounds plausible.
Did you re-test 12100 and 13100 using the same exact OS version? I am also thinking if the silicon lottery may play a role, I have seen a test of multiple cpus of the exact same model, and the results were somewhat divergent. (Do not remember the source, I think it was Der8auer.)
No, we haven’t re-tested the Ci3-12100F and Ci3-13100F, but we are still using Windows 10 (22H2). For the reason that it does not change as dynamically as W11 and is therefore more suitable for building a massive database. It would never be possible to make with the kind of tests we do here otherwise.
I can assure you that it is definitely not about “silicon lottery”. I find Der8baer’s tests very irresponsible and unreliable. To accurately analyze such things, a controlled, consistent testing environment is essential, which he does not have. This is extremely important for processors such as the Ryzen 5 7600, which are particularly sensitive to temperature changes. Rather than real differences between processors, his results are more a reflection of how ambient conditions change during individual tests. These include, among other things, the mounting of the cooler on the processor, which is defined by an always equal heat transfer from the processor to the cooler. This is also very difficult to achieve.
We also used to deal with the dependence of the cooling performance on the different techniques of applying thermal paste and its different quantities. And also the influence of different pressure. All this is necessary to control in order to analyze the properties across different pieces of the same processor. It doesn’t seem that these things are any special concern of the author testing with the motherboard “installed” on its box and components randomly spread on the table. 🙂