There is no easy answer to this situation since fractional frame rates are deeply embedded into current production workflows, camera hardware, and the ASICs (Application Specific Integrated Circuits) that make those cameras.
Some current-generation professional cameras incur tradeoffs when shooting at 24.0 fps. This may include focus behavior, aspect ratio when shooting at high frame rate, etc. E.g, my documentary team has several fully-built Sony FX6 rigs and the Ronin R4D. As configured, those are all $10k cameras. Yet they have some narrow feature limitations if shooting at 24.0. Even the $25k Sony Burano has certain restricted features on 24.0 fps.
If you are a solo operator, you can do whatever you want. But if you are shooting as part of a team (or might be in the future), there are issues of commonality, standardization and collaborative post-production efficiency. Even with a new-generation camera fleet, having them all on 23.98 avoids certain problems.
A big issue is intermixing 23.98 and 24.0 material. Since those clips must be rate-conformed, on a 23.98 timeline, the NLE will force the 24.0 clip to run faster to match frame-by-frame. On a 24.0 timeline, the 23.98 clips will run slightly slower. That means the audio from those clips won't stay in sync with each other or with external audio recorders.
My team regularly leverages the hundreds of terabytes of 23.98 material we've shot over the past 10 years. Even if all our cameras worked perfectly at 24.0, we would then be facing a constant issue of editing mixed 23.98 and 24.0 material. For us, it's easier and safer to just keep shooting 23.98.
If all our pro cameras could shoot at 24.0 fps with no tradeoffs, and if we had no issues with 23.98 archival material, we still might not use 24.0. That's because lots of cameras today simply cannot shoot 24.0. We prioritize using the pro cameras, but we often are forced to use material from mirrorless cameras that can only shoot 23.98 fps. We don't want the hassle of constantly editing mixed frame-rate material except for unavoidable cases.
The issues about measuring real-world playout time on fractional frame rate material are well known and are unrelated to FCP. Even back in the days of FCP 7, the manual described these in Appendix D:
"Remember that these [timecode] numbers don't reflect time; they are simply unique identifiers. The first frame of NTSC video is labeled 00:00:00:00. The 29th frame is labeled 00:00:00:29, and the 30th frame is labeled 00:00:01:00. Again, just because a frame is labeled 00:00:01:00 does not mean that 1 second has passed. The frame could just as easily have been named AAABD, in which case there would be no temptation to read the label as a time value. Only the frame rate of the video can determine how much time has passed by the 30th frame."
"Timecode is merely a method of labeling frames with unique identifiers to easily find them again later. It is a convenient way of giving each frame a name that can be referred to later without having to verbally describe and visually search for it. Even though frame rate and timecode are independent, people commonly confuse the two, which can lead to frustrating problems in post-production."
"The Difference Between Frame Rate and Timecode:
The frame rate of your film or video describes how rapidly frames are photographed or played back. It refers to the physical speed of image capture and playback. Timecode is merely a method of labeling frames with unique identifiers to easily find them again later. It is a convenient way of giving each frame a name that can be referred to later without having to verbally describe and visually search for it. Even though frame rate and timecode are independent, people commonly confuse the two, which can lead to frustrating problems in post-production. Before you start a project, be certain that you understand the difference between these two terms." (emphasis added)