I completely agree with @cwillden's conclusion about your analysis. You are not performing a Gage R&R or any kind of Measure System Analysis (MSA) in this case. You cannot judge the performance of the tool by itself. The same tool might provide very good measurements in one situation and poor ones in another. The difference? What is it expected to distinguish? That is based on the amount of part variation.
Where did your get research information? I have never seen any presentation of Gage R&R in the absence of part variation. (I admit that it might exist, though.)
What is the basis given for an upper bound of 30% EV? Is it relative to the total variation? Your EV is over 80%:
Such 'rules of thumb' are often not meaningful or helpful in a particular situation. For example, if my EV is high but the measurement is cheap (e.g., cost, time), I can determine the number of replicate measurements to average and reduce this component. On the other hand, if my EV is high and the measurement is expensive, then I might have to find the root source of this variation and reduce or eliminate it or find another tool.