A few things that I noticed from my very quick look at your results.
First, this is a JMP community, so a JMP file would be preferable over an Excel file. After all, the first thing that I did is put the data into JMP so that I could see it! Plus, you could have saved your different models as scripts that could be re-run in order to see the options you tried.
I do not know how you implemented SMOTE+TOMEK to balance the responses, but there are other ways you could attempt to balance the responses. You could have oversampled, ignored some of the more prevalent cases, assign rows to training/validation/test based a stratified sampling of the response, etc. Regardless, if you are concerned by the "perfect" model, could these models be picking up on how the balancing was done? If an algorithm is doing the balancing, a model/algorithm just might be able to explain it.
Assuming that the balancing cannot affect the modeling effort, you had split your data into training, validation, and test. Your test set should be indicative of real-world performance, so I do not believe there is any over-fitting going on. The test set is not used at all in the fitting process. If you believe the model is too good, then make sure that your data (and especially the test set) is representative of your process.
There are 89 rows without a validation value. Why? Are you excluding those rows from your analysis? Why?
It looks like you used all of the default values for a boosted tree. Notice that a default boosted tree has 50 layers and your final results show that all 50 layers were used. That means that you can have a better tree by using more boosted layers. Here is a result when I specified a very high number of layers.
The support vector machine also provided a perfect fit according to misclassification rate. Why not consider that?
The only way to truly have confidence in a model is to start using it on new data. Why not go forward with all three (more layered boosted tree, neural net, and support vector machines) and let the data determine which one truly works better? Do you need to consider the speed of model evaluation in determining which one you use?
Dan Obermiller