First, welcome to the community. There are a couple of issues with the information you present:
1. You don't have an estimate of the within rater variation, intrarater reliability (e.g., what if you were to give the same case to the same rater without their knowledge and had them rate the case again. Would they rate it the same?)
2 The variation between raters, interrater reliability. Since the cases are different for the different raters, you won't be able to discriminate between whether their agreement (or lack there of) was due to rater or case. The variation due to rater and case are confounded.
You might find this an interesting read:
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3900052/
"All models are wrong, some are useful" G.E.P. Box