For the two critical values, we determined absolute agreement (z.B Liao et al., 2010) as a percentage of statistically non-different evaluations. The absolute approval rate was 100% if we take into account the CIR calculated on the basis of CCI for our sample. On the other hand, the absolute approval rate was 43.4% if the reliability of the test manual was used to estimate the critical difference. With this more conservative measure of absolute agreement, the probability of obtaining a consistent rating was not too fortuitous. This probability did not differ statistically for the two scoring subgroups (parents-teachers and parent-father-assessments) and therefore on the entire study population, regardless of the calculation of the RCI chosen. These results support the hypothesis that the parents and teachers of the daycare were, in this case, equally competent counsellors with regard to the children`s early expressive vocabulary. Nevertheless, RCI, which was identified by different reliability estimates, differed significantly from specific estimates of absolute compliance. The very divergent amounts of absolute agreement obtained through the use of the reliability of the Inter-Rater division within a relatively small sample or the reliability of the instrument test, obtained with a larger and more representative sample, underline the need for caution in calculating reliable differences. Figure 2. Comparison of Interrater`s reliability.

Intra-class correlation coefficients (CCI presented as points) and confidence intervals corresponding to α -0.05 (CN, presented as error bars) for parent-teacher evaluations, mother-father evaluations, and all evaluation pairs in the subgroups of spleeners. The overlapping IABs suggest that CCIs did not systematically differ. the variance of evaluations between children is 2, with variance being in the subordinate elements and k the number of spleens. Confidence intervals for all CCIs have been calculated to determine if they flatten each other. Measuring the reliable difference between ratings based on Inter-Rater`s reliability in our study led to a 100% rating agreement. On the other hand, a considerable number of different evaluations were identified in the RCI calculation based on the more conservative reliability of the manual tests; The absolute approval rate was 43.4%. The use of this conservative RCI estimate did not result in a significantly higher number of identical or divergent assessments, either for a single rating subgroup, or for the entire population studied. (see Table 2 for the results of the corresponding binomial tests). Therefore, the probability of a child receiving a matching assessment was no different from chance.

When the reliability of the study was used, the probability of obtaining correlated ratings was 100%, which is significantly higher than random. Brown, J. D., Wissow, L. S., Gadomski, A., Zachary, C., Bartlett, E., and Horn, I. (2006). Evaluation of children`s mental health that focuses on primary care services: Interrater agreement and impact on mental health screening. ambulatory. Pediatr. 6, 347-351. doi: 10.1016/j.ambp.2006.09.004 While the correlation analyses used (mostly pearson correlations) provide information on the strength of the relationship between two groups of values, they do not capture the agreement between the seats at all (Bland and Altman, 2003; Kottner et al., 2011).