Next Article in Journal
The Interaction of Vitamin D and Corticosteroids: A Mortality Analysis of 26,508 Veterans Who Tested Positive for SARS-CoV-2
Previous Article in Journal
Relationships between Risk Events, Personality Traits, and Risk Perception of Adolescent Athletes in Sports Training
Previous Article in Special Issue
Movement and Physical Activity in Early Childhood Education and Care Policies of Five Nordic Countries
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Inter-Rater and Intra-Rater Reliability of the Actual Aquatic Skills Test (AAST) for Assessing Young Children’s Motor Competence in the Water

1
Research Unit on Movement and Nutrition for Health and Performance, Department of Movement and Sport Sciences, Faculty of Physical Education and Physiotherapy, Vrije Universiteit Brussel, Pleinlaan 2, 1050 Brussels, Belgium
2
Faculty of Sport and Health Sciences, University of Jyväskylä, P.O. Box 35 (L), FI-40014 Jyväskylä, Finland
*
Author to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2022, 19(1), 446; https://0-doi-org.brum.beds.ac.uk/10.3390/ijerph19010446
Submission received: 28 October 2021 / Revised: 18 December 2021 / Accepted: 22 December 2021 / Published: 31 December 2021
(This article belongs to the Special Issue Physical Education in Early Childhood)

Abstract

:
As children’s actual aquatic skills are important for the prevention of drowning as well as their engagement in lifelong aquatic physical activity, researchers and practitioners should be able to assess this vital concept accurately and reliably. Therefore, this study aimed to investigate the inter-rater and intra-rater reliability of the Actual Aquatic Skills Test (AAST), consisting of 17 different test items for the assessment of young children’s motor competence in the water. Six raters received a training and evaluation session on scoring the AAST, after which five of them assessed four test videos (of various children (n = 38) performing the test items) twice, with one to two weeks in between (i.e., test and re-test). Inter-rater and intra-rater reliability were determined per test video and for the different AAST test items across videos using Gwet’s Agreement Coefficient 2 (Gwet’s AC2). The Gwet’s AC2 for inter-rater reliability at the test varied from 0.414 to 1.000, indicating a moderate to perfect agreement between raters. For intra-rater reliability, it ranged from 0.628 to 1.000, demonstrating a good to perfect agreement between test and re-test scoring. In conclusion, the AAST is a promising tool to reliably assess young children’s actual aquatic skills in an indoor swimming pool.

1. Introduction

The concept of ‘water competence’ was first introduced and described by Langendorfer & Bruya (1995), focusing on fundamental motor skills in the water serving as the basis for further aquatic development [1]. Subsequently, Moran (2013) considered the concept of water competence from a drowning prevention perspective, defining it as “the sum of all personal aquatic movements that help prevent drowning, as well as the associated water safety knowledge, attitudes and behavior that facilitate safety in, on and around the water” [2]. More recently, Stallman et al. (2017) made the definition of the term water competence more tangible through their evidence-based proposal of 15 specific aquatic competencies, including physical, cognitive, as well as affective competencies [3]. In this paper, our focus will specifically be on the physical aspect within the concept of water competence, hereafter referred to as children’s actual aquatic skills.
Drowning is the world’s third leading cause of injury-related death, claiming more than 320,000 lives per year [4,5]. As to fatal and non-fatal drownings, young children are the most vulnerable age group, with one to four-year-olds and five to nine-year-olds experiencing the highest and second highest risk of drowning, respectively [4,5]. The World Health Organization (WHO) (2017 and 2020) highlights water safety education as one of the key approaches to prevent drowning amongst children [4,6]. The value of aquatic education as a preventive measure against drowning has also been supported by two-large scale retrospective studies conducted in different continents, comparing the aquatic educational backgrounds of young children who were victims of fatal drownings against that of a demographically equivalent control group [7,8]. Although these studies did not investigate or report the content of swimming lessons as well as the protective value of specific aquatic skills, they both indicate that aquatic education has the ability to reduce young children’s risk of drowning [7,8].
Numerous studies concerning land based motor skill performances already demonstrated a clear positive relationship between children’s level of motor competence and both their health-related physical fitness and physical activity levels [9,10,11,12]. This beneficial association between developing foundational skills and physical activity across the lifespan might also be valid in an aquatic environment, because an adequate aquatic skill level would enable children to engage in a broad array of aquatic activities and water-based recreation and sports, both now and in the longer term [13,14,15,16,17,18]. As children’s actual level of aquatic skills may not only be important for their water safety but also for their general health and well-being through engagement in aquatic oriented physical activity, it is crucial that both researchers and practitioners (e.g., swimming instructors and physical education teachers) are able to assess this vital concept accurately and reliably.
In addition, swimming instructors and physical education teachers should be able to clearly communicate young children’s actual level of aquatic skills and its importance to the child(ren)’s parents or primary caregivers, since this awareness might influence the degree of supervision deemed necessary when their child is being active in, on or around water [19]. A lack of supervision is a known risk factor for drowning [20]. Being correctly informed and aware of their child(ren)’s actual aquatic skill level, parents can correspondingly stimulate and support their child(ren)’s engagement in aquatic activities in view of pursuing a physically active lifestyle, considering that parental support is an important factor for children’s physical activity engagement [21,22]. Finally, lifeguards must also have the ability to correctly estimate or determine children’s actual aquatic skill levels in view of guaranteeing their water safety.
A promising new tool for assessing young children’s actual aquatic skill level that has yet to be proven reliable and valid, is the Actual Aquatic Skills Test (AAST), first administered by D’Hondt et al. (2021) [23]. The AAST [23] was based on transferring the 17 aquatic skills or test items included in the Pictorial Scale for Perceived Water Competence (PSPWC) [24] to an actual performance by the child in the water of an indoor swimming pool. Each of these 17 aquatic skills has to be assessed based on three possible levels of execution. For each level, a pictogram and corresponding performance criteria have been described in the PSPWC manual of Morgado et al. (2020) [24]. The derived AAST [23] is a tool worth investigating for a number of reasons, as elaborated below.
First, to the best of our knowledge and also according to other authors [23,25,26], there are currently only a few validated instruments available for assessing (pre)school children’s actual aquatic skills, which is in sharp contrast to the wide variety of assessment tools available for evaluating their actual level of motor competence on land [25,27,28,29]. To date, the majority of research investigating water competence in children, and more specifically their aquatic skills, is based on self-reported or parent proxy-reported estimates thereof [3,30,31,32]. This approach is generally found to be less labor-intensive and time-consuming than actually assessing children’s aquatic skill level in the water [30,33,34]. Nonetheless, it is known that both children’s self-perception as well as the parental perception of children’s level of (aquatic) skills are not always in accordance with reality [23,35,36,37]. Estevan et al. (2017) [36], for example, only found weak to moderate positive relationships between children’s actual skill level (on land) and their self-perception as well as the parental perception and the perception of physical education teachers, with the children themselves being the most inaccurate when estimating their own skill level. Additionally, in practice, the evaluation of aquatic skills by teachers or coaches often refers to their experience, if not their intuition [29].
Second, the main and often sole determinant of children’s actual aquatic skill level considered in previous studies, but also by most swimming instructors, lifeguards and parents or caregivers, is the ability of a child to swim a certain distance [38]. Although propulsion in the water is a fundamental aquatic skill, the evaluation of propulsion alone is inadequate in terms of assessing a child’s water safety [3,30,38]. Mastering propulsion does not guarantee that the child is able to breathe, rest, call for help or climb out of the water [3,39,40]. Furthermore, Stallman (2008) [41] identified the underestimation of possible danger and unexpected occurrences, such as cold water, a strong current or falling from great height, as common causes of drowning.
Third, the content of the AAST [23] aligns perfectly with that of the PSPWC [24], allowing research such as that of D’Hondt and colleagues [23], who examined the relationship between the actual and the self- and parent-perceived aquatic skill level of children. According to Stallman et al. (2017) [3], a vital element of water competence is the individual’s ability to accurately estimate their own actual aquatic skill level. More specifically, one’s self-perceived aquatic skill level has to correspond closely to one’s actual skill level in order to appropriately judge potential risks and constraints when being in, on and around water. As such, children who overestimate their actual aquatic skill level may have a higher chance of drowning [3,42,43,44,45]. Moreover, children who underestimate their actual aquatic skill level might avoid water and participation in aquatic activities, therefore losing out on chances to be physically active in this particular setting or movement environment. When swimming instructors and physical education teachers are able to map an individual child’s actual and perceived aquatic skill profile, they can counter these potential threats and/or inform the parents or caregivers about them. In order to investigate the relationship between one’s actual and perceived level of aquatic skills, an alignment between test instruments measuring both constructs is recommended [46,47]. However, only a few of the scarce studies examining and comparing children’s actual and perceived motor competence in the water used aligned instruments, since there are currently few (aligned) test instruments available in this regard [23,37,48].
Furthermore, D’Hondt et al. (2021) already demonstrated a good internal consistency of the 17 test items included in both the PSPWC [24] and the AAST [23]. Besides, given that the AAST [23] has been derived from a pictorial scale questionnaire developed for children, the evaluation of the different test items is fairly uncomplicated. Additionally, the pictograms included in the PSPWC [24] may facilitate an easy understanding and administration of the AAST [23] in areas with low literacy levels.
Consequently, the AAST [23] can be considered a promising tool for assessing the motor competence of young children in the water. Therefore, the purpose of this study is to investigate both the inter-rater and intra-rater reliability of the AAST [23] for the assessment of young children’s actual aquatic skill level (using video footage).

2. Materials and Methods

Our research approach was approved by the local Medical Ethics Committee at VUB (B.U.N. 143201942643) on 16 September 2020.

2.1. Raters

In total, six participant raters (i.e., one female and five males, aged 21–24 years) were conveniently and voluntarily recruited from students enrolled in the Master in Movement and Sport Sciences program of the Vrije Universiteit Brussel (VUB, Belgium) to provide AAST [23] evaluations. Study participants were excluded if they were not able to attend an organized training session for raters. As a part of obtaining their Bachelor’s degree in Physical Education and Movement Sciences, all of the recruited participant raters (n = 6) followed weekly one-hour swimming classes themselves for a total of five consecutive semesters, with each semester consisting of 13 class weeks. In addition, all of the participant raters, with the exception of rater F, had experience teaching at least 10 swimming classes. Furthermore, one of them was a certified physical education teacher (i.e., rater B). Another participating rater completed a specialized water safety course, achieving his lifeguard license (i.e., rater E). Lastly, a third rater was a former competitive swimmer, and was still an active swimming coach during the conduct of this reliability study (i.e., rater D).

2.2. Subjects

Due to COVID-19 and the subsequent closure of public swimming pools during the envisaged data collection period, live administration of the Actual Aquatic Skills Test (AAST) [23] with children in the water of an indoor swimming pool was not possible. Therefore, the raters were asked to assess previously obtained video recordings of different children performing the 17 aquatic skill test items included in the AAST [23]. These recordings were grouped into multiple composite videos (n = 8), using the Shotcut video editing application [49], including three composite videos for the purpose of training the raters and familiarizing them with the testing protocol (i.e., training videos), one composite video to evaluate the accuracy of the raters against experts (i.e., evaluation video) and four composite videos for determining the inter-rater and intra-rater reliability of the AAST (i.e., test video 1, test video 2, test video 3 and test video 4).

2.3. Procedure

The video recordings, employed for this investigation, were originally obtained for the purpose of a previous study of our research group [23], for which 134 children between six and nine years of age were recruited through convenience sampling, in collaboration with multiple swimming schools located in Flanders and the Brussels Capital Region (Belgium), by either addressing their parent(s) or guardian(s) personally at the swimming pool or via e-mail. These swimming schools were intentionally selected because of their focus on developing children’s fundamental and survival aquatic skills before teaching them competitive swimming strokes. The live administration and video recording of the AAST [23] performed by these children took place under the supervision of one of the main researchers who was also an experienced swimming teacher. Subsequently, all collected video recordings were carefully screened by the lead author of this reliability study. Bad quality recordings and recordings in which the child gave the impression not to understand the AAST task(s) [23], were excluded in view of composing the training, evaluation and test videos.
Each composite video contained the test instructions and 17 included recordings (i.e., one for every test item of the AAST [23]), applying the same order of test items as presented in the PSPWC manual [24]. Every test item within the evaluation video and each composite test video was performed by a different child (i.e., 17 different children per composite video or set of recordings). In total, for the four composite test videos, video recordings of the performances of 38 different children were used. A white screen (lasting for 4.0 s) always separated the performance of two subsequent tasks or test items. This was the only moment where the raters were allowed to pause the video, in order (1) to be able to (re)check the task criteria in the PSPWC manual [24], (2) to fill in the provided score-sheet (Table 1), and (3) to prepare themselves for assessing the next test item performed by a different child. Under no circumstances were raters allowed to fast-forward or rewind any parts of the video. Every recorded task performance of each child per test item was shown twice in the composite video with these identical performances being divided by a black screen (lasting for 1.5 s). The scores to be awarded and the associated scoring criteria were exactly the same as described in the PSPWC manual [24]. That is, ‘1’—the child is not able to perform the aquatic skill, ‘2’—the child is partly able to perform the aquatic skill and thus still in progress, and ‘3’—the child is able to correctly perform the aquatic skill. Table 2 provides an overview of the three different mastery levels for each of the 17 aquatic skills of the PSPWC [24]. In the PSPWC manual [24], the minimum criteria for each mastery level of every aquatic skill or test item are described in more detail (see Table 3). The AAST assessment used in this reliability study was based upon these minimum criteria. In order to achieve a score of ‘3’ on a certain aquatic skill or test item, the child’s performance had to meet all the criteria described for level 3 of that particular test item. Consequently, when one or more criteria were not fulfilled, a score of ‘2’ was granted, provided that all the criteria for level 2 were met. When this was not the case, the child’s performance was to be awarded a score of ‘1’.

2.4. Training

Prior to the actual video assessments in the context of this reliability study, all participating raters received a two-hour online training session (in separate groups of three), provided by the lead investigator. During this online training session, the PSPWC [24] and the AAST [23] testing protocols were explained in detail. Consequently, the raters independently assessed a total of three different composite videos (i.e., training videos), each of which contained the 17 PSPWC/AAST test items [23,24] performed by different children in the water of an indoor swimming pool. The raters were urged to read the minimum criteria per score level for each aquatic skill task in the PSWPC manual [24] both before and after watching a child’s performance on a specific test item to be assessed. After individually scoring the 17 performances within one composite training video, the differences in scoring between the lead investigator and the raters as well as between the raters themselves were discussed after each video assessment until consensus was reached. When the scoring criteria were not applied correctly (e.g., a rater assigning a score of 3 to a child’s performance for a particular aquatic skill and at the same time stating that a certain criterion for level 3 of that test item was not met), the correct assessment method was explained again.

2.5. Evaluation

One week after the online training session, the participating raters received the so- called evaluation video. In order to further progress and participate in this reliability study as a rater, participants had to establish an 80% agreement with the scores awarded by international experts in the field, which is in accordance with the criterion used in similar reliability studies investigating land based motor skill assessment tools [50,51]. These experts were four co-developers of the PSPWC [24], who also assessed the same evaluation video or set of recorded test items performed by 17 different children. These four international experts awarded the very same score to 11 out of 17 child performances on the AAST [23] (i.e., test items 1, 2, 3, 4, 5, 7, 9, 11, 12, 13 and 14). Regarding four other aquatic skill performances (i.e., test items 6, 8, 15 and 17) their opinions were divided, with three experts opposed to one. In this case, the score awarded by the majority of the experts was used as the ‘gold standard’. When the opinions were split, two experts opposed to the other two (i.e., test items 10 and 16), there was a group discussion until consensus was reached or until the majority agreed on the final score to be awarded. All six participant raters, who followed the online training session for AAST [23] assessment (i.e., rater A, B, C, D, E and F), achieved an accuracy above 80% against the experts regarding the scoring of the evaluation video (i.e., 94% for rater A and C, 88% for rater B, C and D, and 82% for rater F). The only female rater (i.e., rater F) dropped out with regard to the next study phase due to personal matters. Therefore, one week after successfully completing the evaluation video, five remaining raters individually assessed the four test videos within a timeframe of one week after receiving the video footage (i.e., test). After another week in which the raters were not allowed to look at or score any of the videos, these raters were asked to score the same four test videos a second time, again being provided with a one-week timeframe to complete this task (i.e., re-test). This minimum test-retest interval of one week is a commonly used in reliability studies of motor skill assessment tools [52].

2.6. Statistical Analysis

Both inter-rater and intra-rater reliability were determined for each of the four test videos and separately for each individual test item across the four composite videos. Inter-rater reliability was solely determined for the test, and thus not for the re-test. To investigate both the inter-rater and intra-rater reliability of the ordinally scored AAST [23] data, the Gwet’s Agreement Coefficient 2 (Gwet’s AC2) [53] was determined by means of the Real Statistics Resource Pack for Microsoft Excel [54]. The use of Gwet’s AC2 was a deliberate decision, given its resistance to an unbalanced distribution of data across the scoring categories in contrast to other reliability measures [55,56]. Gwet’s AC2 values can range from 0 to 1, where 0 represents absolute disagreement and 1 represents perfect agreement. In accordance with the classification used by Thoenen et al. (2021) [57], Gwet’s AC2 values below 0.20 were considered to represent a poor level of agreement, whereas Gwet’s AC2 values ranging between 0.21–0.40, 0.41–0.60, 61–0.80 and 0.80–1.00 were considered to demonstrate a fair, moderate, good and very good level of agreement, respectively. Additionally, for inter-rater reliability, the scoring distribution for each test item per test video was reported and expressed as the percentage of raters per awarded score presented in an agreement table. For intra-rater reliability, the percent agreement between test and re-test scoring was also reported for each rater (i.e., the percentage of test items that were awarded the exact same score for the test and the re-test on an individual rater basis).

3. Results

3.1. Inter-Rater Reliability

As demonstrated in Table 4, Gwet’s AC2 values for inter-rater reliability of the test videos varied between 0.646 and 0.922, demonstrating a good level of agreement for test videos 1, 2 and 3, whereas test video 4 showed a very good level of agreement between the different raters.
Regarding the inter-rater reliability per test item across the four test videos, Gwet’s AC2 values ranged from 0.414 (i.e., moderate agreement) to 1.000 (i.e., perfect agreement). Test items 1, 2, 4 and 9 demonstrated perfect agreement among all raters. A very good level of agreement was found for test items 8, 11, 15 and 17, whereas test items 5, 7, 10, 12, 13, 14 and 16 showed good inter-rater reliability. A moderate agreement was found for test items 3 and 6. The scoring distribution for each aquatic skill task per test video, expressed as the percentage of raters per awarded score, is presented in Table 5.

3.2. Intra-Rater Reliability

The Gwet’s AC2 values and the percent agreements for the intra-rater reliability are presented in Table 6 and Table 7, respectively. The intra-rater reliability across the four test videos was found to be very good for all raters, with rater B obtaining the highest level of agreement, followed by rater A, C, D and E, respectively. The Gwet’s AC2 values for the intra-rater reliability per test video varied between 0.834 (i.e., very good agreement) and 1.000 (i.e., perfect agreement). The intra-rater reliability per test item ranged from 0.628 (i.e., good agreement) to 1.000 (i.e., perfect agreement). All raters achieved perfect intra-rater reliability for test items 2, 4, 11, 14, 15 and 17. For test items 1, 6, 7, 8, 12 and 16, four out of the five raters achieved perfect agreement, whereas the remaining rater always achieved a good level of agreement between test and re-test scoring. For test items 3, 5, 9, 10 and 13, three out of the five raters obtained a Gwet’s AC2 of 1.000, indicating a perfect agreement between test and re-test scoring, whereas the remaining two raters always obtained a good level of agreement in terms of their intra-rater reliability. In other words, rater A and rater B demonstrated perfect agreement between test and re-test scoring for all but two test items of the AAST [23]. Rater C and D obtained perfect intra-rater reliability for 14 out of the 17 aquatic skills, whereas rater E obtained perfect intra-rater-reliability for 11 out of the 17 AAST items [23].

4. Discussion

The purpose of this study was to assess both the inter-rater and intra-rater reliability of the AAST [23], a test instrument for assessing young children’s actual aquatic skill level based on transferring the 17 test items from the PSPWC [24] to an actual performance of those aquatic skill tasks in the water of an indoor swimming pool.
Regarding inter-rater reliability, all composite test videos demonstrated a good to very good agreement. When looking at the inter-rater reliability of the individual test items across videos, four out of the 17 test items (i.e., test items 1,2, 4 and 9) demonstrated perfect agreement, another four test items demonstrated a very good level of agreement (i.e., test items 8, 11, 15 and 17) and six test items (i.e., test items 5, 7, 10, 12, 13,14 and 16) demonstrated a good level of agreement. It must be noted that all raters awarded the very same score (i.e., score 3) to all child performances of test item 2 (i.e., standing and submersion in the water), indicating that there was little to no variation in children’s level of skill performance for this particular aquatic skill across the test videos to be scored. The lowest level of agreement (i.e., moderate agreement) was found for test items 3 and 6. The moderate agreement for test item 3 (i.e., blowing bubbles under water) resulted predominantly from the stricter assessment of one rater (i.e., rater E) when compared to the other four raters. Rater E awarded a lower score to the performance of test item 3 in test videos 1 and 2. For test video 3, both rater A and rater E awarded a lower score to the child’s performance when compared to the other three raters. This degree of disagreement between the raters regarding test item 3 was likely caused by the use of two-dimensional video recordings taken from the edge of the swimming pool, making it more difficult to determine whether the child was actually blowing bubbles under water or whether the bubbles were a result of the child moving in the water. This is in accordance with the findings of Vogt & Straub (2020), who also found a lower inter-rater reliability for the video recorded task of blowing bubbles under water, compared to other basic aquatic skills [29]. Another possible explanation is that the short duration for which the children performed the aquatic skill (with two out of four children blowing bubbles for less than 1.5 s), made the assessment more challenging. Adding an extra criterion to the PSPWC [24] regarding the minimal duration a child has to blow bubbles might enhance the reliability of scoring this item. Despite the finding of a moderate level of agreement, test item 6 (i.e., front star) showed the lowest inter-rater reliability. Two of the four children performing this particular test item only briefly maintained the front star or were moving while floating. Therefore, the inclusion of additional criteria in the PSPWC test manual [24] regarding a minimal duration and also regarding the amount of movement that is (not) allowed while a child is performing the front star might benefit the agreement between raters. Regarding intra-rater reliability, the level of agreement between test and re-test scoring per test video was very good to perfect. Considering the intra-rater reliability per test item across all four test videos, the level of agreement was either perfect or very good. A perfect level of agreement between test and re-test scoring was obtained by all the raters for six test items (i.e., test items 2, 4, 11, 14, 15 and 17), by four out of the five raters for five test items (i.e., test items 1, 6, 7, 8, 12 and 16) and by three out of the five raters regarding the remaining five test items (i.e., test items 3, 5, 9, 10 and 13). However, due to the limited amount of child performances per test item (n = 4), it was not possible to obtain a very good level of agreement for a specific test item. Once a participant rater awarded one of the four child performances on a particular test item a different score during the re-test than during the test, the intra-rater reliability always dropped from a perfect to a good level of agreement. Therefore, future reliability research regarding the AAST [23] should pursue a greater amount of child performances to be assessed for each of its test items.
D’Hondt et al. (2021) [23] already demonstrated a good internal consistency of the 17 test items included in both the PSPWC [24] and the AAST [23], indicating that these tools measure one overall construct (i.e., perceived vs. actual water motor competence in children, respectively). In combination with the findings of the present study, indicating that the AAST [23] also is a reliable assessment tool, it can be said that the AAST [23] has the ability to meet the need for a sound test instrument to assess young children’s actual aquatic skills, being appropriate for use by both researchers and practitioners.
In the present study, previously obtained video recordings of children performing the 17 different test items of the AAST [23] in the water of an indoor swimming pool were used. Although the AAST [23] does not require the video recording of these performances, since live administration is the usual way of working, the video recordings composed with the Shotcut video editing application [49] enforced the participating raters to watch the same performance of a child for a second time. Furthermore, the use of video recordings made it possible to include the performances of 17 different children within each composite test video, with each child performing only one of the 17 test items of the AAST [23] to ensure sufficient variation in aquatic skill level within one overall AAST test video. As such, the participant raters were stimulated to look at each aquatic skill task performance with a fresh pair of eyes, not being biased by the child’s performance on one of the previous test items. Besides, the composition of video recordings of different children reduced the chance of the raters recalling any of the children and their performances [58]. As shown in the present study, video recordings might be especially useful for training purposes, allowing both researchers and practitioners to become familiar with the PSPWC-based AAST test protocol and its scoring [23,24]. Likewise, video recordings have also been used effectively in previous studies investigating children’s motor competence assessment on land [51]. Finally, video recordings can also be used as an educational tool for teaching children and their aquatic skill assessors as provided in the context of blended learning [59].

4.1. Strengths and Weaknesses of the Study

Our study demonstrated several strengths. First of all, the reliability research conducted contributes to the limited literature regarding test instruments for assessing young children’s actual aquatic skill level, not relying on self-reports or parent-proxy reports. Second, the included skills in the AAST [23] and PSPWC [24] cover more than solely a child’s swimming ability, as they were carefully selected for their importance as fundamentals for further aquatic skill development and in regard to water safety, which is especially important for young children considering they have the highest chance of drowning compared to other age groups [3,4,5]. Third, the AAST [23], which serves as an instrument for assessing young children’s actual aquatic skill level (i.e., the physical aspect of water competence), perfectly aligns with the PSPWC [24] for assessing their perceived aquatic skill level, allowing (1) future research to further examine the important and possibly developmental variable relationship between these two vital concepts, and (2) practitioners and parents or caregivers to compare children’s actual and perceived aquatic skill level to better estimate and respond to the child’s individual needs [3]. Finally, the raters participating in this study were all students enrolled in the Master in Movement and Sport Sciences, having experience with aquatic education and movement analysis being or potentially becoming practitioners (e.g., physical education teachers, swimming instructors and lifeguards) for whom it is important to know how to assess and thus estimate young children’s actual aquatic skill level correctly.
On the other hand, this reliability study also has some limitations to be considered. A downside of using video footage is that the raters were not able to adjust their point of view and had to settle for the pool edge camera’s perspective of the child(ren)’s performance in the water. Furthermore, the video recordings only allowed for a two-dimensional perception of the performances of the different test items. As highlighted above, the lack of depth perception might have had a negative influence on the assessment of test item 3 (i.e., blowing bubbles under water), for example. The use of video footage allowed the raters to read the task criteria both before and after watching a child’s performance, which is rather unrealistic during live administration with large(r) groups of children and a limited amount of time, such as in a school or training context. Moreover, in this study the determination of the intra-rater reliability lacked sensitivity due to the limited amount of children (n = 4) performing a specific test item.

4.2. Future Directions and Practical Implications

Future research should thus further investigate the reliability of the AAST [23], when used to assess young children’s aquatic skills during live administration in the water of an indoor swimming pool [51]. Moreover, in order to examine the inter-rater and intra-rater reliability of the individual test items of the AAST (n = 17), the inclusion of a greater amount of individual child performances for each test item as well as a greater variation in mastery levels of aquatic skill between those performances than in the present study is advised. Investigating the AAST’s [23] reliability between and within raters who are active physical education teachers and swimming instructors is a direction for future research. Another avenue for future research is to explore the relationship and possible (mis)match between children’s actual AAST (measured by physical education teachers and/or swimming instructors) and the “intuition” of parents and lifeguards. Furthermore, the administration of the AAST [23] in other aquatic environments also seems a relevant avenue for future research, as drownings frequently occur in open water [20]. A long term vision on research regarding the AAST includes investigating its predictive validity in terms of drowning incidences. From a research perspective, the AAST [23] also presents itself as a useful tool for examining the correlates of young children’s actual aquatic skill levels, for determining and (internationally) comparing young childhood populations’ actual aquatic skill level as well as for evaluating intervention programs targeting to improve young children’s actual aquatic skill level. The importance of sharing data internationally is stressed by the United Nations General Assembly (2021), as it is critical for the successful global implementation of children’s water safety interventions and programs [60].
From a practitioner’s point of view, the AAST [23] is applicable for evaluation purposes and tracking the progress of a child’s actual aquatic skill level. Moreover, the instrument lends itself to assessment for learning progress (i.e., stimulating the learning process through assessment [61]) and for planning and/or individually tailoring the lessons with the appropriate lesson content (i.e., skill levels or aquatic activities). Using the AAST [23], the physical education teacher and the swimming instructor can also (better) inform parents and caregivers about their child(ren)’s actual aquatic skill level and its significance in terms of water safety and future participation in aquatic oriented activities [62]. Additionally, the PSPWC [24] as a pictorial scale can be administered by the practitioners but also the parents themselves. In combination with the results of a child’s performance on the AAST [23] in the water, this empowers practitioners and parents to sufficiently supervise, support and stimulate their child(ren) regarding their engagement in aquatic activities. Administering the AAST [23] among a large group of students might seem impractical and time-consuming, however, not every skill of every child has to be assessed in one single lesson. As stated earlier, the AAST [23] can be used to plan and tailor aquatic education lessons. In that case, each lesson might specifically focus on a particular subset of skills (e.g., the rotations), which can then be assessed. The AAST [23] might also be used as a continuous progress tracker over a period of time. When a child shows their ability to perform a particular aquatic skill from the AAST [23] at a certain level during a lesson or even during free play, this can simply be noted and registered to keep track of the individual progress being made in terms of the physical aspect of water competence.

5. Conclusions

Regarding the inter-rater reliability of the AAST [23], this study found a good to very good level of agreement between raters for the four AAST test videos included. For the inter-rater reliability of the individual AAST test items [23] across these test videos, a moderate to perfect level of agreement was found, with test item 3 and test item 6 being the only test items demonstrating a moderate level of agreement. Considering the intra-rater reliability of the AAST [23], the level of agreement between test and re-test scoring for the four AAST test videos ranged from very good to perfect for all the raters involved. For the individual AAST test items [23], the level of agreement between test-and retest scoring was considered to be good to perfect. In conclusion, this study demonstrated that the AAST [23] is a useful and reliable tool for assessing and monitoring young children’s actual aquatic skill levels. A translation of this reliability study to a live administration and scoring of the AAST [23] in the context of a swimming pool is required to confirm these promising results and to move both research and practice within this specific field forward.

Author Contributions

Conceptualization, L.M., K.D.M. and E.D.; Methodology, L.M., K.D.M. and E.D.; Formal Analysis, L.M.; Investigation, L.M.; Resources, K.D.M. and E.D.; Data Curation, L.M.; Writing—Original Draft Preparation, L.M.; Writing—Review & Editing, K.D.M., A.S. and E.D; Visualization, L.M.; Supervision, K.D.M. and E.D.; Project Administration, E.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Academic Ethics Committee of the Faculty of Medicine and Pharmacy at Vrije Universiteit Brussel (B.U.N. 143201942643) on16 September 2020.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

We are grateful to Aldo M. da Costa, Arja Sääkslahti, Boris Jidovtseff and Lisa M. Barnett as co-developers of the Pictorial Scale of Perceived Water Competence (PSPWC) for lending their expertise and assessing the evaluation video from their expertise, allowing us to optimally evaluate the participating raters. Furthermore, we would like to thank the six participating raters for their time, effort and a job well done. Finally, we extend our appreciation to all children and their parents involved, without whom the conduct of this Actual Aquatic Skills Test (AAST) reliability study would not have been possible.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Langendorfer, S.; Bruya, L.D. Aquatic Readiness: Developing Water Competence in Young Children; Human Kinetics Publishers: Champaign, IL, USA, 1995. [Google Scholar]
  2. Moran, K. Defining ‘Swim and Survive’in the Context of New Zealand Drowning Prevention Strategies: A Discussion Paper. Available online: https://www.dpanz.org.nz/research/water-competencies/ (accessed on 12 March 2020).
  3. Stallman, R.K.; Moran, K.; Quan, L.; Langendorfer, S. From swimming skill to water competence: Towards a more inclusive drowning prevention future. Int. J. Aquat. Res. 2017, 10, 3. [Google Scholar] [CrossRef] [Green Version]
  4. World Health Organization. Drowning. Available online: https://www.who.int/news-room/fact-sheets/detail/drowning (accessed on 12 February 2020).
  5. Garrido, N.D.; Costa, A.M.; Stallman, R.K. Drowning: A leading killer! Motricidade 2016, 12, 2–7. [Google Scholar] [CrossRef]
  6. World Health Organization. Preventing Drowning: An Implementation Guide; WHO: Geneva, Switzerland, 2017. [Google Scholar]
  7. Yang, L.; Nong, Q.-Q.; Li, C.-L.; Feng, Q.-M.; Lo, S.K. Risk factors for childhood drowning in rural regions of a developing country: A case–control study. Inj. Prev. 2007, 13, 178–182. [Google Scholar] [CrossRef] [Green Version]
  8. Brenner, R.A.; Taneja, G.S.; Haynie, D.L.; Trumble, A.C.; Qian, C.; Klinger, R.M.; Klebanoff, M.A. Association Between Swimming Lessons and Drowning in Childhood: A Case-Control Study. Arch. Pediatrics Adolesc. Med. 2009, 163, 203–210. [Google Scholar] [CrossRef]
  9. Haga, M. The relationship between physical fitness and motor competence in children. Child Care Health Dev. 2008, 34, 329–334. [Google Scholar] [CrossRef]
  10. Stodden, D.; Goodway, J.; Langendorfer, S.; Roberton, M.A.; Rudisill, M.; Garcia, C.; Garcia, L. A Developmental Perspective on the Role of Motor Skill Competence in Physical Activity: An Emergent Relationship. Quest 2008, 60, 290–306. [Google Scholar] [CrossRef]
  11. Lubans, D.R.; Morgan, P.J.; Cliff, D.P.; Barnett, L.M.; Okely, A.D. Fundamental Movement Skills in Children and Adolescents. Sports Med. 2010, 40, 1019–1035. [Google Scholar] [CrossRef] [Green Version]
  12. Slykerman, S.; Ridgers, N.D.; Stevenson, C.; Barnett, L.M. How important is young children’s actual and perceived movement skill competence to their physical activity? J. Sci. Med. Sport 2016, 19, 488–492. [Google Scholar] [CrossRef]
  13. Hulteen, R.M.; Morgan, P.J.; Barnett, L.M.; Stodden, D.F.; Lubans, D.R. Development of Foundational Movement Skills: A Conceptual Model for Physical Activity Across the Lifespan. Sports Med. 2018, 48, 1533–1540. [Google Scholar] [CrossRef]
  14. Sato, D.; Kaneda, K.; Wakabayashi, H.; Nomura, T. The water exercise improves health-related quality of life of frail elderly people at day service facility. Qual. Life Res. 2007, 16, 1577–1585. [Google Scholar] [CrossRef]
  15. Bergamin, M.; Zanuso, S.; Alvar, B.A.; Ermolao, A.; Zaccaria, M. Is water-based exercise training sufficient to improve physical fitness in the elderly? Eur. Rev. Aging Phys. Act. 2012, 9, 129–141. [Google Scholar] [CrossRef] [Green Version]
  16. Bergamin, M.; Ermolao, A.; Tolomio, S.; Berton, L.; Sergi, G.; Zaccaria, M. Water- versus land-based exercise in elderly subjects: Effects on physical performance and body composition. Clin. Interv. Aging 2013, 8, 1109–1117. [Google Scholar] [CrossRef] [Green Version]
  17. Tamm, S.; Cervenka, S.; Forsberg, A.; Estelius, J.; Grunewald, J.; Gyllfors, P.; Karshikoff, B.; Kosek, E.; Lampa, J.; Lensmar, C.; et al. Evidence of fatigue, disordered sleep and peripheral inflammation, but not increased brain TSPO expression, in seasonal allergy: A [11C]PBR28 PET study. Brain Behav. Immun. 2018, 68, 146–157. [Google Scholar] [CrossRef]
  18. Skelton, D.A.; Dinan, S.M. Exercise for falls management: Rationale for an exercise programme aimed at reducing postural instability. Physiother. Theory Pract. 1999, 15, 105–120. [Google Scholar] [CrossRef]
  19. Johnson, M.B.; Boriack, E.D.; McConnell, C.M.; Williams, S.R.; Naiditch, J.A.; Lawson, K.A. Predictors of swimming pool supervision for caregivers of toddlers. J. Inj. Violence Res. 2021, 13, 141–150. [Google Scholar] [CrossRef]
  20. World Health Organization. Global Report on Drowning: Preventing a Leading Killer; World Health Organization: Geneva, Switzerland, 2014. [Google Scholar]
  21. Zecevic, C.A.; Tremblay, L.; Lovsin, T.; Michel, L. Parental Influence on Young Children’s Physical Activity. Int. J. Pediatrics 2010, 2010, 468526. [Google Scholar] [CrossRef] [Green Version]
  22. Wilk, P.; Clark, A.F.; Maltby, A.; Tucker, P.; Gilliland, J.A. Exploring the effect of parental influence on children’s physical activity: The mediating role of children’s perceptions of parental support. Prev. Med. 2018, 106, 79–85. [Google Scholar] [CrossRef]
  23. D’Hondt, E.; Buelens, L.; Barnett, L.; Howells, K.; Sääkslahti, A.; Costa, A.; Jidovtseff, B.; Mertens, L.; Martelaer, K. Differences between Young Children’s Actual, Self-perceived and Parent-perceived Aquatic Skills. Percept. Mot. Ski. 2021, 128, 003151252110178. [Google Scholar] [CrossRef]
  24. Morgado, L.; Martelaer, K.; D’Hondt, E.; Barnett, L.M.; Costa, A.M.; Howells, K.; Sääkslahti, A.; Jidovtseff, B. Pictorial Scale of Perceived Water Competence (PSPWC): Testing Manual, 2nd ed.; Early Years SIG AIESEP, 2020. [Google Scholar] [CrossRef]
  25. Bardid, F.; Vannozzi, G.; Logan, S.W.; Hardy, L.L.; Barnett, L.M. A hitchhiker’s guide to assessing young people’s motor competence: Deciding what method to use. J. Sci. Med. Sport 2019, 22, 311–318. [Google Scholar] [CrossRef] [Green Version]
  26. Moreno-Murcia, J.A.; de Paula Borges, L.; Huéscar Hernández, E. Design and Validation of the Scale to Measure Aquatic Competence in Children (SMACC). Int. J. Environ. Res. 2020, 17, 6188. [Google Scholar] [CrossRef]
  27. Maeng, H.; Webster, E.K.; Pitchford, E.A.; Ulrich, D.A. Inter-and Intrarater Reliabilities of the Test of Gross Motor Development-Third Edition Among Experienced TGMD-2 Raters. Adapt. Phys. Activ. Q. 2017, 34, 442–455. [Google Scholar] [CrossRef]
  28. Hulteen, R.M.; Barnett, L.M.; True, L.; Lander, N.J.; del Pozo Cruz, B.; Lonsdale, C. Validity and reliability evidence for motor competence assessments in children and adolescents: A systematic review. J. Sports Sci. 2020, 38, 1717–1798. [Google Scholar] [CrossRef]
  29. Vogt, T.; Staub, I. Assessment of basic aquatic skills in children: Inter-rater reliability of coaches, teachers, students and parents. J. Phys. Educ. Sport 2020, 20, 577–583. [Google Scholar]
  30. Chan, D.K.C.; Lee, A.S.Y.; Macfarlane, D.J.; Hagger, M.S.; Hamilton, K. Validation of the swimming competence questionnaire for children. J. Sports Sci. 2020, 38, 1666–1673. [Google Scholar] [CrossRef]
  31. Fife, D.; Goldoft, M. Swimming capabilities and swimming exposure of New Jersey children. J. Saf. Res. 1994, 25, 159–165. [Google Scholar] [CrossRef]
  32. Irwin, C.C.; Irwin, R.L.; Ryan, T.D.; Drayer, J. Urban minority youth swimming (in)ability in the United States and associated demographic characteristics: Toward a drowning prevention plan. Inj. Prev. 2009, 15, 234–239. [Google Scholar] [CrossRef]
  33. Stallman, R.; Moran, K.; Brenner, R.; Rahman, A. Swimming and Water Survival Competence. In Drowning: Prevention, Rescue, Treatment; Bierens, J.J.L.M., Ed.; Springer: Berlin/Heidelberg, Germany, 2014; pp. 197–206. [Google Scholar]
  34. Moran, K.; Stallman, R.K.; Kjendlie, P.-L.; Dahl, D.; Blitvich, J.D. Can You swim? An Exploration of Measuring Real and Perceived Water Competency. Int. J. Aquat. Res. 2012, 6, 122–135. [Google Scholar] [CrossRef] [Green Version]
  35. Mercado, M.C.; Quan, L.; Bennett, E.; Gilchrist, J.; Levy, B.A.; Robinson, C.L.; Wendorf, K.; Gangan Fife, M.A.; Stevens, M.R.; Lee, R. Can you really swim? Validation of self and parental reports of swim skill with an inwater swim test among children attending community pools in Washington State. Inj. Prev. 2016, 22, 253–260. [Google Scholar] [CrossRef]
  36. Estevan, I.; Molina-García, J.; Bowe, S.; Alvarez, O.; Castillo, I.; Barnett, L. Who can best report on children’s motor competence: Parents, teachers, or the children themselves? Psychol. Sport Exerc. 2017, 34, 1–9. [Google Scholar] [CrossRef]
  37. Costa, A.M.; Frias, A.; Ferreira, S.S.; Costa, M.J.; Silva, A.J.; Garrido, N.D. Perceived and Real Aquatic Competence in Children from 6 to 10 Years Old. Int. J. Environ. Res. 2020, 17, 6101. [Google Scholar] [CrossRef]
  38. Brenner, R.; Moran, K.; Stallman, R.; Gilchrist, J.; McVan, J.T. Swimming Ability, Water Safety Education, and Drowning Prevention. In Handbook on Drowning: Prevention, Rescue, Treatment; Bierens, J.J.L.M., Ed.; Springer: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  39. Junge, M.; Blixt, T.; Stallman, R. The construct validity of a traditional 25m test of swimming competence. In Proceedings of the XI International Symposium for Biomechanics and Medicine in Swimming, Oslo, Norway, 16–19 June 2010; pp. 331–332. [Google Scholar]
  40. Moran, K. Getting out of the water: How hard can that be? Int. J. Aquat. Res. Educ. 2014, 8, 4. [Google Scholar]
  41. Stallman, R.K.; Junge, M.; Blixt, T. The teaching of swimming based on a model derived from the causes of drowning. Int. J. Aquat. Res. Educ. 2008, 2, 11. [Google Scholar] [CrossRef] [Green Version]
  42. McCool, E.; Moran, K.; Ameratunga, S.; Robinson, E. New Zealand beachgoers’ swimming behaviours, swimming abilities, and perception of drowning risk. Int. J. Aquat. Res. 2008, 2, 7–15. [Google Scholar] [CrossRef] [Green Version]
  43. Moran, K. Will They Sink or Swim? New Zealand Youth Water Safety Knowledge and Skills. Int. J. Aquat. Res. 2008, 2, 113–126. [Google Scholar] [CrossRef] [Green Version]
  44. McCool, J.; Ameratunga, S.; Moran, K.; Robinson, E. Taking a Risk Perception Approach to Improving Beach Swimming Safety. Int. J. Behav. Med. 2009, 16, 360–366. [Google Scholar] [CrossRef]
  45. Moran, K. (Young) Men behaving badly: Dangerous masculinities and the risk of drowning in aquatic leisure activities. J. Leis. Res. 2011, 14, 260–272. [Google Scholar] [CrossRef]
  46. Estevan, I.; Barnett, L.M. Considerations Related to the Definition, Measurement and Analysis of Perceived Motor Competence. Sports Med. 2018, 48, 2685–2694. [Google Scholar] [CrossRef]
  47. De Meester, A.; Barnett, L.M.; Brian, A.; Bowe, S.J.; Jiménez-Díaz, J.; Van Duyse, F.; Irwin, J.M.; Stodden, D.F.; D’Hondt, E.; Lenoir, M.; et al. The Relationship Between Actual and Perceived Motor Competence in Children, Adolescents and Young Adults: A Systematic Review and Meta-analysis. Sports Med. 2020, 50, 2001–2049. [Google Scholar] [CrossRef]
  48. Rejman, M.; Kwaśna, A.; Chrobot, M.; Kjendlie, P.-L.; Stalmann, R.K. Perceived Versus Real Swimming Skills of Adolescents under Standard and Challenging Conditions: Exploring Water Competencies as an Approach to Drowning Prevention. Int. J. Environ. Res. 2020, 17, 3826. [Google Scholar] [CrossRef]
  49. Meltytech, LLC. Shotcut (Version 20.07.11) [Computer Software]; Meltytech, LLC: Oceanside, CA, USA, 2011–2021. [Google Scholar]
  50. Barnett, L.M.; Minto, C.; Lander, N.; Hardy, L.L. Interrater reliability assessment using the Test of Gross Motor Development-2. J. Sci. Med. Sport 2014, 17, 667–670. [Google Scholar] [CrossRef]
  51. Rintala, P.; Sääkslahti, A.; Iivonen, S. Reliability Assessment of Scores From Video-Recorded TGMD-3 Performances. J. Mot. Learn. Dev. 2017, 5, 59–68. [Google Scholar] [CrossRef]
  52. Robertson, S.J.; Burnett, A.F.; Cochrane, J. Tests examining skill outcomes in sport: A systematic review of measurement properties and feasibility. Sports Med. 2014, 44, 501–518. [Google Scholar] [CrossRef] [Green Version]
  53. Hayes, A.F. Answering the call for a standard reliability measure for coding data. Commun. Methods Meas. 2007, 1, 77–89. [Google Scholar] [CrossRef]
  54. Zaiontz, C. Real Statistics Resource Pack. Available online: https://www.real-statistics.com/free-download/real-statistics-resource-pack/ (accessed on 27 May 2021).
  55. Quarfoot, D.; Levine, R.A. How Robust Are Multirater Interrater Reliability Indices to Changes in Frequency Distribution? Am. Stat. 2016, 70, 373–384. [Google Scholar] [CrossRef]
  56. Tran, D.; Dolgun, A.; Demirhan, H. Weighted inter-rater agreement measures for ordinal outcomes. Commun. Stat.—Simul. Comput. 2020, 49, 989–1003. [Google Scholar] [CrossRef]
  57. Thoenen, J.; Stevens, K.J.; Turmezei, T.D.; Chaudhari, A.; Watkins, L.E.; McWalter, E.J.; Hargreaves, B.A.; Gold, G.E.; MacKay, J.W.; Kogan, F. Non-contrast MRI of synovitis in the knee using quantitative DESS. Eur. Radiol. 2021, 31, 9369–9379. [Google Scholar] [CrossRef]
  58. Hatfield, B.D.; Landers, D.M. Observer Expectancy Effects upon Appraisal of Gross Motor Performance. Res. Quart. Am. Alliance Health Phys. Educ. Recreat. 1978, 49, 53–61. [Google Scholar] [CrossRef]
  59. De Martelaer, K.; Bardyn, B.; Cools, W.; Van Duijn, T. Een blended opleiding tot zwemleraar. Licham. Opvoeding 2021, 270, 31–36. [Google Scholar]
  60. United Nations General Assembly. General Assembly Resolution, 75/273, Global Drowning Prevention; A/RES/75/273; United Nations: New York, NY, USA, 2021. [Google Scholar]
  61. AIESEP Position Statement on Physical Education Assessment. Available online: https://aiesep.org/scientific-meetings/position-statements/ (accessed on 12 October 2021).
  62. De Pasquale, C.; De Sousa Morgado, L.; Jidovtseff, B.; De Martelaer, K.; Barnett, L.M. Utility of a scale to assess Australian children’s perceptions of their swimming competence and factors associated with child and parent perception. Health Promot. J. Aust. 2021, 32, 106–115. [Google Scholar] [CrossRef]
Table 1. Rater’s score sheet example.
Table 1. Rater’s score sheet example.
What Is the Appropriate Level for the Child Observed Per Test Item?
‘1’ = not able to perform the aquatic skill
‘2’ = partly able to perform the aquatic skill
‘3’ = able to correctly perform the aquatic skill
Aquatic skill orDescriptionScore:
test item‘1’, ‘2’ or ‘3’.
1Lying down in a prone position using hands on the bottom to move forward (as a crocodile)
2Standing and submersion in the water
3Blowing bubbles under water
4Catching an object under water
5Floating on the back (i.e., back star)
6Floating on the front (i.e., front star)
7Water entry by slide
8Pushing from the wall and gliding under water
9Leg propulsion on the back
10Leg propulsion on the front
11Water entry by jumping
12Water entry by diving
13Water exit by climbing out
14Vertically treading water
15Turning from the front to the back in an aligned position
(i.e., longitudinal axis rotation)
16Changing direction while swimming on the front
(i.e., transverse axis rotation)
17Turning from the back to the front (i.e., sagittal axis rotation)
Table 2. Overview of the three different mastery levels per test item as included in the Pictorial Scale of Perceived Water Competence (PSPWC) [24].
Table 2. Overview of the three different mastery levels per test item as included in the Pictorial Scale of Perceived Water Competence (PSPWC) [24].
Level
Aquatic Skill
or Test Item
Unable to Execute the
Aquatic Skill
‘1’
Partly Able to Execute the Aquatic Skill (in Progress)
‘2’
Fully Able to Execute
the Aquatic Skill
‘3’
Test item 1
Move forward as a crocodile
Ijerph 19 00446 i001 Ijerph 19 00446 i002 Ijerph 19 00446 i003
Test item 2
Standing and submersion in the water
Ijerph 19 00446 i004 Ijerph 19 00446 i005 Ijerph 19 00446 i006
Test item 3
Blowing bubbles under water
Ijerph 19 00446 i007 Ijerph 19 00446 i008 Ijerph 19 00446 i009
Test item 4
Catching an object under water
Ijerph 19 00446 i010 Ijerph 19 00446 i011 Ijerph 19 00446 i012
Test item 5
Floating on the back (i.e., back star)
Ijerph 19 00446 i013 Ijerph 19 00446 i014 Ijerph 19 00446 i015
Test item 6
Floating on the front (i.e., front star)
Ijerph 19 00446 i016 Ijerph 19 00446 i017 Ijerph 19 00446 i018
Test item 7
Water entry by slide
Ijerph 19 00446 i019 Ijerph 19 00446 i020 Ijerph 19 00446 i021
Test item 8
Pushing from the wall and gliding under water
Ijerph 19 00446 i022 Ijerph 19 00446 i023 Ijerph 19 00446 i024
Test item 9
Leg propulsion on the back
Ijerph 19 00446 i025 Ijerph 19 00446 i026 Ijerph 19 00446 i027
Test item 10
Leg propulsion on the front
Ijerph 19 00446 i028 Ijerph 19 00446 i029 Ijerph 19 00446 i030
Test item 11
Water entry by jumping
Ijerph 19 00446 i031 Ijerph 19 00446 i032 Ijerph 19 00446 i033
Test item 12
Water entry by diving
Ijerph 19 00446 i034 Ijerph 19 00446 i035 Ijerph 19 00446 i036
Test item 13
Water exit by climbing out
Ijerph 19 00446 i037 Ijerph 19 00446 i038 Ijerph 19 00446 i039
Test item 14
Vertically treading water
Ijerph 19 00446 i040 Ijerph 19 00446 i041 Ijerph 19 00446 i042
Test item 15
Longitudinal axis rotation
Ijerph 19 00446 i043 Ijerph 19 00446 i044 Ijerph 19 00446 i045
Test item 16
Transverse axis rotation
Ijerph 19 00446 i046 Ijerph 19 00446 i047 Ijerph 19 00446 i048
Test item 17
Sagittal axis rotation
Ijerph 19 00446 i049 Ijerph 19 00446 i050 Ijerph 19 00446 i051
Permission to publish these pictures was granted by the developers of the PSPWC [24]. Permission to publish this table was granted by D’Hondt et al. (2021) [23].
Table 3. Example of the minimum criteria per mastery level for test item 1 of the PSPWC [24], when performed by the child in the water of an indoor swimming pool during the Actual Aquatic Skills Test (AAST) [23].
Table 3. Example of the minimum criteria per mastery level for test item 1 of the PSPWC [24], when performed by the child in the water of an indoor swimming pool during the Actual Aquatic Skills Test (AAST) [23].
Test Item 1
Lying down in a prone position using hands on the bottom to move forward
(as a crocodile)
Level Minimum criteria
1 Ijerph 19 00446 i052The child stands up in the shallow water but does not dare to lie down.
2 Ijerph 19 00446 i053The child moves forward on all fours but is not completely submerged in the water.
3 Ijerph 19 00446 i054The child lies down in a prone position with arms or hands in contact with the bottom of the wading (paddling) pool, with the body extended and immersion up to the shoulders.
Permission to publish these pictures and criteria was granted by the developers of the PSPWC [24].
Table 4. Gwet’s AC2 test results for inter-rater reliability.
Table 4. Gwet’s AC2 test results for inter-rater reliability.
Test ComponentAC2LoASE SubjCI 95% SubjSE TotCI 95% Tot
Test video 10.794G0.083[0.617; 0.971]0.093[0.596; 0.993]
Test video 20.646G0.091[0.452; 0.839]0.126[0.379; 0.912]
Test video 30.701G0.104[0.480; 0.921]0.123[0.440; 0.962]
Test video 40.922VG0.055[0.806; 1.000]0.062[0.792; 1.000]
Test item 1
Move forward as a crocodile
1.000P0.000[1.000; 1.000]0.000[1.000; 1.000]
Test item 2
Standing and submersion in the water
1.000P0.000[1.000; 1.000]0.000[1.000; 1.000]
Test item 3
Blowing bubbles under water
0.507M0.199[−0.127; 1.000]0.382[−0.708; 1.000]
Test item 4
Catching an object under water
1.000P0.000[1.000; 1.000]0.000[1.000; 1.000]
Test item 5
Floating on the back (i.e., back star)
0.671G0.194[0.052; 1.000]0.262[−0.164; 1.000]
Test item 6
Floating on the front (i.e., front star)
0.414M0.207[−0.246; 1.000]0.310[−0.571; 1.000]
Test item 7
Water entry by slide
0.713G0.194[0.097; 1.000]0.254[−0.094; 1.000]
Test item 8
Pushing from the wall and gliding under water
0.835VG0.179[0.264; 1.000]0.187[0.240; 1.000]
Test item 9
Leg propulsion on the back
1.000P0.000[1.000; 1.000]0.000[1.000; 1.000]
Test item 10
Leg propulsion on the front
0.634G0.235[−0.114; 1.000]0.273[−0.234; 1.000]
Test item 11
Water entry by jumping
0.863VG0.143[0.408; 1.000]0.199[0.230; 1.000]
Test item 12
Water entry by diving
0.777G0.219[0.079; 1.000]0.228[0.051; 1.000]
Test item 13
Water exit by climbing out
0.702G0.207[0.043; 1.000]0.237[−0.053; 1.000]
Test item 14
Vertically treading water
0.657G0.257[−0.161; 1.000]0.288[−0.261; 1.000]
Test item 15
Longitudinal axis rotation
0.806VG0.196[0.181; 1.000]0.204[0.155; 1.000]
Test item 16
Transverse axis rotation
0.780G0.138[0.341; 1.000]0.200[0.143; 1.000]
Test item 17
Sagittal axis rotation
0.828VG0.193[0.215; 1.000]0.196[0.203;1.000]
AC2 = Gwet’s Agreement Coefficient 2; LoA = Level of Agreement; M = moderate (0.41–0.60); G = good (0.61–0.80); VG = very good (0.81–0.99); P = perfect (1.00); SE = standard error; CI = confidence interval; subj = subjects; tot = total.
Table 5. Agreement table for inter-rater reliability.
Table 5. Agreement table for inter-rater reliability.
Test ItemTest VideoPercentage of Raters Per Awarded Score (%)
Score ‘1’Score ‘2’Score ‘3’
Test item 1
Move forward as a crocodile
Test video 100100
Test video 201000
Test video 300100
Test video 400100
Test item 2
Standing and submersion in the water
Test video 100100
Test video 200100
Test video 300100
Test video 400100
Test item 3
Blowing bubbles under water
Test video 102080
Test video 202080
Test video 340600
Test video 401000
Test item 4
Catching an object under water
Test video 110000
Test video 200100
Test video 310000
Test video 400100
Test item 5
Floating on the back (i.e., back star)
Test video 104060
Test video 202080
Test video 300100
Test video 401000
Test item 6
Floating on the front (i.e., front star)
Test video 106040
Test video 260400
Test video 320800
Test video 400100
Test item 7
Water entry by slide
Test video 100100
Test video 202080
Test video 304060
Test video 400100
Test item 8
Pushing from the wall and gliding under water
Test video 100100
Test video 204060
Test video 300100
Test video 400100
Test item 9
Leg propulsion on the back
Test video 101000
Test video 210000
Test video 300100
Test video 401000
Test item 10
Leg propulsion on the front
Test video 101000
Test video 220080
Test video 360400
Test video 400100
Test item 11
Water entry by jumping
Test video 110000
Test video 200100
Test video 302080
Test video 410000
Test item 12
Water entry by diving
Test video 106040
Test video 210000
Test video 301000
Test video 400100
Test item 13
Water exit by climbing out
Test video 102080
Test video 206040
Test video 300100
Test video 400100
Test item 14
Vertically treading water
Test video 100100
Test video 200100
Test video 3206020
Test video 402080
Test item 15
Longitudinal axis rotation
Test video 100100
Test video 204060
Test video 300100
Test video 401000
Test item 16
Transverse axis rotation
Test video 100100
Test video 202080
Test video 302080
Test video 400100
Test item 17
Sagittal axis rotation
Test video 100100
Test video 200100
Test video 300100
Test video 406040
Table 6. Gwet’s AC2 test results for intra-rater reliability.
Table 6. Gwet’s AC2 test results for intra-rater reliability.
TestRater ARater BRater CRater DRater E
ComponentAC2LoASECI 95%AC2LoASECI 95%AC2LoASECI 95%AC2LoASECI 95%AC2LoASECI 95%
Test video 10.848VG0.112[0.610; 1.000]1.000P0.000[1.000; 1.000]0.920VG0.081[0.749; 1.000]0.839VG0.114[0.597; 1.000]0.837VG0.112[0.600; 1.000]
Test video 21.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]0.834VG0.116[0.589; 1.000]0.914VG0.086[0.733; 1.000]0.837VG0.112[0.600; 1.000]
Test video 31.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]0.918VG0.082[0.743; 1.000]
Test video 41.000P0.000[1.000; 1.000]0.855VG0.102[0.637; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]0.919VG0.082[0.744; 1.000]
Test item 1
Move forward as a crocodile
1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]0.660G0.357[−0.477; 1.000]
Test item 2
Standing and submersion
1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]
Test item 3
Blowing bubbles under water
1.000P0.000[1.000; 1.000]0.673G0.332[−0.383; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]0.660G0.357[−0.477; 1.000]
Test item 4
Catching an object under water
1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]
Test item 5
Floating on the back (i.e., back star
1.000P0.000[1.000; 1.000]0.719G0.310[−0.268; 1.000]1.000P0.000[1.000; 1.000]0.673G0.332[−0.383; 1.000]1.000P0.000[1.000; 1.000]
Test item 6
Floating on the front (i.e., front star)
1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]0.660G0.357[−0.477; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]
Test item 7
Water entry by slide
1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]0.719G0.310[−0.268; 1.000]1.000P0.000[1.000; 1.000]
Test item 8
Pushing from the wall and gliding under water
1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]0.719G0.310[−0.268; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]
Test item 9
Leg propulsion on the back
0.628G0.367[−0.539; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]0.628G0.367[−0.539; 1.000]
Test item 10
Leg propulsion on the front
0.644G0.382[−0.572; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]0.660G0.357[−0.477; 1.000]1.000P0.000[1.000; 1.000]
Test item 11
Water entry by jumping
1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]
Test item 12
Water entry by diving
1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]0.628G0.367[−0.539; 1.000]
Test item 13
Water exit by climbing out
1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]0.673G0.332[−0.383; 1.000]1.000P0.000[1.000; 1.000]0.719G0.310[−0.268; 1.000]
Test item 14
Vertically treading water
1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]
Test item 15
Longitudinal axis rotation
1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]
Test item 16
Transverse axis rotation
1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]0.719G0.310[−0.268; 1.000]
Test item 17
Sagittal axis rotation
1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]1.000P0.000[1.000; 1.000]
AC2 = Gwet’s Agreement Coefficient 2; LoA = Level of Agreement; M = moderate (0.41–0.60); G = good (0.61–0.80); VG = very good (0.81–0.99); P = perfect (1.00); SE = standard error; CI = confidence interval.
Table 7. Test re-test agreement per rater.
Table 7. Test re-test agreement per rater.
Test ComponentRater ARater BRater CRater DRater E
Test video 188.24%100%94.12%88.24%88.24%
Test video 2100%100%88.24%94.12%88.24%
Test video 3100%100%100%100%94.12%
Test video 4100%88.24%100%100%94.12%
Test item 1
Move forward as a crocodile
100%100%100%100%75%
Test item 2
Standing and submersion
100%100%100%100%100%
Test item 3
Blowing bubbles under water
100%75%100%100%75%
Test item 4
Catching an object under water
100%100%100%100%100%
Test item 5
Floating on the back (i.e., back star)
100%75%100%75%100%
Test item 6
Floating on the front (i.e., front star)
100%100%75%100%100%
Test item 7
Water entry by slide
100%100%100%75%100%
Test item 8
Pushing from the wall and gliding under water
100%100%75%100%100%
Test item 9
Leg propulsion on the back
75%100%100%100%75%
Test item 10
Leg propulsion on the front
75%100%100%75%100%
Test item 11
Water entry by jumping
100%100%100%100%100%
Test item 12
Water entry by diving
100%100%100%100%75%
Test item 13
Water exit by climbing out
100%100%75%100%75%
Test item 14
Vertically treading water
100%100%100%100%100%
Test item 15
Longitudinal axis rotation
100%100%100%100%100%
Test item 16
Transverse axis rotation
100%100%100%100%75%
Test item 17
Sagittal axis rotation
100%100%100%100%100%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mertens, L.; De Martelaer, K.; Sääkslahti, A.; D’Hondt, E. The Inter-Rater and Intra-Rater Reliability of the Actual Aquatic Skills Test (AAST) for Assessing Young Children’s Motor Competence in the Water. Int. J. Environ. Res. Public Health 2022, 19, 446. https://0-doi-org.brum.beds.ac.uk/10.3390/ijerph19010446

AMA Style

Mertens L, De Martelaer K, Sääkslahti A, D’Hondt E. The Inter-Rater and Intra-Rater Reliability of the Actual Aquatic Skills Test (AAST) for Assessing Young Children’s Motor Competence in the Water. International Journal of Environmental Research and Public Health. 2022; 19(1):446. https://0-doi-org.brum.beds.ac.uk/10.3390/ijerph19010446

Chicago/Turabian Style

Mertens, Lisa, Kristine De Martelaer, Arja Sääkslahti, and Eva D’Hondt. 2022. "The Inter-Rater and Intra-Rater Reliability of the Actual Aquatic Skills Test (AAST) for Assessing Young Children’s Motor Competence in the Water" International Journal of Environmental Research and Public Health 19, no. 1: 446. https://0-doi-org.brum.beds.ac.uk/10.3390/ijerph19010446

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop