Skip to main content

The nutrition for sport knowledge questionnaire (NSKQ): development and validation using classical test theory and Rasch analysis



Appropriate dietary intake can have a significant influence on athletic performance. There is a growing consensus on sports nutrition and professionals working with athletes often provide dietary education. However, due to the limitations of existing sports nutrition knowledge questionnaires, previous reports of athletes’ nutrition knowledge may be inaccurate.


An updated questionnaire has been developed based on a recent review of sports nutrition guidelines. The tool has been validated using a robust methodology that incorporates relevant techniques from classical test theory (CTT) and Item response theory (IRT), namely, Rasch analysis.


The final questionnaire has 89 questions and six sub-sections (weight management, macronutrients, micronutrients, sports nutrition, supplements, and alcohol). The content and face validity of the tool have been confirmed based on feedback from expert sports dietitians and university sports students, respectively. The internal reliability of the questionnaire as a whole is high (KR = 0.88), and most sub-sections achieved an acceptable internal reliability. Construct validity has been confirmed, with an independent T-test revealing a significant (p < 0.001) difference in knowledge scores of nutrition (64 ± 16%) and non-nutrition students (51 ± 19%). Test-retest reliability has been assured, with a strong correlation (r = 0.92, p < 0.001) between individuals’ scores on two attempts of the test, 10 days to 2 weeks apart. Three of the sub-sections fit the Rasch Unidimensional Model.


The final version of the questionnaire represents a significant improvement over previous tools. Each nutrition sub-section is unidimensional, and therefore researchers and practitioners can use these individually, as required. Use of the questionnaire will allow researchers to draw conclusions about the effectiveness of nutrition education programs, and differences in knowledge across athletes of varying ages, genders, and athletic calibres.


Appropriate dietary intake can improve athletic performance, enhance adaptations to training and augment recovery from exercise [1, 2]. However, athletes have been known to consume diets that do not to meet their energy and nutrient needs [3], and a mismatch between contemporary expert recommendations and athletes’ dietary practices have previously been demonstrated [4]. Nutrition education programs improve nutrition knowledge [5,6,7] and higher levels of knowledge are correlated with better diet quality [7,8,9]. Accordingly, professionals working with sports people often provide nutrition advice [10]. Parks et al. [11] reported that the number of dietitians employed by collegiate athletic departments has quadrupled since 2010. However, globally there is limited information regarding athletes’ access to relevant and appropriate nutrition advice; ostensibly, this may vary according to the level of professionalism of their respective sport and their immediate support network. Hamilton et al. [12] reported that elite athletes in New Zealand had higher levels of knowledge than non-elite athletes. In contrast, Andrews et al. [13] found no differences between sub-elite and elite Australian soccer players. Trakman et al. [14] conducted a systematic literature review on nutrition knowledge of athletes and coaches and reported a possible relationship between athletic calibre and knowledge. However, the authors concluded that due to the heterogeneity and poor quality of Nutrition Knowledge Questionnaires (NKQ’s), athletes’ nutrition knowledge (and the factors that influence this) are difficult to ascertain [0–22]. The poor quality of NKQs is also likely to influence researchers’ ability to accurately quantify the correlation between knowledge and dietary intake [8, 15] and impact practitioners’ ability to evaluate nutrition education programs.

Trakman et al. [14] noted that a key factor affecting the quality of NKQs was a lack of adequate validation. The maximum validation score of a sports nutrition knowledge questionnaire (SNKQ) used with athletes was three out of six. More recently, Furber et al. [16] developed an SNKQ for British track and field athletes undertaking four of the six recommended validation methods; face validity testing and item analysis were not performed. Of note, the rating system used by Trakman et al. [14] was based solely on classical test theory (CTT). The CTT framework focuses on the questionnaire as a whole. It is based on correlations and assumes that all questions are equal indicators of an individual’s nutrition knowledge [17]. A key aspect of CTT is the use of the Cα statistic to measure internal reliability; however, Cα is only suitable for scales with 20 or fewer items and is frequently incorrectly used on much longer questionnaires [18]. Moreover, it is not an inherent property of a questionnaire and needs to be re-assessed each time a new sample completes the tool [18].

An alternative to CTT is Rasch analysis. Rasch analysis is a technique that was first developed in education, has been utilised to develop psychological assessment tools [19] and health related patient-reported outcomes (HR-PRO) [12, 20], and more recently has been utilised to validate questionnaires that assess knowledge of the energy content of meals and balanced meals [21, 22]. Rasch analysis offers several advantages over CTT; it allows shorter scales with multiple response formats to be developed, and because it does not rely on measures of central tendency, it is said to be more ‘stable’ across varying populations [23]. The aim of Rasch analysis is to create a unidimensional (i.e. assessing one concept) questionnaire. During Rasch analysis it is necessary to test that the questionnaire concurs with the assumptions that (1) difficult items are less likely to be answered correctly, and (2) individuals with higher levels of knowledge are more likely to answer questions correctly. These expectations are tested by assessing a range of statistics which provide feedback on: the differences between observed and expected responses; whether the difficulty of items is consistent across participants (i.e. whether items are good at discriminating between well-scoring and poor-scoring respondents); and whether items are answered consistently on the basis of participant characteristics, such as age and gender. The present study will use a novel method that evaluates items based both on CTT and Rasch analysis. To our knowledge, no SNKQ has been validated using Rasch analysis.

In addition to issues pertaining to validation, many existing SNKQs have problems with their actual content. While 13 (out of 36) studies in the review by Trakman et al. [14] covered 75% of the nutrition sub-sections that were deemed relevant, the comprehensiveness assessment was limited because the researchers did not assess the extent to which each topic was assessed or the quality of individual items. Indeed, many items appear to test out-dated dietary recommendations that are not in line with recently published guidelines such as the American College of Sports Medicine (ACSM), the International Olympic Committee (IOC), and the International Society for Sports Nutrition (ISSN) review on Sport Nutrition [24] and the multiple ISSN, IOC review papers and consensus statements on nutrition and athletic performance [1, 2, 25,26,27,28,29,30,31,32,33]. As above, current guidelines expound that carefully choosing the amount, type and timing of foods and fluids will optimise an athlete’s adaptations to training, performance outcomes, and recovery from exercise. They emphasise the importance of individualising nutrition, especially with regards to carbohydrate intake and hydration, and acknowledge that some supplements (e.g. creatine, caffeine, and bicarbonate) can enhance athletes’ performance, but encourage a prudent approach to supplementation [1, 2, 25, 26, 34]. The present study has based questions on these recommendations.

Further to the issues pertaining to the quality and content of existing SNKQ’s, many tools have limited cultural applicability and/or focus on measuring the knowledge of a single sport. This limits the ability of tools to be used to compare knowledge of athletes from different countries and knowledge of athletes between sports.

The aim of this study was to address the deficiencies in existing SNKQ’s by developing a new SNKQ that:

  1. (1)

    Has been validated using a robust methodology that includes both CTT techniques and Rasch analysis

  2. (2)

    Assesses knowledge of current consensus recommendations on sports nutrition

  3. (3)

    Assesses knowledge of all relevant aspects of sports nutrition and is generalizable to multiple sports

  4. (4)

    Is likely to be understood by individuals from various cultural backgrounds

It was hypothesized that the questionnaire would represent a significant improvement on currently available measures. From a research perspective, a high-quality nutrition knowledge measure will allow for more accurate assessment of factors that influence knowledge and a more a more reliable assessment of the impact of nutrition knowledge on diet quality. Moreover, for individuals working with athletes, a quality measure is likely to have practical implications, allowing for the evaluation of nutrition education programs and therefore development of more targeted education strategies that are based on gaps in knowledge.


Ethics approval and consent to participate

The research was approved (S16/267) by the La Trobe University’s SHE College Human Ethics Sub-Committee (SHE CHESC). Participants in the expert panel review and think-out-loud focus groups read the Participant Information Statement and signed the Consent Form. Participants who were involved in steps six to eight read the Participant Information Statement and provided consent electronically.

Novel eight-step validation process

A novel eight-step validation method for the development of a nutrition knowledge questionnaire was designed based on an extensive review of the literature and used to validate this questionnaire [35]. The steps include: (1) Definition of Sports Nutrition Knowledge (2) Generation of items to represent sports nutrition knowledge (3) Choice of scoring system (4) Assessment of content validity by panel of experts (5) Assessment of face validity by student athletes (6) CTT analysis: Removal of items on the basis of item difficulty, item discrimination and distractor utility (7) Rasch analysis: Assessment of dimensionality and removal of item on the basis of not meeting assumptions that difficult questions are less likely to be answered correctly, and well-scoring participants are more likely to answer individual items correctly (8) Assessment of construct validity by comparing nutrition and non-nutrition students; assessment of test-retest reliability (consistency over time) by assessing correlation of test on two attempts; and re-checking of steps six and seven. The steps that make this methodology novel are the quantitative assessment of content validity, the assessment of distractor utility (how feasible incorrect multiple choice options are) and the inclusion of Rasch analysis.

Figure 1 provides a summary of the methods and results.

Fig. 1
figure 1

Flow chart of 8-step methadology used to develop and validate the Nutrition for Sport Questionnaire (UNSQ). * Content Validity = the measure covers all relevant topics related to sports nutrition. † CVI = Number of experts who rated an item ‘very relevant’ or ‘relevant’ divided by total number of experts; > 0.78 is adequate. ‡ Face Validity = the measure, on face value is an adequate reflection of sports nutrition. § Difficulty index = frequency with which items were answered correctly; <20% = too hard; >80% = too easy. ǁ Discrimination index = average score of top 10% of participants minus average score of bottom 10% of participants; > 0.3 is adequate. ¶ Distractor utility = frequency with which each multi-choice option is selected; > 5% = effective distractor. **Fit residuals between −2.5 and 2.5 indicate observed = expected responses. ††DIF assessed using ANOVA; non-significant p-value = no differences in response pattern based on participant characteristics; ‡‡ Disordered thresholds are assessed graphically. §§ Perc5% statistic <5% = scale is unidimensional (assessing one concept). ǁ ǁ SD of 0 and Mean of 1for the overall item/person interaction = perfect fit to Rasch model; a SD > 1.5 = misfit. ¶¶ Significant differences in known-group comparison scores = construct validity (questionnaire test what it is supposed to). *** Pearson’s r > 0.7 = test-retest reliability (stability overtime). ††† KR-20 > 0.7 = Internal reliability (consistency in items)


Experts (step four) were recruited using purposive sampling between April and May 2016 and student-athletes (step five) were recruited using convenience sampling in June 2016. For steps six and seven, Australian Football League (AFL) Victoria community football players, La Trobe University student-athletes and other recreational athletes were invited to complete the questionnaire via email, Facebook groups, and online athlete forums. Data collection occurred between July 2016 and October 2016. For step eight, La Trobe University undergraduate and postgraduate business, health science and nutrition students were recruited via email and their Learning Management System (LMS) notice boards and recreational athletes were invited to participate via email and their team Facebook pages. Data collection took place from November 2016 until January 2017.

Sample size calculations

To calculate a CVI (step four) three to 10 experts are needed [36]. The ideal number of participants for focus groups (step five) are six to 10 [37]. Parmenter et al. [38] recommend that to carry out CTT analysis at least one more person than the number of items are required. Pallant [23] recommends that to carry out Rasch analysis, 240 participants are ideal. At steps 6 and 7, the questionnaire had 178 items; a target of 200 participants was set to account for both the CTT and Rasch estimates. For step eight, a power analysis for an independent sample t-test was conducted in G-POWER to determine a sufficient sample size using an alpha of 0.05, a power of 0.80, a large effect size (d = 0.8), and two tails. Based on these assumptions, the desired sample size for each group (nutrition versus non-nutrition students) was 51 [39, 40].

Statistical analysis

Missing values were assessed for normality using Kolmogorov-Smirnov statistic and compared across participant characteristics using independent t-test and ANOVA or Mann–Whitney and Kruskal-Walis, as appropriate (see Additional file 1: Supplementary Material S1). Total and sub-section scores were assessed for normality using Kolmogorov-Smirnov statistic. Parametric tests (independent t-test; ANOVA) were used for normal data. Where assumptions of normality were violated, non-parametric tests (Mann–Whitney U test; Kruskal-Wallis) were used. Normal results were reported as mean ± SD, non-normal results were reported as median and IQR. For correlations, Pearson’s r was used for parametric data, and Spearman’s r was used for non-parametric data. Individuals with and without nutrition education were compared across participant characteristics using chi-square test, to account for potential confounding. Differences in scores based on age, gender, country of birth, level of education, and history of playing sports were also assessed.



Ten experts were invited to be involved in the content validity assessment. Six sports dietitians agreed to participate and three of these returned feedback forms; two respondents were Australian and the other was Swiss. They worked in private consultancy, research and education, and industry. Eight students participated in the retrospective think-out-loud focus group. One student who was unable to attend the session met with the researcher on a separate occasion. For the item analysis based on CTT and Rasch scaling (steps six and seven), 462 athletes started the questionnaire; after excluding data with more than 11% missing values (n = 259) and participants who did not meet the eligibility criteria for age (n = 15) there were 188 usable responses (Table 1). For the assessment of construct validity, test-retest reliability and re-evaluation of CTT and Rasch parameters (step eight), 287 students and athletes started the questionnaire and there were 181 usable responses, including 28 responses from individuals who had completed the questionnaire on two occasions (Table 2).

Table 1 Characteristics of participants included in the first item analysis using CTT and Rasch analysis (n = 188)
Table 2 Characteristics of participants included in Study Two analysis (n = 181)

There did not appear to be any differences between individuals who completed and did not complete the questionnaire (See Additional file 1: Supplementary Material S2).

Individuals who studied nutrition were more likely to be female, have a tertiary education and were younger (See Additional file 1: Supplementary Material S3).

The final questionnaire

Sports nutrition knowledge was defined as “Knowledge of concepts and processes related to nutrition for optimal athletic performance including knowledge of weight management; hydration and fuelling strategies for before, during and after training/performance; supplementation and alcohol use”. The original test plan (see Additional file 1: Supplementary Material S3) follows logically from this. Each correct answer was awarded a point. The final questionnaire had 89 items and six unidimensional sub-sections; weight management (n = 13); macronutrients (n = 30); micronutrients (n = 13); sports nutrition (n = 13); supplements (n = 12) and alcohol (n = 8). The sports nutrition section covers hydration, the pre-completion meal, nutrition during exercise and recovery nutrition. The item response formats include agree/disagree/not sure, multiple choice, and effective/not effective/not sure. The questionnaire is designed to be administered online and includes pictures to reduce responder fatigue. The number of items at each stage of validation is represented in Fig. 1.


The experts reported that the questionnaire covered all relevant topics and therefore content validity was confirmed (step four). Most items that were rated poorly for accuracy, relevance, appropriateness and clarity were removed (n = 25) although some exceptions were made. The number of items that did not meet the CVI and examples of changes made based on expert’s feedback are included as Additional file 1: Supplementary Material S4.

Student athletes were able to identify what each section was attempting to capture. Where there were issues with wording of questions, items were removed (n = 17) or modified (n = 16), so face validity was confirmed (step five). Examples of changes made based on students’ feedback are included as Additional file 1: Supplementary Material S5.

Construct validity was demonstrated (step eight) because there was a significant difference in the total scores for those who had nutrition education and those who did not have nutrition education (P < 0.001) (Table 3). Individuals with nutrition education also scored better across all sub-sections except alcohol (Table 3). No other participant characteristics were found to have a significant effect on scores (See Additional file 1: Supplementary Material S6).

Table 3 Comparison of scores between individuals with and without previous nutrition education; test-retest reliability; internal reliability analysis; results of Rasch analysis

Internal reliability of sub-sections ranged from 0.51 to 0.78 (Table 3). All sections except alcohol achieved or approached adequate values for internal reliability given the number of items in each sub-section. Test re-test reliability ranged from 0.35 to 0.81; all sections except supplementation achieved or approached adequate values for test-retest reliability (Table 3).

Item analysis based on CTT

Item analysis using CTT was carried out at steps 7 and 8. Most items that did not meet the criteria for item difficulty, item discrimination and distractor utility were removed or modified, although some exceptions were made, and the final tool included 15 items that did not meet these criteria. Examples of changes made on the basis of item analysis are included as Additional file 1: Supplementary Materials S7 and S8.

Fit to the rash model

Based on the Rasch analysis conducted on the tool administered at step eight, the questionnaire as a whole was multidimensional. Eleven items were removed because they were shown to be poor discriminators and/or were answered differently (demonstrated DIF) on the basis of age or gender. With these items removed, the questionnaire as a whole was still mutli-dimensional, but the six sub-sections were shown to be unidimensional (Table 3). Each sub-section also had an adequate mean, SD for the overall item/trait interaction statistics, meaning that the expectation that difficult items were more likely to be answered incorrectly and well scoring participants were more likely to answer questions correctly was met. The chi-square probability for macronutrients, micronutrients and alcohol was significant (Table 3); this occurred because some items that did not meet the Rasch indicator requirements were kept because they were deemed important to assess in an SNKQ.


The questionnaire

Due to the limitations with existing SNKQs, researchers have raised concerns regarding the accuracy of previous reports of athletes’ nutrition knowledge, and have postulated that the relationship between nutrition knowledge and dietary intake may have been misjudged [8, 14, 41, 42]. The aim of this study was to create an SNKQ that tested awareness of current consensus recommendations, was adequately validated and could be used with athletes from a range of sports. The newly developed tool, the Nutrition for Sport Knowledge Questionnaire (NSKQ), has 89 individual items (44 questions, with some having multiple parts) covering six distinct subsections The questionnaire takes around 25 min to complete and is comparable in length to the GNKQ [40], which has 113 items and the SNKQ developed by Zinn et al. [43], which has 88 items. Since the questions are based on consensus guidelines, several items assess theoretical knowledge. Questions that assess practical knowledge have also been included. The tool is more comprehensive than existing measures as it includes an alcohol sub-section and adheres to a detailed test plan. The questions are based on current guidelines. For example, rather than ask about carbohydrate requirements as % total calorie intake, we have included a question on requirements in g/kg/day and specified the type (‘endurance’) and intensity (‘moderate to high’) of activity. Likewise, in contrast to other existing tools, our hydration question reflects findings that using thirst to judge fluid needs can maximise performance and current recommendations that hydration plans should be tailored to the individual [31, 44]. Moreover, questions on the timing of the pre-completion meal and recovery snack have intentionally been omitted. It is too difficult to assess the correct answer to these questions given the increasing evidence to support the positive benefits of periodizing nutrition based on the goals of individual sessions and overall training schedule [4, 28].

Validity and reliability

The questionnaire has demonstrated face and content validity based on student-athletes and sports dietitians’ judgements. In contrast to previous tools, the content validity has been assessed quantitatively, using a CVI [36]. Individuals who reported undertaking studies in human nutrition achieved higher scores across all sections, except alcohol, indicating that the questionnaire has good construct validity. The group with a nutrition education were younger and were more likely to be female and tertiary educated. In contrast to previous reports, there was no significant difference in performance based on age or gender [41, 40]. Therefore, the variations in knowledge between groups are unlikely to be due to underlying differences in participant characteristics. Test-retest reliability was assessed based on the correlation between individuals’ (who repeated the questionnaire) test scores. A limitation of this method is that motivated individuals may upskill between attempts. The average total score was higher for attempt two. Nevertheless, overall test-retest reliability was high, and all sub-sections, except ‘supplements’, achieved (or approached) adequate test-retest reliability. Participants performed most poorly on the supplement section. Therefore, it is feasible that the supplement test-retest result occurred due to participants guessing answers. The overall internal reliability was very high, and the internal reliability of most sub-sections (except alcohol) reached or approached the requisite 0.7 value. As expected, there appeared to be a relationship between the number of items and KR-20. Streiner [18] recommends that KR-20 be interpreted with caution if there are more than 20 items; the overall scale and macronutrient sub-section exceeded this value.

Based on the Rasch analysis, the overall Item/person interaction statistics were adequate, indicating compliance with the expectations that difficult questions were less likely to be answered correctly, and individuals who performed well overall were more likely to answer individual questions correctly. The questionnaire as a whole was multi-dimensional, but with problematic items removed, each section was shown to be unidimensional. Therefore, sections can be used independently, as required. Where the whole tool is used, sub-sections rather than total score should be reported.

All items were written so that units and food names were generic and likely to be understood by individuals of varying cultural backgrounds; however, additional evaluation is required to confirm the functionality of the tool in groups who differ from the present cohort. The fact that country of birth did not influence scores, and the use of Rasch analysis, which produces questionnaires that are independent of the sample used for validation, give some indication that the tool is likely to also be valid in other groups.


A limitation of this study is that we were unable to calculate response rates because we distributed the questionnaire using Facebook groups and online athlete forums, making total exposure unclear. The completion rates for step 7 (where item analysis and Rasch analysis was undertaken) were relatively low (~45%), but there did not appear to be any relationship between participant characteristics (other than sport played, country of birth) and completion rate (S2). The completion rate for step eight (~66%) was adequate [45]. The sample size is another potential limitation. For step seven, there were 188 responses for a 176 item measure; for step eight, there were 181 responses for a 100 item measure. The target sample size for both studies was 200. However, there is some evidence to recommend that samples as small as 30–50 are appropriate for CTT [46]. Similarly, Chen et al. [47] modelled Rasch with varying sample sizes and found that stable results can be achieved with samples of around 100.

A limitation of the questionnaire itself is that the length may be prohibitive, especially for athletes balancing training and work/study who are often time-poor. In addition, some items were poor discriminators. This was reflected by low item discrimination in CTT and the significant chi-square probability of the micronutrient, macronutrient and alcohol sections. For several questions, the poor item discrimination can be explained by the item’s relatively high or low difficulty index. That is, when a question is answered correctly (or incorrectly) by a large proportion of individuals, the overall range of responses is minimal, and therefore it is hard to achieve a meaningful difference between high and low scoring individuals. Many of these items were kept because they tested important concepts, providing valuable feedback on gaps in knowledge. Item discrimination is worth re-evaluating using larger samples of predominantly athletes (not including nutrition students). Likewise, future studies may focus on creating a short-form tool that can be used for rapid assessment of nutrition knowledge. A short-form tool would be useful in research settings where the correlation between knowledge and other factors is being assessed. A short-form tool may also have utility in the elite setting as a ‘screening’ tool for professionals working with athletes, i.e. to identify individuals who need nutrition education and extra support.

At present the NSKQ has only been validated in an Australian population. Future studies could focus on validation to confirm reliability and validity in other regions.


A key advantage of the questionnaire is that it has been validated using a robust methodology. To our knowledge, this is one of very few NKQ to be assessed against the Rasch Model. Likewise, it is the only tool to assess content validity qualitatively and to assess distractor utility - a distractor that is too obviously wrong will significantly increase the chances of respondents guessing a correct answer; this type of analysis is valuable. Importantly, the authors have considered the limitations of the statistics and accordingly made decisions that focused on the quality of the overall tool. In addition, the questions (and their correct answers) are based on the most recent evidence and recommendations with regards to sports nutrition; they are generalizable to most sports and enable comparison across disciplines. The tool uses food terms and measurement units that are likely to be understood by athletes from a range of countries. Moreover, the tool is detailed and therefore can assess gaps in knowledge. The NSKQ has been designed to be administered online and can provide participants with immediate feedback with regards to correct answers to questions. This is likely to be especially helpful for athletes who do not have access to professional support. The online format provides unique opportunity to direct participants to reputable and relevant resources based on their outcomes.


An 89-point general and sports nutrition knowledge questionnaire with six distinct sub-sections has been developed and validated using multiple relevant methods. Three (weight management, sports nutrition, supplements) of the six sub-sections fit the Rasch model. The steps the researchers have taken to ensure the tool is current and adequately validated were robust, and the questionnaire represents an improvement on available measures. Coaches, scientists and nutrition counsellors will benefit from this tool because it will allow them to target their education based on gaps in athletes’ knowledge. In a team sports setting, the NSKQ may also be useful as a screening tool, to identify players who require additional educational support. Widespread utilisation of the tool in the long-term will allow for more accurate evaluation of nutrition knowledge, education programs and comparisons across athletes of varying genders, ages, education levels, and calibres.



Nutrition knowledge questionnaire


Sports nutrition knowledge questionnaire


  1. Rodriguez NR, DiMarco NM, Langley S. Position of the American dietetic association, dietitians of Canada, and the American college of sports medicine: nutrition and athletic performance. J Am Diet Assoc. 2009;3:509–27.

    Google Scholar 

  2. Kreider RB, Wilborn CD, Taylor L, Campbell B, Almada AL, Collins R, Cooke M, Earnest CP, Greenwood M, Kalman DS, et al. ISSN exercise & sport nutrition review: research & recommendations. J Int Soc Sports Nutr. 2010; doi:10.1186/1550-2783-7-7.

  3. Spronk I, Heaney SE, Prvan T, O'Connor HT. Relationship between general nutrition knowledge and dietary quality in elite athletes. Int J Sport Nutr Exerc Metab. 2015;3:243–51.

    Article  Google Scholar 

  4. Heikura IA, Stellingwerff T, Mero AA, Uusitalo ALT, Burke LM. A mismatch between athlete practice and current sports nutrition guidelines among elite female and male middle-and long-distance athletes. Int J Sport Nutr Exerc Metab. 2017;4:351–60.

  5. Kunkel ME, Bell LB, Luccia BHD. Gem no. 330. Peer nutrition education program to improve nutrition knowledge of female collegiate athletes. J Nutr Educ. 2001;2:114–5.

    Article  Google Scholar 

  6. Collison SB, Kuczmarski MF, Vickery CE. Impact of nutrition education on female athletes. Am J Health Behav. 1996;1:14–23.

    Google Scholar 

  7. Cholewa JM, Landreth A, Beam S, Jones T, MacDonald CJ. The effects of a sports nutrition education intervention on nutritional status, sport nutrition knowledge, body composition, and performance in NCAA division I baseball players. J Int Soc Sports Nutr. 2015. doi:10.1186/1550-2783-12-S1-P44.

  8. Spronk I, Kullen C, Burdon C, O'Connor H. Relationship between nutrition knowledge and dietary intake. Br J Nutr. 2014;10:1713–26.

    Article  Google Scholar 

  9. Alaunyte I, Perry JL, Aubrey T. Nutritional knowledge and eating habits of professional rugby league players: does knowledge translate into practice? J Int Soc Sports Nutr. 2015;12:18.

    Article  PubMed  PubMed Central  Google Scholar 

  10. Clark KS. Sports nutrition counseling: documentation of performance. Top Clin Nutr. 1999;2:34–40.

    Article  Google Scholar 

  11. Parks RB, Helwig D, Dettmann J, Taggart T, Woodruff B, Horsfall K, Brooks MA. Developing a performance nutrition curriculum for collegiate athletics. J Nutr Educ Behav. 2016;6:419–24.

    Article  Google Scholar 

  12. Hamilton CB, Maly MR, Giffin JR, Clark JM, Speechley M, Petrella RJ, Chesworth BM. Validation of the questionnaire to identify knee symptoms (QuIKS) using Rasch analysis. Health Qual Life Outcomes. 2015. doi:10.1186/s12955-015-0358-6.

  13. Andrews MC, Itsiopoulos C. Room for improvement in nutrition knowledge and dietary intake of male football (soccer) players in Australia. Int J Sport Nutr Exerc Metab. 2016;1:55–64.

    Article  Google Scholar 

  14. Trakman G, Forsyth A, Devlin B, Belski R. A systematic review of athletes’ and coaches’ nutrition knowledge and reflections on the quality of current nutrition knowledge measures. Nutr. 2016. doi:10.3390/nu8090570.

  15. Heaney S, O’Connor H, Michael S, Gifford J, Naughton G. Nutrition knowledge in athletes: a systematic review. Int J Sport Nutr Exerc Metab. 2011;21:248–61.

    Article  PubMed  Google Scholar 

  16. Furber MJW, Roberts JD, Roberts MG. A valid and reliable nutrition knowledge questionnaire for track and field athletes. BMC Nutr. 2017. doi:10.1186/s40795-017-0156-0.

  17. Petrillo J, Cano SJ, McLeod LD, Coon CD. Using classical test theory, item response theory, and Rasch measurement theory to evaluate patient-reported outcome measures: a comparison of worked examples. Value Health. 2015;1:25–34.

    Article  Google Scholar 

  18. Streiner DL. Starting at the beginning: an introduction to coefficient alpha and internal consistency. J Pers Assess. 2003;1:99–103.

    Article  Google Scholar 

  19. Lambert SD, Pallant JF, Boyes AW, King MT, Britton B, Girgis A. A Rasch analysis of the Hospital Anxiety and Depression Scale (HADS) among cancer survivors. US: Am Psychol Assoc. 2013;25:379–90.

    Google Scholar 

  20. Tennant A, Conaghan PG. The Rasch measurement model in rheumatology: what is it and why use it? When should it be applied, and what should one look for in a Rasch paper? Arthritis Care Res (Hoboken). 2007;8:1358–62.

    Article  Google Scholar 

  21. Mötteli S, Barbey J, Keller C, Bucher T, Siegrist M. Measuring practical knowledge about balanced meals: development and validation of the brief PKB-7 scale. Eur J Clin Nutr. 2015. doi:10.1038/ejcn.2015.173.

  22. Mötteli S, Barbey J, Keller C, Bucher T, Siegrist M. Measuring practical knowledge about balanced meals: development and validation of the brief PKB-7 scale| NOVA. Univ Newcastle Digit Repository. 2016. doi:10.1038/ejcn.2015.173.

  23. Pallant J. Scale development, Rasch analysis and item response theory. Melbourne Australian consortium for Social & Political Research inc (ACSPRI). Melbourne: ACCPRI; 2016.

    Google Scholar 

  24. Potgieter S. Sport nutrition: a review of the latest guidelines for exercise and sport nutrition from the American College of Sport Nutrition, the International Olympic Committee and the International Society for Sports Nutrition. South Afr J Clin Nutr. 2013;1:6–16.

    Article  Google Scholar 

  25. Buford TW, Kreider RB, Stout JR, Greenwood M, Campbell B, Spano M, Ziegenfuss T, Lopez H, Landis J, Antonio J. International Society of Sports Nutrition position stand: creatine supplementation and exercise. J Int Soc Sports Nutr. 2007. doi:0.1186/1550-2783-4-6.

  26. Goldstein ER, Ziegenfuss T, Kalman D, Kreider R, Campbell B, Wilborn C, Taylor L, Willoughby D, Stout J, Graves BS, Wildman R, Ivy JL, Spano M, Smith AE, Antonio J. International society of sports nutrition position stand: caffeine and performance. J Int Soc Sports Nutr. 2010. doi:10.1186/1550-2783-7-5.

  27. International Olympic Comitte. IOC consensus statement on sports nutrition 2010. Int J Sport Nutr Exerc Metab. 2010;1003:552.

    Google Scholar 

  28. Burke LM, Hawley JA, Wong SH, Jeukendrup AE. Carbohydrates for training and competition. J Sports Sci. 2011;29(sup1):S17–27.

    Article  PubMed  Google Scholar 

  29. Loucks AB, Kiens B, Wright HH. Energy availability in athletes. J Sports Sci. 2011;29(sup1):S7–S15.

    Article  PubMed  Google Scholar 

  30. Phillips SM, Van Loon LJ. Dietary protein for athletes: from requirements to optimum adaptation. J Sports Sci. 2011;29(sup1):S29–38.

    Article  PubMed  Google Scholar 

  31. Shirreffs SM, Sawka MN. Fluid and electrolyte needs for training, competition, and recovery. J Sports Sci. 2011;29(sup1):S39–46.

    Article  PubMed  Google Scholar 

  32. Slater G, Phillips SM. Nutrition guidelines for strength sports: sprinting, weightlifting, throwing events, and bodybuilding. J Sports Sci. 2011;29(sup1):S67–77.

    Article  PubMed  Google Scholar 

  33. Sundgot-Borgen J, Garthe I. Elite athletes in aesthetic and Olympic weight-class sports and the challenge of body weight and body compositions. J Sports Sci. 2011;29(sup1):S101–14.

    Article  PubMed  Google Scholar 

  34. Maughan R, Greenhaff P, Hespel P. Dietary supplements for athletes: emerging trends and recurring themes. J Sports Sci. 2011;29(sup1):S57–66.

    Article  PubMed  Google Scholar 

  35. Trakman GT, Forsyth A, Hoye R, Belski R. Developing and validating a nutrition knowledge questionnaire: Key methods and considerations. Public Health Nutr. 2017. doi:10.1017/S1368980017001471.

  36. Polit DF, Beck CT, Owen SV. Is the CVI an acceptable indicator of content validity? Appraisal and recommendations. Res Nurs Health. 2007;4:459–67.

    Article  Google Scholar 

  37. Krueger RA, Casey MA. Focus groups: a practical guide for applied research. New York: Sage publications; 2014.

    Google Scholar 

  38. Parmenter K, Wardle J. Evaluation and design of nutrition knowledge measures. J Nutr Educ. 2000;5:269–77.

    Article  Google Scholar 

  39. Faul F, Erdfelder E, Lang A-G, Buchner A. G* power 3: a flexible statistical power analysis program for the social, behavioral, and biomedical sciences. Behav Res Methods. 2007;2:175–91.

    Article  Google Scholar 

  40. Parmenter K, Wardle J. Development of a general nutrition knowledge questionnaire for adults. Eur J Clin Nutr. 1999;4:298–308.

    Article  Google Scholar 

  41. Heaney S, O’Connor H, Michael S, Gifford J, Naughton G. Nutrition knowledge in athletes: a systematic review. Int J Sport Nutr Exerc Metab. 2011;3:248–61.

    Article  Google Scholar 

  42. Spendlove JK, Heaney SE, Gifford JA, Prvan T, Denyer GS, O'Connor HT. Evaluation of general nutrition knowledge in elite Australian athletes. Br J Nutr. 2012;12:1871–80.

    Article  Google Scholar 

  43. Zinn C, Schofield G, Wall C. Development of a psychometrically valid and reliable sports nutrition knowledge questionnaire. J Sci Med Sport. 2005;3:346–51.

    Article  Google Scholar 

  44. Goulet ED. Effect of exercise-induced dehydration on time-trial exercise performance: a meta-analysis. Br J Sports Med. 2011; doi:10.1136/bjsm.2010.077966.

  45. Fincham JE. Response rates and responsiveness for surveys, standards, and the journal. Am J Pharm Educ. 2008. doi:0.5688/aj720243

  46. Cappelleri JC, Lundy JJ, Hays RD. Overview of classical test theory and item response theory for quantitative assessment of items in developing patient-reported outcome measures. Clin Ther. 2014;5:648–62.

    Article  Google Scholar 

  47. Chen W-H, Lenderking W, Jin Y, Wyrwich KW, Gelhorn H, Revicki DA. Is Rasch model analysis applicable in small sample size pilot studies for assessing item characteristics? An example using PROMIS pain behavior item bank data. Qual Life Res. 2014;2:485–93.

    Article  Google Scholar 

Download references


We would like to thank all the participants who were involved in the various stages of the study. A special thanks to the Sports Dietitians who took part in the expert panel review, and to Dr. Brooke Lea Devlin, who contributed to an internal review of an early version of the questionnaire.


This work was supported by an Australian Government Research Training Program Scholarship (GT).

Availability of data and materials

The data used/and or analyses during the current study are available from the corresponding author on reasonable request. The final version of the UNSQ is available from the corresponding author.

Author information

Authors and Affiliations



GT designed the study, drafted the first version of the questionnaire, undertook all data analysis, wrote the majority of the manuscript and oversaw the manuscript preparation. AF assisted with the study design, decision making regarding modifications to the questionnaire, and manuscript preparation. RH assisted with the study design and manuscript preparation. RB is the Chief Investigator of the study, assisted with the study design, decision making regarding modifications to the questionnaire, and manuscript preparation. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Gina Louise Trakman.

Ethics declarations

Ethics approval and consent to participate

The research was approved (S16/267) by the La Trobe University’s SHE College Human Ethics Sub-Committee (SHE CHESC). Participants in the expert panel review and think-out-loud focus groups read the Participant Information Statement and signed the Consent Form. Participants who were involved insteps six to eight read the Participant Information Statement and provided consent electronically.

Consent for publication

No individual’s data has been collected or reported.

Competing interests

The authors declare that they have no competing interests.

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Additional file

Additional file 1: Supplementary Material S1.

Methods used for missing variables analysis. In the original design, the demographic questions were asked last; therefore, it was not possible to complete a full assessment of factors that led participants to drop-out. Using available data, missing values (%) were compared across several participant characteristics. The continuous variables age, hours training per week, years playing sport, and country of birth were transformed into categorical variables. Country of birth was transformed into a dichotomous variable (Born in Australia/Not Born in Australia). The dependent variable, missing data, was assessed for normality using the Kolmogorov-Smirnov statistic. Assumptions of normality were violated (Statistic 0.245, df 462, p < 0.001). Therefore non-parametric tests were used; Mann-Whitney was used dichotomous variables (gender [male/female], being born in Australia [yes/no], being paid to play sport [yes/no], previously having undertaken nutrition studies [yes/no], previously having been given advice to change diet [yes/no]) and Kruskal-Wallis was used for categorical variables with three or more response options (sport played, highest level of education, highest level sport played, years playing sport, hours training). Supplementary Material S2. Relationship between missing values (%) and participant characteristics. Supplementary Material S3. Original ‘Universal Sports Nutrition Knowledge Questionnaire’ test plan. Supplementary Material S4. Summary of items that did not meet the requisite 0.3 value for CVI. Supplementary Material S5. Examples of changes made to the questionnaire based on the student athletes’ feedback. Supplementary Material S6. Comparison of scores across participant demographics at step 8. Supplementary Material S7. Examples of changes made based on item analysis using classical test theory at step 7. Supplementary Material S8. Examples of changes made to questionnaire based on CTT and Rasch re-analysis (step 8). (DOCX 36 kb)

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (, which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver ( applies to the data made available in this article, unless otherwise stated.

Reprints and Permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Trakman, G.L., Forsyth, A., Hoye, R. et al. The nutrition for sport knowledge questionnaire (NSKQ): development and validation using classical test theory and Rasch analysis. J Int Soc Sports Nutr 14, 26 (2017).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI:


  • Nutrition knowledge
  • Sports nutrition
  • Questionnaire
  • Measure
  • Valid
  • Classical test theory
  • Rasch analysis