Abstract
National student feedback surveys are administered in a number of countries, and several of these encompass both campus‐based and distance learning students. The UK Open University achieves a high ranking in the annual National Student Survey (NSS), but there are some anomalies in the results. The NSS questionnaire was administered to three samples of students to explore the origins of these anomalies. The reliability and the validity of the questionnaire were satisfactory, but a substantial proportion of the respondents did not produce valid data. There was a high proportion of ‘not applicable’ responses to items concerned with learning resources and personal development. Other students responded ‘neither agree nor disagree’ to these items when ‘not applicable’ would have been more appropriate. This would have led to low rankings on these scales in the national surveys. Distance learning institutions need to advise their students about the importance of only rating items they have actually experienced or to negotiate changes to the questionnaires themselves if they are to ensure that their activities and achievements are properly represented in national student feedback surveys.
Acknowledgements
The authors are grateful to Nick Haycox for identifying the samples of students to be used in this investigation and to his staff in the Open University’s Survey Office for distributing the surveys and for processing the responses. They are also grateful to Don Dillman and Fern Willits for their advice and to Paul Ginns and Tarla Shah for their comments on an earlier version of this article.