skip to main content
10.1145/1953163.1953200acmconferencesArticle/Chapter ViewAbstractPublication PagessigcseConference Proceedingsconference-collections
research-article

The FCS1: a language independent assessment of CS1 knowledge

Published: 09 March 2011 Publication History

Abstract

A primary goal of many CS education projects is to determine the extent to which a given intervention has had an impact on student learning. However, computing lacks valid assessments for pedagogical or research purposes. Without such valid assessments, it is difficult to accurately measure student learning or establish a relationship between the instructional setting and learning outcomes.
We developed the Foundational CS1 (FCS1) Assessment instrument, the first assessment instrument for introductory computer science concepts that is applicable across a variety of current pedagogies and programming languages. We applied methods from educational and psychological test development, adapting them as necessary to fit the disciplinary context. We conducted a large scale empirical study to demonstrate that pseudo-code was an appropriate mechanism for achieving programming language independence. Finally, we established the validity of the assessment using a multi-faceted argument, combining interview data, statistical analysis of results on the assessment, and CS1 exam scores.

References

[1]
American Educational Research Association, American Psychological Association, and National Council on Measurement in Education. Standards for educational and psychological testing. American Educational Research Association, Washington, DC, 1999.
[2]
J. Cohen. Statistical Power Analysis for the Behavioral Sciences. Lawrence Erlbaum Associates, Hillsdale, NJ, 2nd edition, 1988.
[3]
A. M. Decker. How Students Measure Up: An Assessment Instrument for Introductory Computer Science. PhD thesis, University at Bu ffalo (SUNY), Buff alo, NY, 2007.
[4]
K. Goldman, P. Gross, C. Heeren, G. Herman, L. Kaczmarczyk, M. C. Loui, and C. Zilles. Identifying important and difficult concepts in introductory computing courses using a Delphi process. In SIGCSE '08: Proceedings of the 39th ACM Technical Symposium on Computer Science Education, pages 256--260, 2008.
[5]
R. K. Hambleton, H. Swaminathan, and H. J. Rogers. Fundamentals of item response theory. Sage Publications, Newbury Park, CA, 1991.
[6]
D. Hestenes, M. Wells, and G. Swackhamer. Force concept inventory. The Physics Teacher, 30:141--158, March 1992.
[7]
L. C. Kaczmarczyk, E. R. Petrick, J. P. East, and G. L. Herman. Identifying student misconceptions of programming. In SIGCSE '10: Proceedings of the 41st ACM Technical Symposium on Computer Science Education, 2010.
[8]
M. T. Kane. Validation. In R. L. Brennen, editor, Educational Measurement, pages 17--64. American Council on Education/Praeger Publishers, Westport, CT, 4th edition, 2006.
[9]
J. C. Libarkin and S. Anderson. Assessment of learning in entry-level geoscience courses: Results from the geoscience concept inventory. Journal of Geoscience Education, 53:394--401, 2005.
[10]
M. D. Miller, R. L. Linn, and N. E. Gronlund. Validity. In Measurement and assessment in teaching, pages 80--99. Pearson Education, Upper Saddle River, NJ, 10th edition, 2009.
[11]
P. A. Moss, B. J. Girard, and L. C. Haniford. Validity in Educational Assessment. Review of Research in Education, 30(1):109--162, 2006.
[12]
A. E. Tew. Assessing fundamental introductory computing concept knowledge in a language independent manner. PhD thesis, Georgia Institute of Technology, Atlanta, GA, 2010.
[13]
A. E. Tew and M. Guzdial. Developing a validated assessment of fundamental CS1 concepts. In SIGCSE '10: Proceedings of the 41st ACM Technical Symposium on Computer Science education, 2010.

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
SIGCSE '11: Proceedings of the 42nd ACM technical symposium on Computer science education
March 2011
754 pages
ISBN:9781450305006
DOI:10.1145/1953163
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 09 March 2011

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. CS1
  2. assessment
  3. programming
  4. validity

Qualifiers

  • Research-article

Conference

SIGCSE '11
Sponsor:

Acceptance Rates

SIGCSE '11 Paper Acceptance Rate 107 of 315 submissions, 34%;
Overall Acceptance Rate 1,595 of 4,542 submissions, 35%

Upcoming Conference

SIGCSE TS 2025
The 56th ACM Technical Symposium on Computer Science Education
February 26 - March 1, 2025
Pittsburgh , PA , USA

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)89
  • Downloads (Last 6 weeks)14
Reflects downloads up to 23 Dec 2024

Other Metrics

Citations

Cited By

View all

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media