Abstract
Introduction
The aim of this study was to better understand the interfaces of being correct or incorrect and confident or unconfident; aiming to point out misconceptions and assure valuable questions.
Methods
This cross-sectional study was conducted using a convenience sample of second-year dental students (n = 29) attending a preclinical endodontics course. Students answered 20 multiple-choice questions (“basic” or “moderate” level) on endodontics, all of which were followed by one confidence question (scale). Our two research questions were: (1) How was the students’ performance, considering correctness, misconceptions, and level of confidence? (2) Were the questions valuable, appropriate and friendly, and which ones led to misconceptions? Four situations arouse from the interrelationship between question correctness and confidence level: (1st) correct and confident, (2nd) correct and unconfident, (3rd) incorrect and confident (misconception) and (4th) incorrect and unconfident. Statistical analysis (α = 5%) considered the interaction between (a) students’ performance with misconceptions and confidence; (b) question’s difficulty with correctness and confidence; and (c) misconceptions with clinical and negative questions.
Results
Students had 92.5% of correctness and 84.6% of confidence level. Nine students were responsible for the 12 misconceptions. Students who had more misconceptions had lower correctness (P < 0.001). High achieving students had low confidence in their incorrect responses (P = 0.047). ‘Moderate’ questions had more incorrectness (P < 0.05) and less confidence (P = 0.02) than ‘basic’. All questions were considered valuable [for example, the ones that presented images or required a mental picture of a clinical scenario, since they induced less misconception (P = 0.007)]. There was no difference in misconceptions between negative questions and other questions (P = 0.96).
Conclusion
Preclinical endodontic students were highly correct and very confident in their responses. Students who had more misconceptions had also the lowest performance in the assessment. Questions were valuable; but some will worth further improvement for the future. A multiple-choice assessment, when combined with confidence questions, provided helpful information regarding misconceptions and questions value.
Similar content being viewed by others
Introduction
The literature is inefficient to produce a unifying theory on confidence. Many concepts have been described, some of them, focusing on a more broad definition, and others, focusing on confidence in specific tasks. Shrauger and Schohn1 elaborated one of the most comprehensive conventions for understanding confidence. They highlighted sources of confidence—including judgments of actual performance—and stated that confidence is a trait detectable by others in social interactions and activities. Those authors observed how confidence (and confidence ratings) is a product of behavior that feeds into subsequent decisions to engage in a behavior again. As a result, they defined confidence as “perceived (assuredness) in competence, skill or ability”. Lately, authors2 have claimed that, the previous definition conceptually overlaps the notion of “self-efficacy” described by Bandura3 and Zimmerman4. Bandura’s self-efficacy comprises for sources (past achievement; comparison with others; what others tell you and feelings or physical states) and four mediating processes (cognitive; motivational; affective and selective).3 This overlap leads to, sometimes, the use of self-confidence as a framework for (or synonym of) self-efficacy.5
When we think in complete a specific task we may define confidence as the degree of certainty about handling something.6 It is important to remember that confidence is individualist and specific, influenced by different physical environments and conceptual contexts. Self-efficacy is more complex than confidence: when confidence is the degree of certainty in content/statements, self-efficacy is impacted by sensitive and emotional factors related to the certainty in those content/statements.2
Studies investigating the relationship on confidence and accuracy/correctness have found a positive relationship between being confident and being correct.7 In addition, the feeling of confidence, together with notions of self-concept, self-efficacy, and anxiety are reliable indicators of academic achievements.8
Inquiring about student’s confidence in assessments has been reported as a way to identify the different confidence-correctness interactions.9,10,11,12,13 Confidence and correctness in multiple-choice questions may interact in four ways, each with different implications and outcomes for student learning. First, when a student chooses the correct answer in an assessment and is confident that the information is correct, it reflects the calibration between confidence and correctness. This calibration is an important attribute for appropriate clinical decision-making and professional development.7,14,15,16 In a second situation, when the student chooses the correct answer but is unconfident about his/her decision, there is the possibility that the student is guessing the answer. If that was the case, educators may not be able to detect the student’s deficiency, in which could result in the faculty not giving appropriate feedback.17 The third situation would be when the student chooses an incorrect answer and is sure that the answer is the correct one. This characterizes a misconception. In this case, the student is less likely to seek additional opinions before a decision-making (in a health-related treatment planning, for instance) and, therefore, is unaware of his/her responsibility of adverse outcomes that can be induced in patients.14 Often, misconceptions are difficult to detect and are resistant to change.18 Students can ignore or distort new information that conflicts with their previous erroneous thinking.19 The fourth situation is when the student chooses the incorrect answer and, at the same time, is unconfident about the chosen answer. This situation shows that the student is uninformed about the concept, but, at least, is aware of his/her deficiencies. In this situation, the student is usually open-minded and receptive to learning.9
Understanding the interrelationship between confidence and correctness helps educators in identifying uninformed and misinformed students, and, the topics that induce the most misunderstandings. Consequently, educators can provide feedback to students regarding their specific lack of knowledge, for example, pointing out a misconception and an implied etiology.14,17,20,21 For students, this can encourage the development of self-reflection and foster critical thinking.14,20 For example, one studied used a confidence score algorithm in multiple-choice examinations whereby leaving a question unanswered (admitting being unsure) rewarded students fractional points.12 In addition to that, the combination of confidence questions with multiple-choice questions with the aim of validating and increasing the standards of questions in assessments is a potential advantage. It is known that writing high standard multiple-choice questions is efficient in enhancing the cognitive levels of assessments.22,23,24
Our previous investigations, conducted in the United States, have already explored dentistry students’ confidence in assessments and were able to identify misconceptions in different disciplines, and showed that students were overconfident.14,20,21 Following those investigations, we found important to understand how generalizable our earlier results are to other environments and contexts (Canada). Therefore, this study was conducted with a second-year class of dentistry students whose answered multiple-choice questions combined with confidence questions. This allowed us to better understand the interfaces of being correct or incorrect and confident or unconfident; aiming to point out misconceptions and assure valuable questions.
Material and methods
The protocol for this study was reviewed and approved by the appropriated University Behavioral Research Ethics Board (Beh-REB App ID #50).
Study design and sample
This cross-sectional study was conducted using a convenience sample of second-year dental students from a Canadian dental school attending a preclinical endodontics course. The entire class (n = 31 students: 16 women and 15 men) was invited to participate and those who agreed to join the study signed a consent form.
Assessment instrument
The assessment instrument consisted of 20 multiple-choice questions on endodontics, all of which were followed by one confidence question. This research wanted to answer the following inquires: (1) How was the students’ performance, considering correctness, misconceptions, and level of confidence? (2) Were the questions valuable, appropriate and friendly, and which ones led to misconceptions?
Each multiple-choice question had four alternatives (a, b, c, and d) and only one option was correct. Questions were developed according to the subjects addressed in class and posted online. Preclinical endodontics embrace technical topics regarding root canal treatment: access opening, instrumentation, intracanal medication, irrigation and obturation. The 20 multiple-choice questions had been previously designated in “basic” or “moderate”, according to the judgment of two faculty members. Item analysis (such as, difficulty index, discrimination index or distracter efficiency) was outside the scope of this study.
The confidence question asked “how confident are you about the accuracy/correctness of your chosen answer?” The student should point one of the options displayed in a horizontal visual scale: very unsure, unsure, sure, or very sure. Further, these options were dichotomized as ‘confident’ (sure, very sure) and ‘unconfident’ (very unsure, unsure). Students were advised that confidence levels would be an important factor to consider; however, if they preferred not to answer the confidence question, their grades would not be affected. Faculty members were blinded to the students’ identification on the assessment until the end of the course.
Four situations arouse from the interrelationship between question correctness and confidence level: (1st) correct and confident, (2nd) correct and unconfident, (3rd) incorrect and confident (misconception), and (4th) incorrect and unconfident.
Data analysis
Aiming to answer our first research inquiry, we measured the frequency/percentage of the four situations described above in relation to the student. Statistical analysis was performed for the following interactions: (i) students’ performance × misconceptions and (ii) students’ performance × confidence. Aiming to answer our second research inquiry, we measured the frequency/percentage of the four situations, in relation to the multiple-choice questions. Statistical analysis was performed for the following interactions: (i) questions level of difficulty × correctness, (ii) questions level of difficulty and correctness × confidence, (iii) negative questions × misconceptions, and (iv) mental scenario/clinical questions × misconceptions.
Data analysis was performed with Chi-square and Mann–Whitney’s tests using Open Source Epidemiologic Statistics for Public Health (OpenEpi) Version 3.01 (www.OpenEpi.com) and Jamovi (www.jamovi.org), respectively. Significance was set at α = 5%.
Results
Students’ performance (correctness, confidence, and misconceptions)
From the 31 students who were invited to the study, one did not agree to participate (did not sign the consent form and did not answer the confidence questions) and another was excluded from the analysis since he/she left many confidence questions unanswered; resulting in 29 participants. Nine students (31.03%) were responsible for the 12 misconceptions responses detected. Students who had more misconceptions had also the lowest performance/correctness in the assessment (Mann–Whitney’s test, P < 0.001). High achieving students (the ones with scores of 85–100% on the assessment) were unsure in their incorrect responses (this means that students with high performance had more ‘situation 4’ than the rest of them) (Chi-square test, P = 0.047). Table 1 shows some additional descriptive the findings in relation to the students.
Value of questions
Tables 2 and 3 show the findings in relation to the multiple-choice questions. ‘Basic’ questions had 95.17% of all correct answers. ‘Moderate’ questions had 67.44% of all incorrect answers. Moderate questions (as opposed to basic questions) induced the majority of incorrect answers (Chi-square test: P < 0.05). In addition, moderate questions also prompted lower confidence level, even when students responded correctly (‘situation 2’) (Chi-square test, P = 0.02).
Table 4 shows examples of multiple-choice questions and important findings demonstrating their value (in other words, if they were appropriate, understandable, and friendly). We classified as valuable all 20 multiple-choice questions. However, the redesign/improvement of some questions for future assessments would be interesting, for instance: the five questions with negatively worded stems had 3 misconceptions from the 144 responses that were possible. Comparing to the rest of the questions (9 misconceptions from the 426 possible responses), the negative ones seemed to be misleading. However, there was not statistical difference in misconceptions between negative questions and the rest of them (Chi-square test, P = 0.96).
At last, questions connecting basic science principles to clinical application (as question 20 for instance), induced some unconfidence. Interesting, questions using images and questions that required a mental picture of a clinical scenario had less number of misconceptions compared to theoretical/descriptive questions (Fisher’s exact test, P = 0.007).
Table 5 summarizes the statistical analysis and shows the main results.
Discussion
This study used a multiple-choice assessment to better understand the interrelationship between being confident (or not) in a chosen answer. Our study used multiple-choice questions combined with confidence scales aiming to point out misconceptions and to assure having valuable, understandable and friendly questions. With the use of confidence scales, we acquired unique information on students’ learning, mainly detecting the ones who were more likely to think critically and identifying the ones who needed additional support. Also, this assessment/study provided us knowledge about the value of multiple-choice questions, helping to improve future assessments.
Even without having a traditional multiple-choice assessment (with no confidence scale) to act as a group comparison in our study, it was undoubtedly, that adding confidence scales provided a more complete assessment. Literature has proved that confidence-based assessment techniques integrate the selection of multiple-choice answers with the student’s self-perceived level of certainty and offer a middle ground between the traditional multiple-choice answer and a lengthy essay response. Research has discovered that confidence-based assessments provide a more comprehensive measure of a person’s knowledge, increases the retainability of learned material and identifies topics in which people are misinformed.25,26,27 In addition, confidence scales also helped us to determine the value of questions, since we could verify if faculty’s understanding on the question was in line with students’ understanding (“basic” or “moderate” level of difficulty).
Having learned from previous studies, we expected that students would exhibit high confidence levels.13,21,28 In addition to foreseeing the students’ overconfidence, we also anticipated high percentage of correctness. Indeed, as expected, we found preclinical endodontic students highly correct (92.5%) and very confident in their responses (84.6%). Students who had more misconceptions had also the lowest performance in the assessment (statistically significant difference). One reason for the high accuracy was the relatively straightforward assessment that we have chosen. The nature of preclinical instruction in our academic setting is technical and focused on modus operandi, on simple recall of information and on developing hand skills; which requires, therefore, an assessment in consonance with. We are aware that straightforward and uncomplicated assessments may sometimes not be the ideal test to perform a discrimination power analysis (which was outside the scope of this study) and may not provide sufficient insights to measure students’ knowledge.29 Another reason for students being highly accurate in their answers is that this study was conducted with a small class, where students were able to receive individual and personalized attention; and the same faculty member delivered all the theoretical content. A single lecturer eliminates the challenges of calibration that might occur when several mentors are involved in the course. Since non-calibrated mentors may not apply the same standardization in protocols, techniques and philosophies,30,31 it could raise students’ misinterpretations and misconceptions. On the other hand, it is important to remember that, one disadvantage of having only one tutor, might be that it does not favor different types of learners, if the tutor uses only one modality of teaching.
The calibration between correctness and confidence is the ideal situation that fosters clever decision-making in a dental clinical setting and positively contributes to professional growth.14,15,16,21 The literature has shown the importance of increasing confidence through learning. Authors, by using a mixed-methods approach, proved how changes in confidence have positively influenced the clinical practice of two cohorts of general dental practitioners—demonstrating the following results: the ability to undertake more complex patient treatment with confidence; a realization that previously unattainable skills were attainable; a greater appreciation of one’s limitations; and a greater sense of job satisfaction, which was reflected by patient acceptance and contentment.32
Nonetheless, other evidences point in favor of a “medium” confidence level, saying that high confidence is not always appropriate. In isolation, overconfidence may be seen as undesirable professional characteristic as it is linked to a lack of accuracy and potential reckless behavior.13 In our study, we found high achieving students (the ones with scores equal or higher than 85% on the assessment) were still unsure in their incorrect responses (statistically significant). This is a situation where being unconfident is somehow worthwhile. It shows that although students did not know the answer, they were aware of their uncertainty. In other words, they were able to self-reflect on their uncertainty. Authors have stated that uncertainty can favor apprenticeship.9 If these dental students move from preclinical to the clinical setting with a ‘medium confidence’ in unfamiliar topics they will be open-minded and more prompt to seek tutoring, which may avoid clinical errors, unnecessary procedures, and harmfulness to the patient.14,21
Some correct answers were linked to under-confidence (unsure or very unsure). If this situation occurs in a traditional multiple-choice assessment (without confidence questions), it may lead to mentors to believe that students know the answer, when they were actually guessing.20,33 In our assessment, because of the confidence scale, we were able to provide feedback on both incorrect and correct student responses.
‘Moderate’ questions (as opposed to ‘basic’ questions) induced the majority of incorrect answers. These questions also prompted lower confidence levels, even when students responded correctly. Both results were statistically significant. Question 20, for instance, is a ‘moderate’ level and ‘requires a mental picture of a clinical scenario’ (see Table 4 for more information). Although it is composed of a very simple statement, students needed to mentally create the situation about access opening, connecting theoretical and clinical information. As the students-participants were in the preclinical level, they did not yet have clinical experience. Therefore, it is natural that they still have difficulties in transferring information from a preclinical to a clinical setting. To overcome this challenge of interfacing theory and technique, one idea is to use videos showing real time endodontic procedures,34 or to provide preclinical students with the opportunity of ‘shadowing’ more advanced-level peers in the clinical setting. On the other hand, the statistical analysis showed that questions requiring clinical scenario induced significant less number of misconceptions—which, therefore, made us to consider these types of questions valuable. Another type of valuable question, deserving to be used in further assessments, was the one that employed images/drawings. Question 5, in comparison to question 14, (see Table 4) facilitated the understanding of an endodontic opening access design, having significant less misconceptions. Therefore, we assumed that visual questions may benefit dentistry students, since they usually encounter a professional work environment that exploits images (photos, videos, radiographs, anatomic features, graphical abstracts, schemes, etc.).
Adversely, we also found questions that will need to be revised and improved for next assessments. And, although there was not statistical difference in misconceptions between negative word stem and the other questions, we did find a considerable number of misconceptions in negative questions. It is already known that negative questions might be potentially ineffective and misinterpreted.35,36 In here, we noticed that this type of question induced error, independent of confidence.
The strengths of this current study included the use of authentic data from an academic assessment, which helps assure students were appropriately engaged. Also, we were able to provide personal feedback for each learner—mainly the ones who got misconceptions. Limitations include having a convenience sample and having results from just one assessment. Our results might not be generalizable to other institutions; however, the findings about the questions may have external validity, since they showed inherent positive and negative characteristics of endodontic preclinical questions. Also, the inherent imprecision of the measurement of the confidence level—that is always a multifaceted and subjective measurement;37 and the multiple-choice assessment—that might foster study habits consistent with superficial or strategic learning (only to answer the questionnaire correctly). When this happen, the student would interested in recognizing isolated facts rather than relationships or higher levels of learning.38
As conclusion, we found that preclinical endodontic students were highly correct (92.5%) and very confident in their responses (84.6%). Students who had more misconceptions had also the lowest performance in the assessment. All questions were considered valuable; but some will need further improvement. This study showed that understanding the interfaces between confidence and correctness grant us with the possibility to point out students’ misconceptions (guaranteeing unique feedback for each student) and to assure we have valuable questions for current and future endodontic preclinical assessments.
References
Shrauger, J. S. & Schohn, M. Self-confidence in college students: conceptualization, measurement and behavioral implications. Assessment 2, 255–78 (1995).
Cramer, R. J., Neal, T. M. S. & Brodsky, S. L. Self-efficacy and confidence: theoretical distinctions and implications for trial consultation. Consulting Psychol. J. 61, 319–334 (2009).
Bandura A. Self-efficacy: The exercise of control. (Freeman, New York, 1997).
Zimmerman, B. J. Self-efficacy: an essential motive to learn. Contem. Educ. Psychol. 25, 82–91 (2000).
Parsons, S., Croft, T. & Harrison, M. Engineering students’ self confidence in mathematics mapped onto Bandura’s self-efficacy. Eng. Educ. 6, 52–61 (2011).
Stajkovic, A. D. Development of a core confidence–higher order construct. J. Appl. Psychol. 91, 1208–24 (2006).
Brewer, W. F., Sampaio, C. & Barlow, M. R. Confidence and accuracy in the recall of deceptive and nondeceptive sentences. J. Mem. Lang. 52, 618–27 (2005).
Stankov, L., Lee, J., Luo, W. & Hogan, D. J. Confidence: a better predictor of academic achievement than self-efficacy, self-concept and anxiety? Learn. Individ. Differ. 22, 747–58 (2012).
Wakabayshi, T. & Guskin, K. The effect of an unsure option on early childhood professionals pre-and post-training knowledge assessments. Am. J. Eval. 31, 486–98 (2010).
Ibabe, I. & Sporer, S. L. How you ask is what you get: on the influence of question form on accuracy and confidence. Appl. Cognit. Psychol. 18, 711–26 (2004).
Barr, D. A. & Burke, J. R. Using confidence-based marking in a laboratory setting: a tool for student self-assessment and learning. J. Chiropr. Educ. 27, 21–6 (2013).
McMahan, C. A., Pinckard, R. N., Jones, A. C. & Hendricson, W. D. Fostering dental student self assessment of knowledge by confidence scoring of multiple-choice examinations. J. Dent. Educ. 78, 1643–54 (2014).
Stankov, L. & Lee, J. Overconfidence across world regions. J. Cross Cult. Psychol. 45, 821–37 (2014).
Curtis, D. A., Lind, S. L., Boscardin, C. K. & Dellinges, M. Does student confidence on multiple choice question assessments provide useful information? Med. Educ. 47, 578–84 (2013).
Mamede, S., Schmidt, H. G. & Rikers, R. Diagnostic errors and reflective practice in medicine. J. Eval. Clin. Pract. 13, 138–45 (2007).
Tweed, M., Thompson-Fawcett, M., Schwartz, P. & Wilkinson, T. J. Determining measures of insight and foresight from responses to multiple-choice questions. Med. Teach. 35, 127–33 (2013).
Curtis, D. A. What do you know and how confident are you that you know it. Int. J. Oral Maxillofac. Implants 28, 327–8 (2013).
Klymkowsky, M., Taylor, L. & Spindler, S. Garvin-Doxas R. Two-dimensional, implicit confidence tests as a tool for recognizing student misconceptions. J. Coll. Sci. Teach. 36, 44–8 (2006).
Kohnle, A., Mclean, S. & Aliotta, M. Towards a conceptual diagnostic survey in nuclear physics. Eur. J. Phys. 32, 55–62 (2011).
Curtis, D. A., Lind, S. L., Dellinges, M. & Schroeder, K. Identifying student misconceptions in biomedical course assessments in dental education. J. Dent. Educ. 76, 1183–94 (2012).
Grazziotin-Soares, R., Lind, S. L., Ardenghi, D. M. & Curtis, D. A. Misconceptions amongst dental students: How can they be identified? Eur. J. Dent. Educ. 22, e.101–e.106 (2017).
Capan Melser, M. et al. Knowledge, application and how about competence? Qualitative assessment of multiple-choice questions for dental students. Med. Educ. Online 25, 1714199 (2020).
Dellinges, M. A. & Curtis, D. A. Will a Short Training Session Improve Multiple-Choice Item-Writing Quality by Dental School Faculty? A Pilot Study. J. Dent. Educ. 81, 948–55 (2017).
Shaikh, S., Kannan, S. K., Naqvi, Z. A., Pasha, Z. & Ahamad, M. The role of faculty development in improving the quality of multiple-choice questions in dental education. J. Dent. Educ. 84, 316–22 (2020).
Khatibi, R., Ghadermarzi, M., Yazdani, S. & Zarezadeh, Y. Assessment of the relation between students’ gender and their scores on selecting confidence choices in confidence-based exams. Iran J. Med. Educ. 11, 926–32 (2012).
Yen, Y.-C., Ho, R.-G., Chen, L.-J., Chou, K.-Y. & Chen, Y.-L. Development and evaluation of a confidence-weighting computerized adaptive testing. Educ. Tech Society 13, 163–76 (2010).
Novacek, P., Confidence-based assessments within an adult learning environment. IADIS International Conference on Cognition and Exploratory Learning in Digital Age (CELDA 2013). https://files.eric.ed.gov/fulltext/ED562245.pdf
Stankov, L., Lee, J. Too many school students are overconfident. The conversation US pilot. Online information. www.theconversation.com
Cornachione, Jr. & Edgard, B. Objective tests and their discriminating power in business courses: a case study. Braz. Admin. Rev. 2, 63–78 (2005).
Lane, B. A. et al. Assessment of the calibration of periodontal diagnosis and treatment planning among dental students at three dental schools. J. Dent. Educ. 79, 16–24 (2015).
Oh, S.-L., Liberman, L. & Mishler, O. Faculty calibration and students’ self-assessments using an instructional rubric in preparation for a practical examination. Eur. J. Dent. Educ. 22, e-400–e-407 (2018).
Fine, P., Leung, A., Bentall, C. & Louca, C. The impact of confidence on clinical dental practice. Eur. J. Dent. Educ. 23, 159–67 (2019).
Butler, A. C. & Roediger, H. L. Feedback enhances the positive effects and reduces the negative effects of multiple-choice testing. Mem. Cognit. 36, 604–16 (2008).
Edrees, H. Y., Ohlin, J., Ahlquist, M., Tessma, M. K. & Zary, N. Patient demonstration videos in predoctoral endodontic education: aspects perceived as beneficial by students. J. Dent. Educ. 79, 928–33 (2015).
Campbell, D. E. How to write good multiple-choice questions. J. Paediatr. Child Health 47, 322–325 (2011).
Case, S. M., Swanson, D. B. Constructing written test questions for the basic and clinical sciences. 3rd Ed. National Board of Medical Examiners: Philadelphia PA. http://www.medbev.umontreal.ca/docimo/DocSource/NBME_MCQ.pdf
Inayah, A. T. et al. Objectivity in subjectivity: do students’ self and peer assessments correlate with examiners’ subjective and objective assessment in clinical skills? A prospective study. BMJ Open 7, e012289 (2017).
Entwistle, N., Skinner, D., Entwistle, D. & Orr, S. Conceptions and beliefs about “good teaching”: an integration of contrasting research areas. Higher Educ Res. Development 19, 5–26 (2000).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Grazziotin-Soares, R., Blue, C., Feraro, R. et al. The interrelationship between confidence and correctness in a multiple-choice assessment: pointing out misconceptions and assuring valuable questions. BDJ Open 7, 10 (2021). https://doi.org/10.1038/s41405-021-00067-4
Received:
Revised:
Accepted:
Published:
Version of record:
DOI: https://doi.org/10.1038/s41405-021-00067-4
This article is cited by
-
Impact of self-confidence and personality on dental students’ radiographic interpretation skills
BMC Medical Education (2026)
-
The Relationship between Performance and Confidence in the Secondary-School Version of the Test of Calculus and Vectors in Mathematics and Physics
International Journal of Science and Mathematics Education (2025)
-
BDJ Open 2021 - our most successful year to date
British Dental Journal (2022)


