Introduction

The literature is inefficient to produce a unifying theory on confidence. Many concepts have been described, some of them, focusing on a more broad definition, and others, focusing on confidence in specific tasks. Shrauger and Schohn1 elaborated one of the most comprehensive conventions for understanding confidence. They highlighted sources of confidence—including judgments of actual performance—and stated that confidence is a trait detectable by others in social interactions and activities. Those authors observed how confidence (and confidence ratings) is a product of behavior that feeds into subsequent decisions to engage in a behavior again. As a result, they defined confidence as “perceived (assuredness) in competence, skill or ability”. Lately, authors2 have claimed that, the previous definition conceptually overlaps the notion of “self-efficacy” described by Bandura3 and Zimmerman4. Bandura’s self-efficacy comprises for sources (past achievement; comparison with others; what others tell you and feelings or physical states) and four mediating processes (cognitive; motivational; affective and selective).3 This overlap leads to, sometimes, the use of self-confidence as a framework for (or synonym of) self-efficacy.5

When we think in complete a specific task we may define confidence as the degree of certainty about handling something.6 It is important to remember that confidence is individualist and specific, influenced by different physical environments and conceptual contexts. Self-efficacy is more complex than confidence: when confidence is the degree of certainty in content/statements, self-efficacy is impacted by sensitive and emotional factors related to the certainty in those content/statements.2

Studies investigating the relationship on confidence and accuracy/correctness have found a positive relationship between being confident and being correct.7 In addition, the feeling of confidence, together with notions of self-concept, self-efficacy, and anxiety are reliable indicators of academic achievements.8

Inquiring about student’s confidence in assessments has been reported as a way to identify the different confidence-correctness interactions.9,10,11,12,13 Confidence and correctness in multiple-choice questions may interact in four ways, each with different implications and outcomes for student learning. First, when a student chooses the correct answer in an assessment and is confident that the information is correct, it reflects the calibration between confidence and correctness. This calibration is an important attribute for appropriate clinical decision-making and professional development.7,14,15,16 In a second situation, when the student chooses the correct answer but is unconfident about his/her decision, there is the possibility that the student is guessing the answer. If that was the case, educators may not be able to detect the student’s deficiency, in which could result in the faculty not giving appropriate feedback.17 The third situation would be when the student chooses an incorrect answer and is sure that the answer is the correct one. This characterizes a misconception. In this case, the student is less likely to seek additional opinions before a decision-making (in a health-related treatment planning, for instance) and, therefore, is unaware of his/her responsibility of adverse outcomes that can be induced in patients.14 Often, misconceptions are difficult to detect and are resistant to change.18 Students can ignore or distort new information that conflicts with their previous erroneous thinking.19 The fourth situation is when the student chooses the incorrect answer and, at the same time, is unconfident about the chosen answer. This situation shows that the student is uninformed about the concept, but, at least, is aware of his/her deficiencies. In this situation, the student is usually open-minded and receptive to learning.9

Understanding the interrelationship between confidence and correctness helps educators in identifying uninformed and misinformed students, and, the topics that induce the most misunderstandings. Consequently, educators can provide feedback to students regarding their specific lack of knowledge, for example, pointing out a misconception and an implied etiology.14,17,20,21 For students, this can encourage the development of self-reflection and foster critical thinking.14,20 For example, one studied used a confidence score algorithm in multiple-choice examinations whereby leaving a question unanswered (admitting being unsure) rewarded students fractional points.12 In addition to that, the combination of confidence questions with multiple-choice questions with the aim of validating and increasing the standards of questions in assessments is a potential advantage. It is known that writing high standard multiple-choice questions is efficient in enhancing the cognitive levels of assessments.22,23,24

Our previous investigations, conducted in the United States, have already explored dentistry students’ confidence in assessments and were able to identify misconceptions in different disciplines, and showed that students were overconfident.14,20,21 Following those investigations, we found important to understand how generalizable our earlier results are to other environments and contexts (Canada). Therefore, this study was conducted with a second-year class of dentistry students whose answered multiple-choice questions combined with confidence questions. This allowed us to better understand the interfaces of being correct or incorrect and confident or unconfident; aiming to point out misconceptions and assure valuable questions.

Material and methods

The protocol for this study was reviewed and approved by the appropriated University Behavioral Research Ethics Board (Beh-REB App ID #50).

Study design and sample

This cross-sectional study was conducted using a convenience sample of second-year dental students from a Canadian dental school attending a preclinical endodontics course. The entire class (n = 31 students: 16 women and 15 men) was invited to participate and those who agreed to join the study signed a consent form.

Assessment instrument

The assessment instrument consisted of 20 multiple-choice questions on endodontics, all of which were followed by one confidence question. This research wanted to answer the following inquires: (1) How was the students’ performance, considering correctness, misconceptions, and level of confidence? (2) Were the questions valuable, appropriate and friendly, and which ones led to misconceptions?

Each multiple-choice question had four alternatives (a, b, c, and d) and only one option was correct. Questions were developed according to the subjects addressed in class and posted online. Preclinical endodontics embrace technical topics regarding root canal treatment: access opening, instrumentation, intracanal medication, irrigation and obturation. The 20 multiple-choice questions had been previously designated in “basic” or “moderate”, according to the judgment of two faculty members. Item analysis (such as, difficulty index, discrimination index or distracter efficiency) was outside the scope of this study.

The confidence question asked “how confident are you about the accuracy/correctness of your chosen answer?” The student should point one of the options displayed in a horizontal visual scale: very unsure, unsure, sure, or very sure. Further, these options were dichotomized as ‘confident’ (sure, very sure) and ‘unconfident’ (very unsure, unsure). Students were advised that confidence levels would be an important factor to consider; however, if they preferred not to answer the confidence question, their grades would not be affected. Faculty members were blinded to the students’ identification on the assessment until the end of the course.

Four situations arouse from the interrelationship between question correctness and confidence level: (1st) correct and confident, (2nd) correct and unconfident, (3rd) incorrect and confident (misconception), and (4th) incorrect and unconfident.

Data analysis

Aiming to answer our first research inquiry, we measured the frequency/percentage of the four situations described above in relation to the student. Statistical analysis was performed for the following interactions: (i) students’ performance × misconceptions and (ii) students’ performance × confidence. Aiming to answer our second research inquiry, we measured the frequency/percentage of the four situations, in relation to the multiple-choice questions. Statistical analysis was performed for the following interactions: (i) questions level of difficulty × correctness, (ii) questions level of difficulty and correctness × confidence, (iii) negative questions × misconceptions, and (iv) mental scenario/clinical questions × misconceptions.

Data analysis was performed with Chi-square and Mann–Whitney’s tests using Open Source Epidemiologic Statistics for Public Health (OpenEpi) Version 3.01 (www.OpenEpi.com) and Jamovi (www.jamovi.org), respectively. Significance was set at α = 5%.

Results

Students’ performance (correctness, confidence, and misconceptions)

From the 31 students who were invited to the study, one did not agree to participate (did not sign the consent form and did not answer the confidence questions) and another was excluded from the analysis since he/she left many confidence questions unanswered; resulting in 29 participants. Nine students (31.03%) were responsible for the 12 misconceptions responses detected. Students who had more misconceptions had also the lowest performance/correctness in the assessment (Mann–Whitney’s test, P < 0.001). High achieving students (the ones with scores of 85–100% on the assessment) were unsure in their incorrect responses (this means that students with high performance had more ‘situation 4’ than the rest of them) (Chi-square test, P = 0.047). Table 1 shows some additional descriptive the findings in relation to the students.

Table 1 Correctness, confidence level, and other important findings in relation to the students.

Value of questions

Tables 2 and 3 show the findings in relation to the multiple-choice questions. ‘Basic’ questions had 95.17% of all correct answers. ‘Moderate’ questions had 67.44% of all incorrect answers. Moderate questions (as opposed to basic questions) induced the majority of incorrect answers (Chi-square test: P < 0.05). In addition, moderate questions also prompted lower confidence level, even when students responded correctly (‘situation 2’) (Chi-square test, P = 0.02).

Table 2 Interrelationship between confidence and correctness and other important findings in relation to the questions.
Table 3 Topic addressed on the question, question designation (degree of difficulty attributed by faculty members), percentage (%) of correctness, confidence, and % for each ‘Situation’ (interrelationship between confidence and correctness).

Table 4 shows examples of multiple-choice questions and important findings demonstrating their value (in other words, if they were appropriate, understandable, and friendly). We classified as valuable all 20 multiple-choice questions. However, the redesign/improvement of some questions for future assessments would be interesting, for instance: the five questions with negatively worded stems had 3 misconceptions from the 144 responses that were possible. Comparing to the rest of the questions (9 misconceptions from the 426 possible responses), the negative ones seemed to be misleading. However, there was not statistical difference in misconceptions between negative questions and the rest of them (Chi-square test, P = 0.96).

Table 4 Examples of multiple-choice questions, which were part of the endodontic assessment and important findings indicating if the question is valuable (appropriate, understandable, and friendly).

At last, questions connecting basic science principles to clinical application (as question 20 for instance), induced some unconfidence. Interesting, questions using images and questions that required a mental picture of a clinical scenario had less number of misconceptions compared to theoretical/descriptive questions (Fisher’s exact test, P = 0.007).

Table 5 summarizes the statistical analysis and shows the main results.

Table 5 Summary of statistical results considering the interaction between different variables.

Discussion

This study used a multiple-choice assessment to better understand the interrelationship between being confident (or not) in a chosen answer. Our study used multiple-choice questions combined with confidence scales aiming to point out misconceptions and to assure having valuable, understandable and friendly questions. With the use of confidence scales, we acquired unique information on students’ learning, mainly detecting the ones who were more likely to think critically and identifying the ones who needed additional support. Also, this assessment/study provided us knowledge about the value of multiple-choice questions, helping to improve future assessments.

Even without having a traditional multiple-choice assessment (with no confidence scale) to act as a group comparison in our study, it was undoubtedly, that adding confidence scales provided a more complete assessment. Literature has proved that confidence-based assessment techniques integrate the selection of multiple-choice answers with the student’s self-perceived level of certainty and offer a middle ground between the traditional multiple-choice answer and a lengthy essay response. Research has discovered that confidence-based assessments provide a more comprehensive measure of a person’s knowledge, increases the retainability of learned material and identifies topics in which people are misinformed.25,26,27 In addition, confidence scales also helped us to determine the value of questions, since we could verify if faculty’s understanding on the question was in line with students’ understanding (“basic” or “moderate” level of difficulty).

Having learned from previous studies, we expected that students would exhibit high confidence levels.13,21,28 In addition to foreseeing the students’ overconfidence, we also anticipated high percentage of correctness. Indeed, as expected, we found preclinical endodontic students highly correct (92.5%) and very confident in their responses (84.6%). Students who had more misconceptions had also the lowest performance in the assessment (statistically significant difference). One reason for the high accuracy was the relatively straightforward assessment that we have chosen. The nature of preclinical instruction in our academic setting is technical and focused on modus operandi, on simple recall of information and on developing hand skills; which requires, therefore, an assessment in consonance with. We are aware that straightforward and uncomplicated assessments may sometimes not be the ideal test to perform a discrimination power analysis (which was outside the scope of this study) and may not provide sufficient insights to measure students’ knowledge.29 Another reason for students being highly accurate in their answers is that this study was conducted with a small class, where students were able to receive individual and personalized attention; and the same faculty member delivered all the theoretical content. A single lecturer eliminates the challenges of calibration that might occur when several mentors are involved in the course. Since non-calibrated mentors may not apply the same standardization in protocols, techniques and philosophies,30,31 it could raise students’ misinterpretations and misconceptions. On the other hand, it is important to remember that, one disadvantage of having only one tutor, might be that it does not favor different types of learners, if the tutor uses only one modality of teaching.

The calibration between correctness and confidence is the ideal situation that fosters clever decision-making in a dental clinical setting and positively contributes to professional growth.14,15,16,21 The literature has shown the importance of increasing confidence through learning. Authors, by using a mixed-methods approach, proved how changes in confidence have positively influenced the clinical practice of two cohorts of general dental practitioners—demonstrating the following results: the ability to undertake more complex patient treatment with confidence; a realization that previously unattainable skills were attainable; a greater appreciation of one’s limitations; and a greater sense of job satisfaction, which was reflected by patient acceptance and contentment.32

Nonetheless, other evidences point in favor of a “medium” confidence level, saying that high confidence is not always appropriate. In isolation, overconfidence may be seen as undesirable professional characteristic as it is linked to a lack of accuracy and potential reckless behavior.13 In our study, we found high achieving students (the ones with scores equal or higher than 85% on the assessment) were still unsure in their incorrect responses (statistically significant). This is a situation where being unconfident is somehow worthwhile. It shows that although students did not know the answer, they were aware of their uncertainty. In other words, they were able to self-reflect on their uncertainty. Authors have stated that uncertainty can favor apprenticeship.9 If these dental students move from preclinical to the clinical setting with a ‘medium confidence’ in unfamiliar topics they will be open-minded and more prompt to seek tutoring, which may avoid clinical errors, unnecessary procedures, and harmfulness to the patient.14,21

Some correct answers were linked to under-confidence (unsure or very unsure). If this situation occurs in a traditional multiple-choice assessment (without confidence questions), it may lead to mentors to believe that students know the answer, when they were actually guessing.20,33 In our assessment, because of the confidence scale, we were able to provide feedback on both incorrect and correct student responses.

‘Moderate’ questions (as opposed to ‘basic’ questions) induced the majority of incorrect answers. These questions also prompted lower confidence levels, even when students responded correctly. Both results were statistically significant. Question 20, for instance, is a ‘moderate’ level and ‘requires a mental picture of a clinical scenario’ (see Table 4 for more information). Although it is composed of a very simple statement, students needed to mentally create the situation about access opening, connecting theoretical and clinical information. As the students-participants were in the preclinical level, they did not yet have clinical experience. Therefore, it is natural that they still have difficulties in transferring information from a preclinical to a clinical setting. To overcome this challenge of interfacing theory and technique, one idea is to use videos showing real time endodontic procedures,34 or to provide preclinical students with the opportunity of ‘shadowing’ more advanced-level peers in the clinical setting. On the other hand, the statistical analysis showed that questions requiring clinical scenario induced significant less number of misconceptions—which, therefore, made us to consider these types of questions valuable. Another type of valuable question, deserving to be used in further assessments, was the one that employed images/drawings. Question 5, in comparison to question 14, (see Table 4) facilitated the understanding of an endodontic opening access design, having significant less misconceptions. Therefore, we assumed that visual questions may benefit dentistry students, since they usually encounter a professional work environment that exploits images (photos, videos, radiographs, anatomic features, graphical abstracts, schemes, etc.).

Adversely, we also found questions that will need to be revised and improved for next assessments. And, although there was not statistical difference in misconceptions between negative word stem and the other questions, we did find a considerable number of misconceptions in negative questions. It is already known that negative questions might be potentially ineffective and misinterpreted.35,36 In here, we noticed that this type of question induced error, independent of confidence.

The strengths of this current study included the use of authentic data from an academic assessment, which helps assure students were appropriately engaged. Also, we were able to provide personal feedback for each learner—mainly the ones who got misconceptions. Limitations include having a convenience sample and having results from just one assessment. Our results might not be generalizable to other institutions; however, the findings about the questions may have external validity, since they showed inherent positive and negative characteristics of endodontic preclinical questions. Also, the inherent imprecision of the measurement of the confidence level—that is always a multifaceted and subjective measurement;37 and the multiple-choice assessment—that might foster study habits consistent with superficial or strategic learning (only to answer the questionnaire correctly). When this happen, the student would interested in recognizing isolated facts rather than relationships or higher levels of learning.38

As conclusion, we found that preclinical endodontic students were highly correct (92.5%) and very confident in their responses (84.6%). Students who had more misconceptions had also the lowest performance in the assessment. All questions were considered valuable; but some will need further improvement. This study showed that understanding the interfaces between confidence and correctness grant us with the possibility to point out students’ misconceptions (guaranteeing unique feedback for each student) and to assure we have valuable questions for current and future endodontic preclinical assessments.