Startseite Medizin Multifactorial assessment of measurement errors affecting intraoral quantitative sensory testing reliability
Artikel Öffentlich zugänglich

Multifactorial assessment of measurement errors affecting intraoral quantitative sensory testing reliability

  • Estephan J. Moana-Filho EMAIL logo , Aurelio A. Alonso , Flavia P. Kapos , Vladimir Leon-Salazar , Scott H. Durand , James S. Hodges und Donald R. Nixdorf
Veröffentlicht/Copyright: 1. Juli 2017
Veröffentlichen auch Sie bei De Gruyter Brill

Abstract

Background and purpose (aims)

Measurement error of intraoral quantitative sensory testing (QST) has been assessed using traditional methods for reliability, such as intraclass correlation coefficients (ICCs). Most studies reporting QST reliability focused on assessingone source of measurement error at a time, e.g., inter- or intra-examiner (test–retest) reliabilities and employed two examiners to test inter-examiner reliability. The present study used a complex design with multiple examiners with the aim of assessing the reliability of intraoral QST taking account of multiple sources of error simultaneously.

Methods

Four examiners of varied experience assessed 12 healthy participants in two visits separated by 48 h. Seven QST procedures to determine sensory thresholds were used: cold detection (CDT), warmth detection (WDT), cold pain (CPT), heat pain (HPT), mechanical detection (MDT), mechanical pain (MPT) and pressure pain (PPT). Mixed linear models were used to estimate variance components for reliability assessment; dependability coefficients were used to simulate alternative test scenarios.

Results

Most intraoral QST variability arose from differences between participants (8.8–30.5%), differences between visits within participant (4.6–52.8%), and error (13.3–28.3%). For QST procedures other than CDT and MDT, increasing the number of visits with a single examiner performing the procedures would lead to improved dependability (dependability coefficient ranges: single visit, four examiners = 0.12–0.54; four visits, single examiner = 0.27–0.68). A wide range of reliabilities for QST procedures, as measured by ICCs, was noted for inter- (0.39–0.80) and intra-examiner (0.10–0.62) variation.

Conclusion

Reliability of sensory testing can be better assessed by measuring multiple sources of error simultaneously instead of focusing on one source at a time. In experimental settings, large numbers of participants are needed to obtain accurate estimates of treatment effects based on QST measurements. This is different from clinical use, where variation between persons (the person main effect) is not a concern because clinical measurements are done on a single person.

Implications

Future studies assessing sensorytestingreliabilityinboth clinicaland experimental settings would benefit from routinely measuring multiple sources of error. The methods and results of this study can be used by clinical researchers to improve assessment of measurement error related to intraoral sensorytesting. This should lead to improved resource allocation when designing studies that use intraoral quantitative sensory testing in clinical and experimental settings.

© 2017 Scandinavian Association for the Study of Pain. Published by Elsevier B.V. All rights reserved.

1 Introduction

Somatosensory system assessment is part of a clinical examination of a patient presenting with pain, including tests to assess various sensory functions [1]. Such evaluation of the orofacial region includes traditional procedures such as thermal/electrical pulp tests, tooth percussion, palpation, and anaesthetic blocks [2] as well as various thermal, mechanical, and chemical stimuli [3]. In clinical settings, these tests are qualitative and lack standardization regarding stimulus application and assessment of evoked sensations [1,3]. When performed in a systematic manner using strictly defined stimulus properties, these tests are called quantitative sensorytesting (QST) [4,5]. A standard QST protocol has been developed by the German Research Network on Neuropathic pain (DFNS) [6], and further developed for intraoral use [7].

Measurement error assessment is important for sensory testing given the multiple sources of variation: variation in stimulus (delivery methods), between examiners (experience, dexterity), between participants (sensitivity, attention, previous experiences), or between multiple visits. Intraoral QST measurement error has been investigated in healthy participants [7] and patients with persistent intraoral pain [8]. These studies focused on two measures of reliability, intra- and inter-examiner, as previously assessed in other studies of QST reliability [9,10]. This approach only accounts for one source of variation at a time – examiner or visit – and thus does not identify or measure other factors, e.g., related to participants, interactions between factors, or random error. Recently studies have investigated multiple sources of variation for sensory testing [11,12]. Such a comprehensive approach can identify factors that, once addressed, can reduce variation and guide resource allocation for studies employing sensory testing and also help evaluate these tests’ applicability in clinical practice [13,14].

Our aim was to assess multiple sources of variation in a battery of intraoral QST procedures to: (i) determine their main source(s) of variation; and (ii) evaluate the influence of the number of examiners and participant visits for QST measurements’ dependability.

2 Methods

2.1 Participants

Healthy participants were recruited from the UMN community. Eligibility criteria were absence of bodily pains in the previous six months and no visible oral disease. Telephone or in-person screening was initially done, then a clinical evaluation determined participant eligibility.

2.2 Examiners

Four examiners with varied clinical experience performed the intraoral QST procedures: one faculty member, one post-doctoral fellow, one dental resident, and one dental student. The faculty examiner underwent a 2-day training session in the intraoral QST protocol at the University of Washington. He then conducted a 1- day training session for the other three examiners, after which all four examiners practiced the procedures together on two further occasions.

2.3 Study design

The intraoral QST protocol was based on the DFNS adapted for intraoral use [7], retaining seven of the 13 original procedures due to time constraints and limited available resources. It included procedures measuring thresholds for thermal (cold detection [CDT], warmth detection [WDT], cold pain [CPT], and heat pain [HPT]) and mechanical (mechanical detection [MDT], mechanical pain [MPT], pressure pain [PPT]) sensory functions. Sensory testing was performed in four intraoral sites, one over the buccal premolar gingival mucosa in each quadrant. Thermal tests were performed in 2 quadrants, which were selected randomly in each participant for each thermal test done by each examiner; mechanical tests were performed in all quadrants.

Each participant was measured on two visits separated by 48 h, with each visit lasting a half-day. Before each session, all examiners convened to review the protocol. Separate dental operatory stations were used for these procedures: (1) PPT, (2) MDT, (3) MPT, and (4) thermal. Each participant remained seated in a given station and received that station’s procedure(s) from each examiner, then moved to the next station to be examined by each examiner with that station’s assigned procedure(s), until all seven procedures were performed on each participant by all four examiners.

2.4 QST procedures

2.4.1 Thermal testing

PATHWAY Pain & Sensory Evaluation System (Medoc, Israel) with an intraoral thermode having a round active contact surface (diameter = 6 mm) was used for all thermal tests, which were performed in the sequence: CDT-WDT-CPT-HPT. For each test, the intraoral thermode was held in place by the examiner, with a baseline temperature of 32 °C, and temperature change rate of 1 °C/s for CDT and WDT; for CPT and HPT, the rate of temperature change from baseline was 1.5 °C/s; the rate of return to baseline was 8 °C/s. Cut-off temperatures for thermal tests were 0 °C and 54°C. Participants were instructed to hold a response unit and press its button once a particular sensation (coolness, warmth, cold pain, heat pain) was first perceived, ending the trial. Detection thresholds were calculated as the temperature difference from baseline; pain thresholds were determined from the absolute temperature reached. Each test included three measurements; the average of the three measurements was used as threshold.

2.4.2 Mechanical testing

2.4.2.1 Mechanical detection threshold

Thirteen modified von Frey monofilaments (OptiHair2, MARSTOCK nervtest, Germany) were used to determine the MDT. The force range was 0.125–512 mN, with each successive filament increasing force by a multiple of 2. Threshold measurement used an adaptation of the method of limits described previously [15], obtaining three infra- and three supra-thresholds. The geometric mean of these six values was used as the MDT.

2.4.2.2 Mechanical pain threshold

MPT was measured using a custom-made weighted set of eight calibrated pinprick instruments with a flat contact surface of 0.2 mm diameter (4–512 mN force range, factor 2 progression). Each instrument was applied perpendicular to the intraoral sites, with a contact time of approximately 2 s. The method of limits was used to determine six values, with their geometric mean used as the MPT.

2.4.2.3 Pressure pain threshold

PPT was measured using a digital pressure algometer (SOMEDIC, Sweden) fitted with a probe (surface area: 0.18 cm2, diameter: 4.8 mm). Participants held a switch connected to the algometer and were instructed to press it at the first painful sensation. After placing the probe tip over the gingiva, pressure was increased at a rate of 50 kPa/s until the participant interrupted the stimulus. The average of three trials was used as the PPT.

2.5 Data analysis

Analyses used mixed linear models with the following variance components

Main effects

  1. Bona fide differences between participants in threshold (“true scores”) (σp2);

  2. Differences between a participant’s quadrants (σq[p]2);

  3. Examiner differences in administering a test (σe2), e.g., knowledge, skill, experience, and biases;

  4. Differences between visits (σv2);

    Interactions:

  5. “Examiner-by-participant” (σe,p2), variation between examiners in their measured differences between participants;

  6. “Visit-by-participant” (σv,p2), variation between visits in a participant’s measurements;

  7. “Examiner-by-visit” (σe,v2), variation between visits in an examiner’s measurements;

  8. “Examiner-by-quadrant within participant” (σe,q[p]2), variation between examiners in the differences between quadrants that they measure;

  9. “Visit-by-quadrant within participant” (σv,q[p]2); variation between visits in differences between quadrants;

  10. “Examiner-by-visit-by-participant” (σe,v,p2); possible influencing factors similar to items 5, 6 and 7 taken in conjunction;

    And finally:

  11. Residual error (σResid2), which includes both the interaction “examiner-by-visit-by-quadrant within participant” and random error.

For these mixed linear models, the dependent variables were the test measurements on their raw scales except for MDT and MPT, for which the common logarithms (log to base 10) were used as dependent variables. Dependability coefficients for alternative settings (different number of examiners or visits) were also computed for each test (see Supplementary file). These coefficients simulate alternative test settings beyond those used in our study, and allow us to estimate how adding more visits would impact the variability of test results. Simple summaries for all seven measures (average, 2.5 and 97.5 percentiles) are presented for descriptive purposes, ignoring the complex study design (i.e., multiple examiners, visits, etc.).

Intraclass correlation coefficients (ICC) describing inter- and intra-examiner (test–retest) reliability were calculated for comparison to previous reliability studies. Study design complexity was taken into account by adapting Shrout and Fleiss’s “ICC (2,1)” model [16] (see Supplemental material). Inter- and intra-examiner levels of agreement were deemed poor (ICC <0.40), fair (0.40≤ ICC≤0.59), good (0.60≤ICC≥0.74), or excellent (ICC≥ 0.75) using published guidelines [17].

All analyses were implemented using SAS (v. 9.3, SAS Institute Inc., Cary, NC) and the R system v.3.3.0 [18].

3 Results

Twelve participants (10 females, average age±SD: 39.3±11.9 years) participated in the first visit; 8 returned for the second visit (6 females, age: 39.8±10.7), as the other participants had time constraints preventing their return for a second visit.

Table 1 lists simple summaries of the QST measures; Table 2 shows intra- and inter-examiner ICCs. Inter-examiner reliability was poor for CDT and MDT, fair for PPT, good for WDT, HPT, and MPT, and excellent for CPT. Intra-examiner (test–retest) ICCs had lower reliability, mostly poor (WDT, CPT, MPT, PPT) and fair (HPT, MDT).

Table 1

Means for QST procedures.

Procedure Mean Percentiles 2.5–97.5
CDT[a] (°C) 9.7 3.1–23.1
WDT[a] (°C) 9.1 3.3–13.9
CPT[b] (°C) 11.5 0.0–23.9
HPT[b] (°C) 44.7 39.0–50.0
MDT[c] (mN) 5.5 0.2–32.0
MPT[c] (mN) 148.6 12.0–406.0
PPT(kPa) 219.0 78.0–454.0

Table 2

Intraclass correlation coefficients (ICC) for QST procedures.

Procedure Inter-examiner Intra-examiner (test–retest)
CDT 0.39 0.62
WDT 0.66 0.10
CPT 0.80 0.34
HPT 0.63 0.42
MDT 0.39 0.52
MPT 0.60 0.27
PPT 0.53 0.37
  1. Study design complexity accounted for by adapting the “ICC 2,1”, model proposed by shrout and fleiss (1979) – see supplemental material for details.

    Levels of agreement: poor (ICC < 0.40), fair (0.40≤ICC≤0.59), good (0.60≤ICC−0.74), and excellent (ICC≥0.75).

Table 3 presents variance component estimates for each procedure both as absolute values and as percentages of total variance. For each procedure, most variance (>75%) arises from the participant main effect (σp2), differences between visits within participant (σv,p2), and residual error (σResid2). Also, CDT had large variation between examiners in the difference between quadrants (σe,q[p]2); MDT had large variation between quadrants and between examiners (σq[p]2,σe2); and PPT had a large examiner-by-visit-by-participant interaction (σe,v,p2). The participant main effect (variation between their “true scores”) accounted for 8.8–33.7% of total variance, depending on the test. As examples, the largest components of variance for CPT were the participant main effect (30.5%) and differences between visits within participant (45%), while for MDT the largest components were the residual (24.7%) and the main effects for examiners (20.6%), participants (18.4%), and quadrants (11.5%).

Table 3

Variance component estimates for each QST procedures.

Variance component CDT WDT CPT HPT
σ p 2 4.42 (20.6) 0.70(8.8) 19.80 (30.5) 2.55 (33.7)
σ q[p] 2 0.42 (2.0) 0.00 (0.0) 2.49 (3.8) 0.01 (0.1)
σ e 2 0.19 (0.9) 0.00 (0.0) 0.00 (0.0) 0.00 (0.0)
σ v 2 0.07 (0.3) 0.00 (0.0) 0.56 (0.8) 0.00 (0.0)
σ e,p 2 2.42(11.3) 0.13 (1.6) 0.00 (0.0) 0.26 (3.4)
σ v,p 2 3.39(15.8) 4.19 (52.8) 29.21 (45.0) 1.86(24.5)
σ e,v 2 0.00 (0.0) 0.00 (0.0) 0.00 (0.0) 0.08(1.1)
σ e,q[p] 2 5.92 (27.7) 0.00 (0.0) 0.00 (0.0) 0.33 (4.3)
σ v,q[p] 2 0.00 (0.0) 0.37 (4.7) 0.00 (0.0) 0.35 (4.6)
σ e,v,p 2 0.00 (0.0) 0.72(9.1) 0.54 (0.8) 0.00 (0.0)
σ Resid 2 4.58(21.4) 1.83 (23.0) 12.36 (19.0) 2.15(28.3)
σ Total 2 21.42(100.0) 7.93(100.0) 64.94(100.0) 7.58(100.0)
Variance component MDT[a] MPT[a] PPT

σ p 2 0.07(18.4) 0.03 (13.7) 2418.11 (30.0)
σ q[p] 2 0.04 (11.5) 0.01 (4.0) 259.32 (3.2)
σ e 2 0.08 (20.6) 0.00(1.4) 421.13 (5.2)
σ v 2 0.01 (2.4) 0.01 (2.8) 88.74 (1.1)
σ e,p 2 0.02 (5.8) 0.02 (8.7) 0.00 (0.0)
σ v,p 2 0.02 (4.6) 0.08 (41.0) 877.09(10.9)
σ e,v 2 0.01 (1.4) 0.00 (0.0) 0.00 (0.0)
σ e,q[p] 2 0.02 (6.0) 0.00 (0.0) 126.06(1.6)
σ v,q[p] 2 0.01 (4.0) 0.00 (0.0) 670.65 (8.3)
σ e,v,p 2 0.00 (0.7) 0.01 (4.5) 2117.68(26.3)
σ Resid 2 0.09 (24.7) 0.05 (23.8) 1074.26(13.3)
σ Total 2 0.37 (100.0) 0.20(100.0) 8053.04(100.0)
  1. Variance(%).

Table 4 shows dependability coefficients for alternative test scenarios. As expected, the worst (1 visit, 1 examiner) and best (4 visits, 4 examiners) scenarios gave a wide range of dependability coefficients, for example 0.23–0.60 for CDT. Except for CDT and MDT, having more participant visits with a single examiner improved the procedure’s dependability coefficient more than having a single visit with 4 examiners. Other test settings can be simulated using the “variance components estimates calculator” available as a supplementary file.

Table 4

Depenability coefficients for QST procedures.

N visits N examiners

CDT WDT


1 2 3 4 1 2 3 4
1 0.23 0.33 0.38 0.42 0.09 0.11 0.11 0.12
2 0.28 0.40 0.48 0.52 0.16 0.19 0.20 0.21
3 0.30 0.44 0.52 0.57 0.22 0.26 0.27 0.28
4 0.31 0.46 0.54 0.60 0.27 0.31 0.33 0.34
N visits N examiners

CPT HPT


1 2 3 4 1 2 3 4

1 0.34 0.38 0.40 0.40 0.34 0.41 0.45 0.47
2 0.51 0.55 0.57 0.57 0.48 0.57 0.61 0.63
3 0.61 0.65 0.66 0.67 0.55 0.65 0.68 0.71
4 0.68 0.71 0.72 0.73 0.60 0.70 0.73 0.75
N visits N examiners

MDT[a] MPT[a]


1 2 3 4 1 2 3 4

1 0.30 0.42 0.49 0.54 0.18 0.22 0.24 0.25
2 0.37 0.51 0.59 0.64 0.28 0.34 0.37 0.39
3 0.40 0.55 0.63 0.68 0.34 0.42 0.45 0.47
4 0.42 0.57 0.65 0.71 0.38 0.47 0.51 0.54
N visits N examiners
PPT

1 2 3 4

1 0.33 0.43 0.48 0.51
2 0.47 0.59 0.64 0.66
3 0.55 0.66 0.71 0.74
4 0.60 0.71 0.76 0.78
  1. Dependability coefficients range: 0 (poor) – 1 (high).

4 Discussion

The present study used a complex design with four examiners having different training levels, examining each participant in two visits, allowing estimation of several variance components for each of seven QST procedures. This allowed detailed examination of multiple error sources that contribute to the procedures’ performance. This approach departs from the usual practice of focusing on one source of error such as variability between examiners (inter-examiner) or between visits by the same examiner (intra-examiner), which others have previously noted [11].

Average intraoral QST measures reported here (Table 1) are slightly lower but comparable to previous studies [7,19], especially considering the 2.5th and 97.5th percentiles. We calculated ICCs to compare our results to those studies, though as noted our calculations needed elaboration to accommodate our study design and use the whole dataset. To the best of our knowledge, this is the first time ICCs have incorporated such design complexity; by doing so we avoided reporting 56 ICCs (4 examiners × 2 visits × 7 procedures) based on subsets of the data, and also obtained more stable estimates.

Despite the aforementioned complexity, all reliability calculations using ICCs follow the same basic formula [20]:

reliability=between subjects variability (signal)between subjects variability (signal)+error (noise)

From this, it follows that an ICC ranges between 0 (no reliability) to 1 (perfect reliability), and its value depends on the magnitude of “signal” relative to “signal + noise”. The mathematical model rationalizing these ICCs in the present study included in “signal” the variance components for the main effects for “participant” and “quadrant”; in addition to these, the inter-examiner ICC included in “signal” the interactions “visit-by-participant” and “visit-by-quadrant within participant”, while the intraexaminer (test–retest) ICC included “examiner-by-participant” and “examiner-by-quadrant” (see supplemental material). This aids understanding of Table 2’s ICC values. Consider CPT for example, with inter- and intra-examiner ICCs of 0.80 and 0.34 respectively. In Table 3’s variance component estimates for CPT, “visit-byparticipant” contributes 45% of the total variance, which means that almost half of the variability for CPT arises from variation between participants in the change in their thresholds between the first and second visits. Because this component is considered “signal” for inter-examiner ICC but “noise” for intra-examiner ICC, it becomes clear why the former ICC is much larger than the latter.

Another feature of ICCs is that low values can arise either from large measurement error (high noise) from any of several sources (methods/instruments, examiners, visits) or from reduced between-subject variability (low signal), i.e., subjects who “look alike” when measured. This does not mean that a measurement method with low ICC has no use, but only that is has limited ability to discriminate the specific collection of subjects that was assessed. Such a method could be valuable for assessing a given patient over time if its measurement error is small; in this case, between-subject variability has no influence on the method’s performance [21]. The above formula also implies that ICC as a measure of reliability reflects a measurement method’s performance for a given population sample with a given heterogeneity (context specific); when comparing a method’s ICCs as measured in studies that used different samples, one needs to assess whether the samples are similarly heterogeneous, for example by evaluating the between- and within-subject standard deviations [22].

Reliability calculations focusing on one error source at a time, such as ICCs, are considered part of a framework known as “classic test theory”. This has been recognized as an important limitation; an extension of this framework, called generalizability theory (GT) [23,24], was proposed so that multiple sources of measurement error could be recognized and estimated. The methods used here can be considered part of this trend, providing several benefits. First, variance components estimation allows not only calculation of reliability measures such as ICCs (how well can participants be distinguished from each other, despite measurement error), but also other aspects of measurement error such as “agreement” (how close are two measurements on the same participant). Agreement is expressed in the same units as the measurement itself and unlike ICC does not depend on the sample’s heterogeneity, thus quantifying measurement error only [20,21,22]. Second, no consensus exists on how reproducibility of QST results should be defined or assessed [9,25,26]. Finally, in a carefully designed study (e.g., number of participants, examiners, and visits), variance components estimation is relatively straightforward and would also allow calculation of dependability coefficients, which can be used to simulate alternative test scenarios to improve allocation of resources in future studies using QST procedures for sensory testing to obtain accurate estimates of treatment effects.

Table 3 shows that most variability in QST procedures arises from three variance components – differences between participants (“true scores”), differences between visits within participant, and error – but there were exceptions. For CDT, 27.7% of variation was variation between examiners in their measured differences between intraoral quadrants. One explanation is that CDT was the first thermal procedure done in all subjects, and placement of the intraoral thermal probe could elicit discomfort, thus impacting overall CDT variability. About a third of MDT’s variation was variation between participants in differences between quadrants and differences between examiners, which could be related, respectively, to difficulties in positioning the Von Frey filaments and in examiners’ experience in using the filaments. Other authors have found poor reliability for MDT, attributing it to the filaments’ design and the method of limits used for threshold determination [7]. The interaction “examiner-by-visit-by-participant” represented 26.3% of PPT total variance, meaning that within-participant differences between visits also varied considerably between participants and examiners. This could be a consequence of the examiners’ varying ability in assessing PPT across visits. Based on these findings and after addressing concerns specific to CDT, MDT, and PPT, one potential way to improve these procedures’ reliability is to perform them on multiple visits. By averaging measurements from all visits, the error attributed to within-participant variability between visits can be reduced [11], allowing true differences between participants to explain a greater portion of the remaining variability of each procedure, i.e., the measurement would become more reliable.

Studies using the GT framework would provide information needed to serve the different needs of clinical decision making and clinical research: when examining a patient, the between-patient component of variation does not contribute to error in determining that patient’s status (diseased vs. non-diseased), but when comparing treatments given to distinct groups of study participants, that component of variation will largely determine the statistical power for a given sample size, as the present study suggests. By knowing the several sources of error in each test’s performance, improved allocation of limited resources (clinician’s chair time, costs associated with procedures, study participant sample size) will ultimately lead to improved diagnosis in clinical settings and more accurate estimates of treatment effects in experimental settings. Dependability coefficients calculation for alternative test scenarios simulation could help researchers to determine which design features would improve the tests’ reliability.

This study has strengths and limitations. Our sample size was relatively small; the mathematical models allowed calculations with missing data but a larger sample would give more stable estimates. A shortened version of the intraoral QST protocol was used [7], with seven out of the 13 procedures originally described and MDT and MPT modified to reduce threshold measurements from 10 to six. These modifications were needed to save time four examiners were included instead of two, as in most studies [7,8,9,10,11]. An additional strength was accounting for multiple sources of error at the same time including testing in different intraoral quadrants, which allows better understanding of factors affecting intraoral QST reliability.

5 Conclusions

To the best of our knowledge, this study was the first to account for multiple sources of measurement error in intraoral QST simultaneously. This allows an improved understanding of the moderate to poor reliability for these tests measured by ICC for intra- and inter-rater scenarios. Using the GT framework to determine the reliability of intraoral sensory tests can help elucidate sources of error to inform clinical decision-making and resources allocation in experimental settings.

Supplemental material

  1. Rationale for ICC and dependability coefficients calculation and Dependability Coefficients.

  2. Variance components estimates and Dependability Coefficients calculator.

Highlights

  • A comprehensive approach to assess multiple sources of intraoral QST variation is proposed.

  • Most variability comefrom differences between participants and visits-within-participant.

  • Comprehensive reliability appraisal aids in clinical decision-making and resources allocation.


Disclosures

Research reported in this publication was supported by the National Institutes of Health grants UL1-TR000114 and K12-RR23247. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.

DOI of refers to article: http://dx.doi.org/10.1016/j.sjpain.2017.04.066.


  1. Ethical issues: The University of Minnesota (UMN) Institutional Review Board approved the study’s protocol (approval #1004M80212) and all participants gave oral and written informed consent before entering the study. Data were collected in May 2010.

  2. Conflict of interest: The authors declare that there is no conflict of interest.

References

[1] Hansson P, Backonja M, Bouhassira D. Usefulness and limitations of quantitative sensory testing: clinical and research application in neuropathic pain states. Pain 2007;129:256–9.Suche in Google Scholar

[2] Hargreaves KM, Berman LH. Cohen’s Pathways of the Pulp Expert Consult. 11th ed. Mosby; 2015.Suche in Google Scholar

[3] Svensson P, Baad-Hansen L, Pigg M, List T, Eliav E, Ettlin D, Michelotti A, Tsukiyama Y, Matsuka Y, Jaaskelainen SK, Essick G, Greenspan JD, Drangsholt M. Guidelines and recommendations for assessment of somatosensory function in oro-facial pain conditions – a taskforce report. J Oral Rehabil 2011;38:366–94.Suche in Google Scholar

[4] Cruz-Almeida Y, Fillingim RB. Can quantitative sensory testing move us closer to mechanism-based pain management? Pain Med 2014;15:61–72.Suche in Google Scholar

[5] Gruener G, Dyck PJ. Quantitative sensory testing: methodology, applications, and future directions. J Clin Neurophysiol 1994;11:568–83.Suche in Google Scholar

[6] Rolke R, Baron R, Maier C, Tolle TR, Treede RD, Beyer A, Binder A, Birbaumer N, Birklein F, Botefur IC, Braune S, Flor H, Huge V, Klug R, Landwehrmeyer GB, Magerl W, Maihofner C, Rolko C, Schaub C, Scherens A, Sprenger T, Valet M, Wasserka B. Quantitative sensory testing in the German Research Network on Neuropathic Pain (DFNS): standardized protocol and reference values. Pain 2006;123:231–43.Suche in Google Scholar

[7] Pigg M, Baad-Hansen L, Svensson P, Drangsholt M, List T. Reliability of intraoral quantitative sensory testing (QST). Pain 2010;148:220–6.Suche in Google Scholar

[8] Baad-Hansen L, Pigg M, Yang G, List T, Svensson P, Drangsholt M. Reliability of intra-oral quantitative sensory testing (QST) in patients with atypical odontalgia and healthy controls – a multicentre study. J Oral Rehabil 2015;42: 127–35.Suche in Google Scholar

[9] Geber C, Klein T, Azad S, Birklein F, Gierthmuhlen J, Huge V, Lauchart M, Nitzsche D, Stengel M, Valet M, Baron R, Maier C, Tolle T, Treede RD. Test–retest and interobserver reliability of quantitative sensory testing according to the protocol of the German Research Network on Neuropathic Pain (DFNS): a multi-centre study. Pain 2011;152:548–56.Suche in Google Scholar

[10] Moloney NA, Hall TM, O’Sullivan TC, Doody CM. Reliability of thermal quantitative sensory testing of the hand in a cohort of young, healthy adults. Muscle Nerve 2011;44:547–52.Suche in Google Scholar

[11] O’Neill S, O’Neill L. Improving QST reliability – more raters, tests, or occasions? A multivariate generalizability study. J Pain 2015;16:454–62.Suche in Google Scholar

[12] Pryseley A, Ledent EY, Drewes AM, Staahl C, Olesen AE, Arendt-Nielsen L. Applying concepts of generalizability theory on data from experimental pain studies to investigate reliability. Basic Clin Pharmacol Toxicol 2009; 105:105–12.Suche in Google Scholar

[13] Backonja MM, Attal N, Baron R, Bouhassira D, Drangholt M, Dyck PJ, Edwards RR, Freeman R, Gracely R, Haanpaa MH, Hansson P, Hatem SM, Krumova EK, Jensen TS, Maier C, Mick G, Rice AS, Rolke R, Treede RD, Serra J, Toelle T, Tugnoli V, Walk D, Walalce MS, Ware M, Yarnitsky D, Ziegler D. Value of quantitative sensory testing in neurological and pain disorders: NeuPSIG consensus. Pain 2013;154:1807–19.Suche in Google Scholar

[14] Birklein F, Sommer C. Pain: quantitative sensory testing – a tool for daily practice? Nat Rev Neurol 2013;9:490–2.Suche in Google Scholar

[15] Baumgartner U, Magerl W, Klein T, Hopf HC, Treede RD. Neurogenic hyperalgesia versus painful hypoalgesia: two distinct mechanisms of neuropathic pain. Pain 2002;96:141–51.Suche in Google Scholar

[16] Shrout PE, Fleiss JL. Intraclass correlations: uses in assessing rater reliability. Psychol Bull 1979;86:420–8.Suche in Google Scholar

[17] Cicchetti DV. Guidelines, criteria, and rules of thumb for evaluating normed and standardized assessment instruments in psychology. Psychol Assess 1994;6:284.Suche in Google Scholar

[18] R Core Team. R: a language and environment for statistical computing. Vienna, Austria: R Foundation for Statistical Computing; 2016.Suche in Google Scholar

[19] Baad-Hansen L, Pigg M, Ivanovic SE, Faris H, List T, Drangsholt M, Svensson P. Intraoral somatosensory abnormalities in patients with atypical odontalgia – a controlled multicenter quantitative sensory testing study. Pain 2013;154:1287–94.Suche in Google Scholar

[20] Weir JP. Quantifying test–retest reliability using the intraclass correlation coefficient and the SEM. J Strength Cond Res 2005;19:231–40.Suche in Google Scholar

[21] de Vet HC, Terwee CB, Knol DL, Bouter LM. When to use agreement versus reliability measures. J Clin Epidemiol 2006;59:1033–9.Suche in Google Scholar

[22] Bartlett JW, Frost C. Reliability, repeatability and reproducibility: analysis of measurement errors in continuous variables. Ultrasound Obstet Gynecol 2008;31:466–75.Suche in Google Scholar

[23] Cronbach LJ, Rajaratnam N, Gleser GC. Theory of generalizability: a liberalization of reliability theory. Br J Stat Psychol 1963;16:137–63.Suche in Google Scholar

[24] Shavelson RJ, Webb NM, Rowley GL. Generalizability theory. Am Psychol 1989;44:922–32.Suche in Google Scholar

[25] Shy ME, Frohman EM, So YT, Arezzo JC, Cornblath DR, Giuliani MJ, Kincaid JC, Ochoa JL, Parry GJ, Weimer LH, Therapeutics, Technology Assessment Subcommittee of the American Academy of N. Quantitative sensory testing: report of the Therapeutics and Technology Assessment Subcommittee of the American Academy of Neurology. Neurology 2003;60:898–904.Suche in Google Scholar

[26] Werner MU, Petersen MA, Bischoff JM. Test–retest studies in quantitative sensory testing: a critical review. Acta Anaesthesiol Scand 2013;57:957–63.Suche in Google Scholar

Appendix A. Supplementary data

Supplementary data associated with this article can be found, in the online version, at http://dx.doi.org/10.1016/j.sjpain.2017.03.007.

Received: 2016-11-03
Revised: 2017-03-11
Accepted: 2017-03-29
Published Online: 2017-07-01
Published in Print: 2017-07-01

© 2017 Scandinavian Association for the Study of Pain

Artikel in diesem Heft

  1. Scandinavian Journal of Pain
  2. Editorial comment
  3. Glucocorticoids – Efficient analgesics against postherpetic neuralgia?
  4. Original experimental
  5. Effect of intrathecal glucocorticoids on the central glucocorticoid receptor in a rat nerve ligation model
  6. Editorial comment
  7. Important new insight in pain and pain treatment induced changes in functional connectivity between the Pain Matrix and the Salience, Central Executive, and Sensorimotor networks
  8. Original experimental
  9. Salience, central executive, and sensorimotor network functional connectivity alterations in failed back surgery syndrome
  10. Editorial comment
  11. Education and support strategies improve assessment and management of pain by nurses
  12. Clinical pain research
  13. Using education and support strategies to improve the way nurses assess regular and transient pain – A quality improvement study of three hospitals
  14. Editorial comment
  15. The interference of pain with task performance: Increasing ecological validity in research
  16. Original experimental
  17. The disruptive effects of pain on multitasking in a virtual errands task
  18. Editorial comment
  19. Analyzing transition from acute back pain to chronic pain with linear mixed models reveals a continuous chronification of acute back pain
  20. Observational study
  21. From acute to chronic back pain: Using linear mixed models to explore changes in pain intensity, disability, and depression
  22. Editorial comment
  23. NSAIDs relieve osteoarthritis (OA) pain, but cardiovascular safety in question even for diclofenac, ibuprofen, naproxen, and celecoxib: what are the alternatives?
  24. Clinical pain research
  25. Efficacy and safety of diclofenac in osteoarthritis: Results of a network meta-analysis of unpublished legacy studies
  26. Editorial comment
  27. Editorial comment on Nina Kreddig’s and Monika Hasenbring’s study on pain anxiety and fear of (re) injury in patients with chronic back pain: Sex as a moderator
  28. Clinical pain research
  29. Pain anxiety and fear of (re) injury in patients with chronic back pain: Sex as a moderator
  30. Editorial comment
  31. Intraoral QST – Mission impossible or not?
  32. Clinical pain research
  33. Multifactorial assessment of measurement errors affecting intraoral quantitative sensory testing reliability
  34. Editorial comment
  35. Objective measurement of subjective pain-experience: Real nociceptive stimuli versus pain expectation
  36. Clinical pain research
  37. Cerebral oxygenation for pain monitoring in adults is ineffective: A sequence-randomized, sham controlled study in volunteers
  38. Editorial comment
  39. Association between adolescent and parental use of analgesics
  40. Observational study
  41. The association between adolescent and parental use of non-prescription analgesics for headache and other somatic pain – A cross-sectional study
  42. Editorial comment
  43. Cancer-pain intractable to high-doses systemic opioids can be relieved by intraspinal local anaesthetic plus an opioid and an alfa2-adrenoceptor agonist
  44. Clinical pain research
  45. Spinal analgesia for severe cancer pain: A retrospective analysis of 60 patients
  46. Editorial comment
  47. Specific symptoms and signs of unstable back segments and curative surgery?
  48. Clinical pain research
  49. Symptoms and signs possibly indicating segmental, discogenic pain. A fusion study with 18 years of follow-up
  50. Editorial comment
  51. Local anaesthesia methods for analgesia after total hip replacement: Problems of anatomy, methodology and interpretation?
  52. Clinical pain research
  53. Local infiltration analgesia or femoral nerve block for postoperative pain management in patients undergoing total hip arthroplasty. A randomized, double-blind study
  54. Editorial
  55. Scientific presentations at the 2017 annual meeting of the Scandinavian Association for the Study of Pain (SASP)
  56. Abstracts
  57. Correlation between quality of pain and depression: A post-operative assessment of pain after caesarian section among women in Ghana
  58. Abstracts
  59. Dynamic and static mechanical pain sensitivity is associated in women with migraine
  60. Abstracts
  61. The number of active trigger points is associated with sensory and emotional aspects of health-related quality of life in tension type headache
  62. Abstracts
  63. Chronic neuropathic pain following oxaliplatin and docetaxel: A 5-year follow-up questionnaire study
  64. Abstracts
  65. Expression of α1 adrenergic receptor subtypes by afferent fibers that innervate rat masseter muscle
  66. Abstracts
  67. Buprenorphine alleviation of pain does not compromise the rat monoarthritic pain model
  68. Abstracts
  69. Association between pain, disability, widespread pressure pain hypersensitivity and trigger points in subjects with neck pain
  70. Abstracts
  71. Association between widespread pressure pain hypersensitivity, health history, and trigger points in subjects with neck pain
  72. Abstracts
  73. Neuromas in patients with peripheral nerve injury and amputation - An ongoing study
  74. Abstracts
  75. The link between chronic musculoskeletal pain and sperm quality in overweight orthopedic patients
  76. Abstracts
  77. Several days of muscle hyperalgesia facilitates cortical somatosensory excitability
  78. Abstracts
  79. Social stress, epigenetic changes and pain
  80. Abstracts
  81. Characterization of released exosomes from satellite glial cells under normal and inflammatory conditions
  82. Abstracts
  83. Cell-based platform for studying trigeminal satellite glial cells under normal and inflammatory conditions
  84. Abstracts
  85. Tramadol in postoperative pain – 1 mg/ml IV gave no pain reduction but more side effects in third molar surgery
  86. Abstracts
  87. Tempo-spatial discrimination to non-noxious stimuli is better than for noxious stimuli
  88. Abstracts
  89. The encoding of the thermal grill illusion in the human spinal cord
  90. Abstracts
  91. Effect of cocoa on endorphin levels and craniofacial muscle sensitivity in healthy individuals
  92. Abstracts
  93. The impact of naloxegol treatment on gastrointestinal transit and colonic volume
  94. Abstracts
  95. Preoperative downregulation of long-noncoding RNA Meg3 in serum of patients with chronic postoperative pain after total knee replacement
  96. Abstracts
  97. Painful diabetic polyneuropathy and quality of life in Danish type 2 diabetic patients
  98. Abstracts
  99. What about me?”: A qualitative explorative study on perspectives of spouses living with complex chronic pain patients
  100. Abstracts
  101. Increased postural stiffness in patients with knee osteoarthritis who are highly sensitized
  102. Abstracts
  103. Efficacy of dry needling on latent myofascial trigger points in male subjects with neck/shoulders musculoskeletal pain. A case series
  104. Abstracts
  105. Identification of pre-operative of risk factors associated with persistent post-operative pain by self-reporting tools in lower limb amputee patients – A feasibility study
  106. Abstracts
  107. Renal function estimations and dose recommendations for Gabapentin, Ibuprofen and Morphine in acute hip fracture patients
  108. Abstracts
  109. Evaluating the ability of non-rectangular electrical pulse forms to preferentially activate nociceptive fibers by comparing perception thresholds
  110. Abstracts
  111. Detection of systemic inflammation in severely impaired chronic pain patients, and effects of a CBT-ACT-based multi-modal pain rehabilitation program
  112. Abstracts
  113. Fixed or adapted conditioning intensity for repeated conditioned pain modulation
  114. Abstracts
  115. Combined treatment (Norspan, Gabapentin and Oxynorm) was found superior in pain management after total knee arthroplasty
  116. Abstracts
  117. Effects of conditioned pain modulation on the withdrawal pattern to nociceptive stimulation in humans – Preliminary results
  118. Abstracts
  119. Application of miR-223 onto the dorsal nerve roots in rats induces hypoexcitability in the pain pathways
  120. Abstracts
  121. Acute muscle pain alters corticomotor output of the affected muscle stronger than a synergistic, ipsilateral muscle
  122. Abstracts
  123. The subjective sensation induced by various thermal pulse stimulation in healthy volunteers
  124. Abstracts
  125. Assessing Offset Analgesia through electrical stimulations in healthy volunteers
  126. Abstracts
  127. Metastatic lung cancer in patient with non-malignant neck pain: A case report
  128. Abstracts
  129. The size of pain referral patterns from a tonic painful mechanical stimulus is increased in women
  130. Abstracts
  131. Oxycodone and macrogol 3350 treatment reduces anal sphincter relaxation compared to combined oxycodone and naloxone tablets
  132. Abstracts
  133. The effect of UVB-induced skin inflammation on histaminergic and non-histaminergic evoked itch and pain
  134. Abstracts
  135. Topical allyl-isothiocyanate (mustard oil) as a TRPA1-dependent human surrogate model of pain, hyperalgesia, and neurogenic inflammation – A dose response study
  136. Abstracts
  137. Dissatisfaction and persistent post-operative pain following total knee replacement – A 5 year follow-up of all patients from a whole region
  138. Abstracts
  139. Paradoxical differences in pain ratings of the same stimulus intensity
  140. Abstracts
  141. Pain assessment and post-operative pain management in orthopedic patients
  142. Abstracts
  143. Combined electric and pressure cuff pain stimuli for assessing conditioning pain modulation (CPM)
  144. Abstracts
  145. The effect of facilitated temporal summation of pain, widespread pressure hyperalgesia and pain intensity in patients with knee osteoarthritis on the responds to Non-Steroidal Anti-Inflammatory Drugs – A preliminary analysis
  146. Abstracts
  147. How to obtain the biopsychosocial record in multidisciplinary pain clinic? An action research study
  148. Abstracts
  149. Experimental neck muscle pain increase pressure pain threshold over cervical facet joints
  150. Abstracts
  151. Are we using Placebo effects in specialized Palliative Care?
  152. Abstracts
  153. Prevalence and pattern of helmet-induced headache among Danish military personnel
  154. Abstracts
  155. Aquaporin 4 expression on trigeminal satellite glial cells under normal and inflammatory conditions
  156. Abstracts
  157. Preoperative synovitis in knee osteoarthritis is predictive for pain 1 year after total knee arthroplasty
  158. Abstracts
  159. Biomarkers alterations in trapezius muscle after an acute tissue trauma: A human microdialysis study
  160. Abstracts
  161. PainData: A clinical pain registry in Denmark
  162. Abstracts
  163. A novel method for investigating the importance of visual feedback on somatosensation and bodily-self perception
  164. Abstracts
  165. Drugs that can cause respiratory depression with concomitant use of opioids
  166. Abstracts
  167. The potential use of a serious game to help patients learn about post-operative pain management – An evaluation study
  168. Abstracts
  169. Modelling activity-dependent changes of velocity in C-fibers
  170. Abstracts
  171. Choice of rat strain in pre-clinical pain-research – Does it make a difference for translation from animal model to human condition?
  172. Abstracts
  173. Omics as a potential tool to identify biomarkers and to clarify the mechanism of chronic pain development
  174. Abstracts
  175. Evaluation of the benefits from the introduction meeting for patients with chronic non-malignant pain and their relatives in interdisciplinary pain center
  176. Observational study
  177. The changing face of acute pain services
  178. Observational study
  179. Chronic pain in multiple sclerosis: A10-year longitudinal study
  180. Clinical pain research
  181. Functional disability and depression symptoms in a paediatric persistent pain sample
  182. Observational study
  183. Pain provocation following sagittal plane repeated movements in people with chronic low back pain: Associations with pain sensitivity and psychological profiles
  184. Observational study
  185. A longitudinal exploration of pain tolerance and participation in contact sports
  186. Original experimental
  187. Taking a break in response to pain. An experimental investigation of the effects of interruptions by pain on subsequent activity resumption
  188. Clinical pain research
  189. Sex moderates the effects of positive and negative affect on clinical pain in patients with knee osteoarthritis
  190. Original experimental
  191. The effects of a brief educational intervention on medical students’ knowledge, attitudes and beliefs towards low back pain
  192. Observational study
  193. The association between pain characteristics, pain catastrophizing and health care use – Baseline results from the SWEPAIN cohort
  194. Topical review
  195. Couples coping with chronic pain: How do intercouple interactions relate to pain coping?
  196. Narrative review
  197. The wit and wisdom of Wilbert (Bill) Fordyce (1923 - 2009)
  198. Letter to the Editor
  199. Unjustified extrapolation
  200. Letter to the Editor
  201. Response to: “Letter to the Editor entitled: Unjustified extrapolation” [by authors: Supp G., Rosedale R., Werneke M.]
Heruntergeladen am 5.2.2026 von https://www.degruyterbrill.com/document/doi/10.1016/j.sjpain.2017.03.007/html?lang=de
Button zum nach oben scrollen