10  Individual Pathologist Performance

10.1 Objective

Analyze each pathologist’s consistency, AI adoption patterns, and performance characteristics.

Note for Pathologist: This section gets personal (anonymously). We look at how each pathologist interacted with the AI. Did they change their mind often? Did they only change for difficult cases? This helps us understand if the AI acts as a “second opinion” that everyone trusts, or if some users are more skeptical than others.

10.2 Setup

10.3 Load Data

10.4 Intra-Pathologist Consistency

Measure how consistent each pathologist is with themselves (within-pathologist variability).

Intra-Pathologist Consistency
Correlation between Pre-AI and Post-AI assessments
Pathologist ER Correlation PR Correlation Ki67 Correlation
Pathologist 1 0.981 0.970 0.899
Pathologist 2 0.987 0.977 0.961
Pathologist 3 0.976 0.975 0.945
Pathologist 4 0.997 0.993 0.955
Mean Absolute Changes
Average magnitude of change after AI
Pathologist ER MAE PR MAE Ki67 MAE
Pathologist 1 4.38 5.56 9.56
Pathologist 2 3.52 4.47 4.97
Pathologist 3 3.53 4.08 6.36
Pathologist 4 0.91 1.80 6.75

Note for Pathologist: The tables and bar chart above show self-consistency metrics. A high Pre-Post correlation (close to 1.0) means the pathologist rarely changed their scores after seeing the AI result. A high MAE (Mean Absolute Error) means they changed by large amounts. The percentage of cases changed shows how often they adjusted any value at all.

10.5 AI Adoption Index

Create a composite metric of how much each pathologist is influenced by AI.

AI Adoption Metrics
How much each pathologist is influenced by AI
Pathologist AI Adoption Score Avg ER Change Avg PR Change Avg Ki67 Change Prop ER Changed Prop PR Changed Prop HER2 Changed
Pathologist 1 40.89 4.38 5.56 9.56 0.73 0.59 0.08
Pathologist 2 34.81 3.52 4.47 4.97 0.64 0.52 0.16
Pathologist 3 28.69 3.53 4.08 6.36 0.31 0.38 0.08
Pathologist 4 19.98 0.91 1.80 6.75 0.11 0.21 0.04

10.6 Extreme Value Analysis

Identify pathologists who tend to give extreme or moderate values.

Tendency for Extreme Values (0% or 100%)
Proportion of cases with extreme scores
Pathologist ER Pre PR Pre ER Post PR Post
Pathologist 1 25.1% 39.8% 15.8% 37.1%
Pathologist 2 52.7% 43.3% 26.5% 38.9%
Pathologist 3 17.1% 44.4% 16.3% 38.5%
Pathologist 4 17.9% 41.0% 17.9% 39.8%

10.7 Agreement with Group Median

Compare each pathologist to the group median.

Mean Absolute Deviation from Group Median
Lower values = better agreement with group
Pathologist ER Pre PR Pre Ki67 Pre ER Post PR Post Ki67 Post
Pathologist 1 2.23 2.95 1.91 1.60 1.19 3.05
Pathologist 2 3.15 4.09 3.79 1.90 1.83 2.86
Pathologist 3 5.37 4.54 2.17 3.82 2.29 3.23
Pathologist 4 2.06 2.90 1.65 1.82 2.34 2.22

Note for Pathologist: The bias plot shows whether each pathologist tends to score higher or lower than the group median. Bars above zero mean the pathologist consistently scores higher than peers for that marker; bars below zero mean they score lower. If AI reduces the bar height, it is pulling the pathologist toward the group consensus.

10.8 Marker-Specific Expertise

Identify if pathologists show different patterns for different markers.

Coefficient of Variation by Marker
Measures scoring variability (lower = more consistent)
Pathologist ER Pre ER Post PR Pre PR Post Ki67 Pre Ki67 Post
Pathologist 1 0.520 0.528 1.182 1.161 0.869 0.678
Pathologist 2 0.521 0.527 1.132 1.146 0.761 0.692
Pathologist 3 0.518 0.522 1.167 1.178 0.849 0.710
Pathologist 4 0.542 0.538 1.226 1.212 0.803 0.685

10.9 Seniority and AI Adoption Patterns

Evaluate how pathologist experience level relates to AI adoption behavior. Wu et al. (2023) found that junior pathologists benefited most from AI assistance (Wu et al. 2023).

Assumption: Pathologist 1 is the most senior and Pathologist 4 is the most junior (P1 → P2 → P3 → P4).

AI Adoption by Seniority Level
P1 (most senior) → P4 (most junior)
Pathologist Seniority Self-Consistency (r)1 Mean |Δ|2 Change Frequency
Pathologist 1 Most Senior 0.950 6.499 0.592
Pathologist 2 Senior 0.975 4.320 0.547
Pathologist 3 Junior 0.965 4.655 0.391
Pathologist 4 Most Junior 0.981 3.155 0.310
1 Self-Consistency: average Pre-Post correlation across ER, PR, Ki67 (higher = less AI influence)
2 Mean |Δ|: average absolute change magnitude across markers

10.9.1 Seniority–Adoption Correlation

Seniority–AI Adoption Correlation
Rank: 1 = most senior, 4 = most junior
Metric Spearman ρ p-value Interpretation
Self-Consistency (r) 0.800 0.333 Junior → higher consistency (less AI influence)
Mean Absolute Change −0.800 0.333 Junior → smaller changes (less AI influence)
Change Frequency −1.000 0.083 Junior → less frequent changes

10.9.2 Seniority Trend Visualization

10.10 Performance Summary

Create a comprehensive performance scorecard for each pathologist.

Pathologist Performance Scorecard
Summary metrics by pathologist and seniority level
Pathologist Seniority Self-Consistency (r)1 AI Influence2 Change Frequency3
Pathologist 1 Most Senior 0.950 0.650 0.761
Pathologist 2 Senior 0.975 0.432 0.675
Pathologist 3 Junior 0.965 0.465 0.493
Pathologist 4 Most Junior 0.981 0.316 0.398
1 Self-Consistency: Correlation between Pre-AI and Post-AI assessments (higher is better)
2 AI Influence: Average magnitude of changes (higher = more influenced)
3 Change Frequency: Proportion of cases changed after AI (0-1)

10.11 Conclusion

10.11.1 Key Insights

  1. Individual Variability: Pathologists differ in:

    • How consistent they are with their own initial assessments
    • How much they are influenced by AI
    • Their scoring patterns (extreme vs moderate)
  2. AI Adoption and Seniority: More junior pathologists may adopt AI suggestions more readily, consistent with Wu et al. (2023) who found that less experienced pathologists benefited most from AI assistance (Wu et al. 2023). This could reflect:

    • Lower confidence in initial assessments among junior pathologists
    • Greater openness to AI technology in less experienced practitioners
    • AI filling a knowledge gap that experience otherwise provides
    • More established scoring habits in senior pathologists
  3. Systematic Biases: Individual pathologists may consistently score higher or lower than the group median for certain markers.

  4. Marker-Specific Patterns: Pathologists may show different levels of expertise or consistency across different biomarkers.

10.11.2 Implications

  • Seniority-tailored implementation: Junior pathologists may benefit from AI as a training tool, while senior pathologists may use it as a quality check
  • Targeted training may be beneficial for pathologists with specific patterns
  • Understanding individual adoption patterns can inform AI implementation strategies
  • Quality assurance should account for individual variability
  • Pathologists with high consistency but low AI influence may benefit from feedback
  • Those with high AI influence but low consistency may need additional training