-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy path_main.tex
5089 lines (4134 loc) · 447 KB
/
_main.tex
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
\documentclass[openright,titlepage,12pt,a4paper]{book}
\usepackage{lmodern}
\usepackage{amssymb,amsmath}
\usepackage{ifxetex,ifluatex}
\usepackage{fixltx2e} % provides \textsubscript
\ifnum 0\ifxetex 1\fi\ifluatex 1\fi=0 % if pdftex
\usepackage[T1]{fontenc}
\usepackage[utf8]{inputenc}
\else % if luatex or xelatex
\ifxetex
\usepackage{mathspec}
\else
\usepackage{fontspec}
\fi
\defaultfontfeatures{Ligatures=TeX,Scale=MatchLowercase}
\fi
% use upquote if available, for straight quotes in verbatim environments
\IfFileExists{upquote.sty}{\usepackage{upquote}}{}
% use microtype if available
\IfFileExists{microtype.sty}{%
\usepackage{microtype}
\UseMicrotypeSet[protrusion]{basicmath} % disable protrusion for tt fonts
}{}
\usepackage{hyperref}
\hypersetup{unicode=true,
pdfauthor={Duco Veen},
pdfborder={0 0 0},
breaklinks=true}
\urlstyle{same} % don't use monospace font for urls
\usepackage{color}
\usepackage{fancyvrb}
\newcommand{\VerbBar}{|}
\newcommand{\VERB}{\Verb[commandchars=\\\{\}]}
\DefineVerbatimEnvironment{Highlighting}{Verbatim}{commandchars=\\\{\}}
% Add ',fontsize=\small' for more characters per line
\usepackage{framed}
\definecolor{shadecolor}{RGB}{248,248,248}
\newenvironment{Shaded}{\begin{snugshade}}{\end{snugshade}}
\newcommand{\AlertTok}[1]{\textcolor[rgb]{0.94,0.16,0.16}{#1}}
\newcommand{\AnnotationTok}[1]{\textcolor[rgb]{0.56,0.35,0.01}{\textbf{\textit{#1}}}}
\newcommand{\AttributeTok}[1]{\textcolor[rgb]{0.77,0.63,0.00}{#1}}
\newcommand{\BaseNTok}[1]{\textcolor[rgb]{0.00,0.00,0.81}{#1}}
\newcommand{\BuiltInTok}[1]{#1}
\newcommand{\CharTok}[1]{\textcolor[rgb]{0.31,0.60,0.02}{#1}}
\newcommand{\CommentTok}[1]{\textcolor[rgb]{0.56,0.35,0.01}{\textit{#1}}}
\newcommand{\CommentVarTok}[1]{\textcolor[rgb]{0.56,0.35,0.01}{\textbf{\textit{#1}}}}
\newcommand{\ConstantTok}[1]{\textcolor[rgb]{0.00,0.00,0.00}{#1}}
\newcommand{\ControlFlowTok}[1]{\textcolor[rgb]{0.13,0.29,0.53}{\textbf{#1}}}
\newcommand{\DataTypeTok}[1]{\textcolor[rgb]{0.13,0.29,0.53}{#1}}
\newcommand{\DecValTok}[1]{\textcolor[rgb]{0.00,0.00,0.81}{#1}}
\newcommand{\DocumentationTok}[1]{\textcolor[rgb]{0.56,0.35,0.01}{\textbf{\textit{#1}}}}
\newcommand{\ErrorTok}[1]{\textcolor[rgb]{0.64,0.00,0.00}{\textbf{#1}}}
\newcommand{\ExtensionTok}[1]{#1}
\newcommand{\FloatTok}[1]{\textcolor[rgb]{0.00,0.00,0.81}{#1}}
\newcommand{\FunctionTok}[1]{\textcolor[rgb]{0.00,0.00,0.00}{#1}}
\newcommand{\ImportTok}[1]{#1}
\newcommand{\InformationTok}[1]{\textcolor[rgb]{0.56,0.35,0.01}{\textbf{\textit{#1}}}}
\newcommand{\KeywordTok}[1]{\textcolor[rgb]{0.13,0.29,0.53}{\textbf{#1}}}
\newcommand{\NormalTok}[1]{#1}
\newcommand{\OperatorTok}[1]{\textcolor[rgb]{0.81,0.36,0.00}{\textbf{#1}}}
\newcommand{\OtherTok}[1]{\textcolor[rgb]{0.56,0.35,0.01}{#1}}
\newcommand{\PreprocessorTok}[1]{\textcolor[rgb]{0.56,0.35,0.01}{\textit{#1}}}
\newcommand{\RegionMarkerTok}[1]{#1}
\newcommand{\SpecialCharTok}[1]{\textcolor[rgb]{0.00,0.00,0.00}{#1}}
\newcommand{\SpecialStringTok}[1]{\textcolor[rgb]{0.31,0.60,0.02}{#1}}
\newcommand{\StringTok}[1]{\textcolor[rgb]{0.31,0.60,0.02}{#1}}
\newcommand{\VariableTok}[1]{\textcolor[rgb]{0.00,0.00,0.00}{#1}}
\newcommand{\VerbatimStringTok}[1]{\textcolor[rgb]{0.31,0.60,0.02}{#1}}
\newcommand{\WarningTok}[1]{\textcolor[rgb]{0.56,0.35,0.01}{\textbf{\textit{#1}}}}
\usepackage{longtable,booktabs}
\usepackage{graphicx,grffile}
\makeatletter
\def\maxwidth{\ifdim\Gin@nat@width>\linewidth\linewidth\else\Gin@nat@width\fi}
\def\maxheight{\ifdim\Gin@nat@height>\textheight\textheight\else\Gin@nat@height\fi}
\makeatother
% Scale images if necessary, so that they will not overflow the page
% margins by default, and it is still possible to overwrite the defaults
% using explicit options in \includegraphics[width, height, ...]{}
\setkeys{Gin}{width=\maxwidth,height=\maxheight,keepaspectratio}
\IfFileExists{parskip.sty}{%
\usepackage{parskip}
}{% else
\setlength{\parindent}{0pt}
\setlength{\parskip}{6pt plus 2pt minus 1pt}
}
\setlength{\emergencystretch}{3em} % prevent overfull lines
\providecommand{\tightlist}{%
\setlength{\itemsep}{0pt}\setlength{\parskip}{0pt}}
\setcounter{secnumdepth}{5}
% Redefines (sub)paragraphs to behave more like sections
\ifx\paragraph\undefined\else
\let\oldparagraph\paragraph
\renewcommand{\paragraph}[1]{\oldparagraph{#1}\mbox{}}
\fi
\ifx\subparagraph\undefined\else
\let\oldsubparagraph\subparagraph
\renewcommand{\subparagraph}[1]{\oldsubparagraph{#1}\mbox{}}
\fi
%%% Use protect on footnotes to avoid problems with footnotes in titles
\let\rmarkdownfootnote\footnote%
\def\footnote{\protect\rmarkdownfootnote}
%%% Change title format to be more compact
\usepackage{titling}
% Create subtitle command for use in maketitle
\providecommand{\subtitle}[1]{
\posttitle{
\begin{center}\large#1\end{center}
}
}
\setlength{\droptitle}{-2em}
\title{}
\pretitle{\vspace{\droptitle}}
\posttitle{}
\author{Duco Veen}
\preauthor{\centering\large\emph}
\postauthor{\par}
\predate{\centering\large\emph}
\postdate{\par}
\date{Department of Methodology \& Statistics, Utrecht University}
% \documentclass[12pt, a4paper, titlepage]{article}
\usepackage[margin=1in]{geometry}
\usepackage{pdflscape} % this with newcommand for landscape pages
\newcommand{\blandscape}{\begin{landscape}} %redefine landscape argument from pdflscape \begin{landscape} doesn't work in bookdown
\newcommand{\elandscape}{\end{landscape}}
\usepackage{float} % to allow you to put tables and figures in specific places
\usepackage{fancyhdr} % for short chapter names https://github.com/rstudio/bookdown/issues/360
% https://tex.stackexchange.com/questions/7276/table-caption-width
\usepackage[width = \textwidth, textfont = it]{caption}
% https://en.wikibooks.org/wiki/LaTeX/Footnotes_and_Margin_Notes
\makeatletter
\def\blfootnote{\xdef\@thefnmark{}\@footnotetext}
\makeatother
\frontmatter % page numbering starts at introduction
% \pagenumbering{roman}
% \usepackage{booktabs}
% \usepackage{longtable}
%
%
% \ifxetex
% \usepackage{letltxmacro}
% \setlength{\XeTeXLinkMargin}{1pt}
% \LetLtxMacro\SavedIncludeGraphics\includegraphics
% \def\includegraphics#1#{% #1 catches optional stuff (star/opt. arg.)
% \IncludeGraphicsAux{#1}%
% }%
% \newcommand*{\IncludeGraphicsAux}[2]{%
% \XeTeXLinkBox{%
% \SavedIncludeGraphics#1{#2}%
% }%
% }%
% \fi
% \usepackage{fullpage}
% \usepackage{pdflscape}
% \usepackage[margin=1in]{geometry} %for marings
% \usepackage{setspace} % for double spacing
% \usepackage{float}
% \doublespacing
%\usepackage{newtxtext,newtxmath} % for times new roman
% \usepackage{makeidx} % make index https://bookdown.org/yihui/bookdown/latex-index.html
% \makeindex
\begin{document}
\pagenumbering{gobble}
% \pagestyle{empty}
\begin{center}
\huge{\textbf{Alternative Information}}
% \huge{\textbf{Alternative Information}}
\Large{Bayesian Statistics, Expert Elicitation and Information Theory in the Social Sciences}
\vspace*{1cm}
\large{\textbf{Alternatieve Informatie}}
\normalsize{Bayesiaanse Statistiek, Expert Elicitatie en Informatie Theorie in de Sociale Wetenschappen}
\vspace*{.3cm}
\normalsize{(met een samenvatting in het Nederlands)}
\vspace*{2cm}
\Large{\textbf{Proefschrift}}
\vspace*{3cm}
\normalsize
ter verkrijging van de graad van doctor aan de \\
Universiteit Utrecht \\
op gezag van de \\
rector magnificus, prof.dr. H.R.B.M. Kummeling, \\
ingevolge het besluit van het college voor promoties \\
in het openbaar te verdedigen op
\vspace*{.5cm}
vrijdag 13 maart 2020 des ochtends te 10.30 uur
\vspace*{1.5cm}
door
\vspace*{1.5cm}
\Large{\textbf{Duco Veen}}
\normalsize
\vspace*{1cm}
geboren op 13 maart 1990
te Zutphen
\end{center}
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% newpage
\newpage
\pagestyle{empty}
\textbf{Promotoren:}
Prof.dr. A.G.J. van de Schoot
\textbf{Copromotoren:}
dr. G. Vink
dr. N.E.E. van Loey
\vspace*{\fill}
\noindent The studies in this thesis were funded by the Netherlands Organization for Scientific Research grant NWO-VIDI-452-14-006.
%
\newpage
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% newpage
\textbf{Beoordelingscommissie:}
Prof. dr. R. Geenen
Prof. dr. I.G. Klugkist
Dr. D.L. Oberski
Prof. dr. S. van der Stigchel
Prof. dr. E.M. Wagenmakers
\vspace*{\fill}
\noindent Alternative Information: Bayesian Statistics, Expert Elicitation and Information Theory in the Social Sciences.\\
\noindent Proefschrift Universiteit Utrecht, Utrecht. \\
\noindent Met samenvatting in het Nederlands. \\
\noindent ISBN: 978-94-6375-796-6 \\
\noindent Copyright \textcopyright 2020, D. Veen. All Rights Reserved.
\newpage
\textbf{De Quasi-neutrale Oplossingsfabriek}
\textit{Een tocht door het land van de ivoren torens, waar de objectieve waarheid wordt gemaakt}
De utopische traditie is de afgelopen decennia drastisch uitgedund. Wat ons resteert zijn technotopia's, klimaatdystopia's en nostalgie naar de sixties. Maar als altijd, hebben wij ook vandaag een baken nodig aan de horizon van tijd, om naartoe te koersen en gezamenlijk grote beslissingen te kunnen nemen. Eeuwenlang was de wetenschap de plek bij uitstek waar Utopia ontstond, vooral ten tijde van een intellectuele revolutie. Dus waarom ontstaan er geen utopieën vandaag? We bezoeken het land van de ivoren torens.
Hoe komen we tot kennis? Uit de wereld om ons heen verzamelen we gegevens die wij omhoog brengen naar het hart van de wetenschap, de nok van de ivoren toren waar de waarheid wordt gemaakt. En wanneer een theorie is gemaakt, gaan we checken of het klopt. Want dat is tenslotte wat wij doen op de universiteit, we construeren waarheden. Allereerst zal de kennis door de objectieve waarheidstrechter gaan, waar het wordt gestript van kwalitatieve en subjectieve aspecten en van normen en waarden. Wat er overblijft, neutrale cijfers, gaat naar de binaire pers, klaar voor de computers van de programmeurs. Met algoritmes transformeren zij de eentjes en nulletjes in efficiëntie, snelheid en groei, en schuiven het dan door naar de economen. Zittend op het GNP en maaiend met grote grijpers plaatsen de economen iedereen netjes ergens in de fabriek van onze maatschappij.
In blinde processie bewegen we voort op het tikken van de klok van de ene cel naar de volgende, gevangen in de oneindigheid van onze dagelijkse routines. Billboards moedigen ons aan om geld dat we niet hebben uit te geven aan zooi die we willen, om indruk te maken op de mensen die we eigenlijk niet uit kunnen staan. Rechts van de kloof vinden we de kunstenaars, muzikanten, filosofen, én de utopisten - bestempeld als dromers of extremisten. Voor hen is het te gevaarlijk om over te steken, zij zijn overbodig hier, in een maatschappij van cijfers, feiten en neutrale waarheden.
Eeuwenlang was de wetenschap bij uitstek de plek waar Utopia ontstond, maar vandaag is dat veranderd. De universiteit van de 21ste eeuw is eigenlijk niets meer dan een fabriek, een quasi-neutrale oplossingsfabriek, waar iedereen te druk is met schrijven om nog iets te lezen, en te druk met publiceren voor een debat. Als wij willen dat er waardevolle utopieën kunnen ontstaan binnen de universiteit, dan zal het proces van kennis verwerven en de interpretatie van waarheid misschien wel moeten veranderen...
\vspace*{\fill}
Illustratie op achterzijde en bijbehorend verhaal door:
Carlijn Kingma (https://www.carlijnkingma.com/)
\newpage
%\frontmatter
% \makeatletter
% \renewcommand\chaptermark[1]{%
% \markboth{\MakeUppercase{#1}}{}
% }
% \makeatother
%
% \markboth{CONTENTS}{CONTENTS}
\setcounter{tocdepth}{1}
\tableofcontents
\thispagestyle{empty}
\addtocontents{toc}{\protect\thispagestyle{empty}}
% https://tex.stackexchange.com/questions/2995/removing-page-number-from-toc
\mainmatter
\pagestyle{headings}
\renewcommand{\chaptername}{}
\hypertarget{introduction}{%
\chapter{Introduction}\label{introduction}}
\chaptermark{INTRODUCTION}
\thispagestyle{empty}
\hypertarget{bayesian-statistics}{%
\section{Bayesian Statistics}\label{bayesian-statistics}}
We all have to make decisions whilst facing uncertainty and incomplete information. To help us interpret and organize available information we use statistics. One framework that is often used to plan the optimal course of action is Bayesian statistics.
Bayesian statistics offers a way to describe our state of knowledge in terms of probability (Jaynes, \protect\hyperlink{ref-jaynes_bayesian_1996}{1996}). Moreover, it can be seen as an extension of logic (Jaynes, \protect\hyperlink{ref-jaynes_probability_2003}{2003}). In addition, Bayesian statistics describes how we ought to learn (Lindley, \protect\hyperlink{ref-lindley_understanding_2013}{2013}). We can do so by using probability distributions to describe our state of knowledge about a parameter. This can be done both before we observe new data, i.e.~by means of a \emph{prior} distribution of probability concerning the parameter, or after we have observed new data and we have updated our state of knowledge, i.e.~the \emph{posterior} distribution of probability concerning the parameter.
To make this more intuitive I very briefly describe learning via Bayesian statistics. I use the example describing how we could learn about the unknown proportion of a sequence of `Bernoulli trials' that result in either 0 or 1, or in case of a coin, tails (\(T\)) for 0 or head (\(H\)) for 1. We say that \(\theta\) is the proportion of coin flips resulting in heads facing upwards. It turns out that we can use the Beta distribution in a very convenient way to update our beliefs, or state of knowledge, concerning \(\theta\). That is, we can express which values are consistent with both our \emph{prior} state of knowledge and the newly \emph{observed data} (Jaynes, \protect\hyperlink{ref-jaynes_bayesian_1996}{1996}). The distribution of probability indicates which values are most consistent with both sources. For mathematical details see for instance Gelman et al. (\protect\hyperlink{ref-gelman_bayesian_2013}{2013} Chapter 2). The intuition is as follows: the Beta distribution has two parameters, \(\alpha\) and \(\beta\), which can be interpreted as follows in our example; there have `Bernoulli trials', and \(\alpha - 1\) of them have been a success whilst \(\beta - 1\) of them have been a failure. In other words we have observed heads \(\alpha - 1\) times and tails \(\beta - 1\) times.
Now let us start with a prior state of ignorance, we have neither observed head nor tails before. We then specify a \(Beta(\alpha = 1, \beta = 1)\) \emph{prior} distribution. It turns out that this neatly coincides with an initial state of ignorance. Every proportion in the interval from 0 up to 1 is assigned equal probability to be the value for \(\theta\) based on no initial evidence, see Figure \ref{fig:binomUninformative} panel A. Now we observe heads four times and tails once (\(THHHH\)) in the first five trials and we learn from this data such that we update to a \emph{posterior} distribution represented by a \(Beta(\alpha = 5, \beta = 2)\), which can be seen in Figure \ref{fig:binomUninformative} panel B. Before we observe more trials and new data we have an updated state of belief. The \emph{posterior} distribution can become our new \emph{prior} distribution, which we, in turn, update with new information to obtain a new posterior distribution. This is what happens in panels C and D of Figure \ref{fig:binomUninformative} where we in turn observe \(HTHHT\) and \(THTTH\) to come to a \(Beta(\alpha = 8, \beta = 4)\) as a posterior in panel C and a \(Beta(\alpha = 10, \beta = 7)\) as posterior in panel D. After 15 trials, and without initial prior knowledge, slightly more heads were observed than tails, thus values just above a proportion of .5 are assigned the largest probability. However, given the few trials that we observed, a wide range of possible values for the proportion of coin flips resulting in heads facing upwards are still assigned probability. Note too, that nowhere do I state which value for \(\theta\) I used to simulate these results, for in practice this is unknown and the best we can do is what we just did, use the knowledge available to us to assign probabilities to values for \(\theta\).
\begin{figure}
{\centering \includegraphics{_main_files/figure-latex/binomUninformative-1}
}
\caption{Example of Bayesian updating. Panel A shows a $Beta(\alpha = 1, \beta = 1)$ distribution representing a prior state of knowledge equal to ignorance. Panels B, C and D show how the state of knowledge updated after new data is observed, each time the previous panel is the prior belief for the next panel, combined with the information from five new observations.}\label{fig:binomUninformative}
\end{figure}
Now, let us suppose that we did not have an initial state of ignorance. The \emph{prior} need not be ignorance as we noticed when the previous posterior became our new prior each time. Would of belief differ if we had more initial information? Figure \ref{fig:binomInf} shows learning from the same data as in the example presented in Figure \ref{fig:binomUninformative} with our initial state of knowledge expressed by a \(Beta(\alpha = 51, \beta = 51)\) distribution. In other words, before the new trials we had initial information equivalent to 100 previous coin flips that were distributed equally between head and tails. The new data is very much in line with our previous data and we only slightly adjust our beliefs, assigning even more probability to values near .5.
\begin{figure}
{\centering \includegraphics{_main_files/figure-latex/binomInf-1}
}
\caption{Example of Bayesian updating. Panel A shows a $Beta(\alpha = 51, \beta = 51)$ distribution. This is updated using the same data as in Figure 1.1, only the initial prior contains more information.}\label{fig:binomInf}
\end{figure}
All of this wonderful nuanced theory is historically summarized by a single equation. The reason for showing you this formula only after the examples, is not to get distracted by the mathematics for those readers not working with statistics every day. For those who do use statistics often, there are many books written in much more detail on this subject that yield a more complete overview (e.g.~Gelman et al., \protect\hyperlink{ref-gelman_bayesian_2013}{2013}; Jaynes, \protect\hyperlink{ref-jaynes_probability_2003}{2003}; Kaplan, \protect\hyperlink{ref-kaplan_bayesian_2014}{2014}; Lindley, \protect\hyperlink{ref-lindley_understanding_2013}{2013}; Lynch, \protect\hyperlink{ref-lynch_introduction_2007}{2007}; Ntzoufras, \protect\hyperlink{ref-ntzoufras_bayesian_2011}{2011}; Press, \protect\hyperlink{ref-press_subjective_2009}{2009}). Without further ado, Bayes' Theorem
\begin{equation}
p(A|BC) = P(A|C)\frac{P(B|AC)}{P(B|C)}
\label{eq:bayestheorem}
\end{equation}
where \(A\), \(B\) and \(C\) are different propositions, \(p(A|C)\) describes the \emph{prior} distribution of probability concerning \(A\), given that we know \(C\). \(p(A|BC)\) is the \emph{posterior} distribution of probability concerning \(A\), updated with the new information that is provided to us by \(B\). Note that \(C\) here has the interpretation of what we know about \(A\) before learning about, or obtaining, the information from \(B\). In the previous example, \(\theta\) took on the role of \(A\) and the new trials took on the role of \(B\). \(C\), in the first example, expressed that we knew that \(\theta\) was a proportion which can only take on values in the interval from 0 up to 1. Equation \eqref{eq:bayestheorem} describes how we can learn about \(A\), how we ought to update our beliefs in the light of new data. It also makes it explicit that this learning effect is dependent on our prior knowledge, just like in the example of Figures \ref{fig:binomUninformative} and \ref{fig:binomInf} Again, the aim here is not to expand on the mathematics, but merely to provide some initial intuition about the concept of learning using Bayes' rule. Next, we turn to the implementation of the concept of prior knowledge, how can we formalize our prior distribution of probability concerning \(A\) given that we know \(C\).
\hypertarget{prior-information}{%
\section{Prior Information}\label{prior-information}}
The topic of prior information in Bayesian statistics is a field of study all on it's own. To illustrate this, note that Tuyl, Gerlach, \& Mengersen (\protect\hyperlink{ref-tuyl_comparison_2008}{2008}) wrote a full paper, discussing only the choice of prior for extreme cases of our example above. There are heavy discussions and different schools of thought, roughly divided into ``objective'' (e.g.~Berger, \protect\hyperlink{ref-berger_case_2006}{2006}) and ``subjective'' (e.g.~Finetti, \protect\hyperlink{ref-de_finetti_theory_1974}{1974}; Goldstein, \protect\hyperlink{ref-goldstein_subjective_2006}{2006}) camps. Discussing the differences within, let alone between, these different approaches is way to much to get into at this point. Simply listing names of different approaches to objective Bayesian priors takes up an extensive paragraph (Berger, \protect\hyperlink{ref-berger_case_2006}{2006}, pp. 387--388) and for a discussion about the (dis)advantages of both method I refer the reader to Chapter 5 of Press (\protect\hyperlink{ref-press_subjective_2009}{2009}). For the reader it suffices to know that in this dissertation we certainly specify priors that would be considered more in line with the ``subjective'' school of Bayesian analysis, even if we do not always use these priors to be updated with new data.
Now let us briefly consider three types of information that could be included in a prior distribution of probability. First, previous research can inform us about certain parameters and including this information in future analyses seems in line with our idea of leaning. In Section 5.4 of Spiegelhalter, Abrams, \& Myles (\protect\hyperlink{ref-spiegelhalter_bayesian_2004}{2004}) they provide a very nice overview on how to include results of previous studies based on similarity, exchangability and bias considerations. It is described not only how you could include information from previous studies if they are exactly on the same topic, but also how to do so if the research differs in specific ways. Second, logical considerations can be taken into account, e.g.~in the coin flipping example we know that a proportion lies between 0 and 1 and no values outside the interval between those two will be assigned any probability. In a similar fashion we could incorporate information with respect to our measurements, e.g.~no negative values for temperature measured in Kelvin, or when calculating air pollution in a city that we do live in, the amount of matter in the air cannot be so much that we could not breath and live there. Third,
information gathered from an expert, or as put by the Dictionary (\protect\hyperlink{ref-cambridge_english_dictionary_expert_2019}{2019}); ``\emph{a person with a high level of knowledge or skill relating to a particular subject or activity}''. This particular knowledge can be translated, or elicited, to be expressed in the form of distribution of probability. There are surely more sources of information to inform our prior probability distributions besides previous research, logical considerations and expert knowledge, but as this dissertations involves experts quite a bit I will elaborate on that specific case somewhat more.
\hypertarget{expert-elicitation}{%
\section{Expert Elicitation}\label{expert-elicitation}}
The process of creating a probabilistic representation of an experts' beliefs is called elicitation (O'Hagan et al., \protect\hyperlink{ref-ohagan_uncertain_2006}{2006}). There is an extensive history of expert elicitation across many different disciplines of sciences (Cooke \& Goossens, \protect\hyperlink{ref-cooke_tu_2008}{2008}; Gosling, \protect\hyperlink{ref-gosling_shelf:_2018}{2018}; O'Hagan et al., \protect\hyperlink{ref-ohagan_uncertain_2006}{2006}, Chapter 10). Expert judgements are for instance used in the case that there is no actual data available (Hald et al., \protect\hyperlink{ref-hald_world_2016}{2016}) or to add information to small sample data (Kadane, \protect\hyperlink{ref-kadane_application_1994}{1994}). However, with many examples, covering many disciplines, expert knowledge still does not seem to be used in the social sciences with Van de Schoot, Winter, Ryan, Zondervan-Zwijnenburg, \& Depaoli (\protect\hyperlink{ref-van_de_schoot_systematic_2017}{2017}) finding only two use cases in the past 25 years of Bayesian statistics in Psychology.
One of the reasons for this limited use can perhaps be found in the traditional way of eliciting expert judgements. One of the traditional ways is to elicit three quantiles from an experts concerning the distribution of probability over a specific parameters (O'Hagan et al., \protect\hyperlink{ref-ohagan_uncertain_2006}{2006}, Chapter 5). These quantiles are then used to determine the distribution that fits best with these described value, and that distribution is used to represent an experts' beliefs. The questions to the experts would be for instance the following:
\begin{quote}
"Q1: Can you determine a value (your median) such that X is equally likely to be less than or greater than this point?
Q2: Suppose you were told that X is below your assessed median. Can you now determine a new value (the lower quartile) such that it is equally likely that X is less than or greater than this value?
Q3: Suppose you were told that X is above your assessed median. Can you now determine a new value (the lower quartile) such that it is equally likely that X is less than or greater than this value?"
O'Hagan et al. (\protect\hyperlink{ref-ohagan_uncertain_2006}{2006}), p.~101
\end{quote}
We believe that this rather abstract thinking in terms of quantiles of distributions might be hard for some experts, depending on their experience with statistics and mathematical background. This naturally bring us to the following section and the outline of this dissertation, how is it that we propose to use expert elicitation in the social sciences and what do we propose to do with this source of additional information.
\hypertarget{aims-and-outline}{%
\section{Aims and Outline}\label{aims-and-outline}}
In this dissertation it is discussed how one can capture and utilize alternative sources of (prior) information compared to traditional method in the social sciences such as survey research. Specific attention is paid to expert knowledge.
In Chapter \protect\hyperlink{fivestep}{2} we propose an elicitation methodology for a single parameter that does not rely on specifying quantiles of a distribution. The proposed method is evaluated using a user feasibility study, a partial validation study and an empirical example of the full elicitation method.
In Chapter \protect\hyperlink{DAC1}{3} it is investigated how experts' knowledge, as alternative source of information, can be contrasted with traditional data collection methods. At the same time, we explore how experts can be assessed and ranked borrowing techniques from information theory. We use the information theoretical concept of relative entropy or Kullback-Leibler divergence which assesses a loss of information when approximating one distribution by another. For those familiar with the concept of model selection, Akaike's Information Criterion is an approximation of this (Burnham \& Anderson, \protect\hyperlink{ref-burnham_model_2002}{2002}, Chapter 2).
In Chapter \protect\hyperlink{Hierarchical}{4} an alternative way of enhancing the amount of information in a model is proposed. We introduce Bayesian hierarchical modelling to the field of infants' speech discrimination analysis. This technique is not new on it's own but was not applied to this field. Implementing this type of modelling enables individual analyses within a group structure. By taking the hierarchical structure of the data into account we can make the most of the, on individual level, small noisy data sets. The analysis methodology estimates if individuals are (dis)similar and takes this into account for all individuals in one single model. Essentially, the estimated group effects serve as priors for the individual estimates. Moreover, we do not need to do a single hypothesis test for every child, which was the most advanced individual analysis in the field up to now, going back to 2007 (Houston, Horn, Qi, Ting, \& Gao, \protect\hyperlink{ref-houston_assessing_2007}{2007}).
In Chapter \protect\hyperlink{Burns}{5} we reflect on issues that come along with the estimation of increasingly complicated models. Techniques and software for estimating more complex models, such as proposed in Chapter \protect\hyperlink{Hierarchical}{4}, need to be carefully used and the results of the analyses need to be carefully checked. But what to do when things actually go awry in your analysis? We show how even with weakly informative priors, adding the information that is available to us, sometimes we do not get a solution with our analysis plan. We guide the reader on what to do when this occurs and where to look for clues and possible causes. We provide some guidance and a textbook example that for once shows things not working out the way you would like. We believe this is important as there are few examples of this.
In Chapter \protect\hyperlink{elicitlgm}{6} we combine the previous chapters. We take more complex models and get experts to specify beliefs with respect to these models. We extend the method developed in Chapter \protect\hyperlink{fivestep}{2} to elicit experts' beliefs with respect to a hierarchical model, which is used in Chapters \protect\hyperlink{Hierarchical}{4} and \protect\hyperlink{Burns}{5}. In specific, we concern ourselves with a Latent Growth Curve model and utilize the information theoretical measures from Chapter \protect\hyperlink{DAC1}{3} to compare the (groups) of experts to one another and to data collected in a traditional way. We do this in the context of Posttraumatic Stress Symptoms development in children with burn injuries.
In Chapter \protect\hyperlink{thesisdiscussion}{7} I reflect on the work and explanations provided within the chapters of this dissertation, including this introduction. The discussion is a reflection of my own personal thoughts, and no other person is responsible for the content, although the personal discussions and collaborations of the past years have certainly contributed to the formulation of these opinions.
\hypertarget{fivestep}{%
\chapter{Proposal for a Five-Step Method to Elicit Expert Judgment}\label{fivestep}}
\chaptermark{FIVE-STEP METHOD}
\thispagestyle{empty}
\blfootnote{This chapter is published as Veen, D., Stoel ,D., Zondervan-Zwijnenburg, M. \& van de Schoot, R. (2017). Proposal for a Five-Step Method to Elicit Expert Judgment. \textit{Front. Psychol.}, 8:2110. doi: 10.3389/fpsyg.2017.02110 \\ \indent DV and RvdS mainly contributed to the study design. All authors have been involved in the design of (part) of the elicitation procedure. DV programmed the elicitation software. All elicitations have been facilitated by DV and DS. DV wrote and revised the paper with feedback and input of DS, MZ-Z, and RvdS. RvdS supervised the project.}
\hypertarget{abstract}{%
\section*{Abstract}\label{abstract}}
\addcontentsline{toc}{section}{Abstract}
\small
Elicitation is a commonly used tool to extract viable information from experts. The information that is held by the expert is extracted and a probabilistic representation of this knowledge is constructed. A promising avenue in psychological research is to incorporated experts' prior knowledge in the statistical analysis. Systematic reviews on elicitation literature however suggest that it might be inappropriate to directly obtain distributional representations from experts. The literature qualifies experts' performance on estimating elements of a distribution as unsatisfactory, thus reliably specifying the essential elements of the parameters of interest in one elicitation step seems implausible. Providing feedback within the elicitation process can enhance the quality of the elicitation and interactive software can be used to facilitate the feedback. Therefore, we propose to decompose the elicitation procedure into smaller steps with adjustable outcomes. We represent the tacit knowledge of experts as a location parameter and their uncertainty concerning this knowledge by a scale and shape parameter. Using a feedback procedure, experts can accept the representation of their beliefs or adjust their input. We propose a Five-Step Method which consists of (1) Eliciting the location parameter using the trial roulette method. (2) Provide feedback on the location parameter and ask for confirmation or adjustment. (3) Elicit the scale and shape parameter. (4) Provide feedback on the scale and shape parameter and ask for confirmation or adjustment. (5) Use the elicited and calibrated probability distribution in a statistical analysis and update it with data or to compute a prior-data conflict within a Bayesian framework. User feasibility and internal validity for the Five-Step Method are investigated using three elicitation studies.
\normalsize
\newpage
\hypertarget{ch02introduction}{%
\section{Introduction}\label{ch02introduction}}
\begin{quote}
``The knowledge held by expert practitioners is too valuable to be ignored. But only when thorough methods are applied, can the application of expert knowledge be as valid as the use of empirical data. The responsibility for the effective and rigorous use of expert knowledge lies with the researchers.''
Drescher et al. (\protect\hyperlink{ref-drescher_toward_2013}{2013}, p. 1)
\end{quote}
According to O'Hagan et al. (\protect\hyperlink{ref-ohagan_uncertain_2006}{2006}) elicitation is the process of extracting and creating a representation of an expert's beliefs. It can be used for a variety of reasons, e.g., to add information to small sample data (Kadane, \protect\hyperlink{ref-kadane_application_1994}{1994}; Schoot, Sijbrandij, et al., \protect\hyperlink{ref-van_de_schoot_bayesian_2018}{2018}; Zondervan-Zwijnenburg et al., \protect\hyperlink{ref-zondervan-zwijnenburg_where_2017}{2017}\protect\hyperlink{ref-zondervan-zwijnenburg_where_2017}{a}), when there is no data for certain confounding parameters in a model (Fischer, Lewandowski, \& Janssen, \protect\hyperlink{ref-fischer_estimating_2013}{2013}), when no data is available (Ho \& Smith, \protect\hyperlink{ref-ho_volcanic_1997}{1997}), as sensitivity analysis to check assumptions about missing data (Mason et al., \protect\hyperlink{ref-mason_development_2017}{2017}), or simply to enrich the available data (Wisniowski, Bijak, \& Shang, \protect\hyperlink{ref-wisniowski_forecasting_2014}{2014}). Expert knowledge is a valuable source of information, as becomes evident in the quote of Drescher et al. (\protect\hyperlink{ref-drescher_toward_2013}{2013}). (Hadorn, Kvizhinadze, Collinson, \& Blakely, \protect\hyperlink{ref-hadorn_useof_2014}{2014}) found that 57\% of health economic decision models included at least one expert knowledge elicitation parameter, showing that in some fields it is even the norm to use expert elicitation. More examples of elicitation practices in many different fields can be found in overview studies by O'Hagan et al. (\protect\hyperlink{ref-ohagan_uncertain_2006}{2006}, Chapter 10) and Bistline (\protect\hyperlink{ref-bistline_energy_2014}{2014}) or the paper by Cooke \& Goossens (\protect\hyperlink{ref-cooke_tu_2008}{2008}) in which they describe the data base of over 67,000 experts' subjective probability distributions.
There are many elicitation procedures available, overviews can be found in for instance O'Hagan et al. (\protect\hyperlink{ref-ohagan_uncertain_2006}{2006}), S. R. Johnson et al. (\protect\hyperlink{ref-johnson_methods_2010}{2010}), and Aspinall \& Cooke (\protect\hyperlink{ref-aspinall_quantifying_2013}{2013}). A popular elicitation method is the trial roulette method (Gore, \protect\hyperlink{ref-gore_biostatistics_1987}{1987}), sometimes also called the chips and bins method or the histogram method, in which experts assign ``chips''" to ``bins'' of a histogram to ascribe probability. In the procedure, used by for instance Diamond et al. (\protect\hyperlink{ref-diamond_expert_2014}{2014}) and Goldstein \& Rothschild (\protect\hyperlink{ref-goldstein_lay_2014}{2014}), the parameter space for which experts can assign probability is divided into equal sections or ``bins''. The experts receive 20 ``chips'', which are to be distributed amongst these ``bins''. For each ``chip'' that is allocated to one of the ``bins'', 5\% of the mass of a probability distribution is ascribed. Based on the input provided by the expert, a probability distribution is fitted. The trial roulette method has been validated by S. R. Johnson, Tomlinson, et al. (\protect\hyperlink{ref-johnson_valid_2010}{2010}) and M. Zondervan-Zwijnenburg et al. (\protect\hyperlink{ref-zondervan-zwijnenburg_application_2017}{2017}\protect\hyperlink{ref-zondervan-zwijnenburg_application_2017}{b}) in a face-to-face setting.
Software that can be used in the elicitation with the trial roulette method is available in the MATCH Uncertainty Elicitation Tool (Morris, Oakley, \& Crowe, \protect\hyperlink{ref-morris_web-based_2014}{2014}). MATCH is an online framework for elicitation procedures. It uses the R-package (R Core Team, \protect\hyperlink{ref-r_core_team_r:_2017}{2017}\protect\hyperlink{ref-r_core_team_r:_2017}{b}) SHELF (Oakley, \protect\hyperlink{ref-R-SHELF}{2019}) to fit appropriate parametric distributions based on input that is provided by experts.
One of the reasons the trial roulette method is popular is that the procedure provides immediate visual feedback to experts. Feedback is important in elicitation procedures to reduce bias and improve the quality of the elicitation (Johnson et al., \protect\hyperlink{ref-johnson_methods_2010}{2010}; O'Hagan et al., \protect\hyperlink{ref-ohagan_uncertain_2006}{2006}). The ``chips''" that are allocated in the trial roulette method by the expert visually approximate a probability distribution. However, the feedback provided to the expert is not on the statistical distribution that is actually used by the researcher in the final analyses. It is important to receive conformation of the expert that the interpretation by the researcher matches their beliefs, or as O'Hagan et al. (\protect\hyperlink{ref-ohagan_uncertain_2006}{2006}, p. 174) state, \emph{``feedback to the expert is the most natural way of evaluating the distribution -- the expert is in the best position to judge whether something corresponds to her opinion.''} Providing instant feedback on the representation of the experts' beliefs, based on the input they provide, and how their beliefs are translated into a statistical distribution can easily be done by using software.
Feedback is believed to improve the quality of the elicitation procedure by making experts; reflect and maintain selfconsistency (Fisher, O'Leary, Low-Choy, Mengersen, \& Caley, \protect\hyperlink{ref-fisher_software_2012}{2012}), by highlighting inconsistencies in judgment and making errors apparent (Morris et al., \protect\hyperlink{ref-morris_web-based_2014}{2014}; O'Hagan et al., \protect\hyperlink{ref-ohagan_uncertain_2006}{2006}) and by allowing for self-correction by experts (Johnson et al., \protect\hyperlink{ref-johnson_methods_2010}{2010}). Despite assumed quality improvement by feedback, systematic reviews on elicitation literature by O'Hagan et al. (\protect\hyperlink{ref-ohagan_uncertain_2006}{2006}) and Johnson et al. (\protect\hyperlink{ref-johnson_methods_2010}{2010}) conclude that measurement properties of elicitation methods have not been adequately evaluated. Moreover, there is no direct research into how accurate experts can assess properties like the mean, mode, or variance for the distribution of an uncertain parameter. Research by S. R. Johnson, Tomlinson, et al. (\protect\hyperlink{ref-johnson_valid_2010}{2010}) and (Zondervan-Zwijnenburg et al., \protect\hyperlink{ref-zondervan-zwijnenburg_application_2017}{2017}\protect\hyperlink{ref-zondervan-zwijnenburg_application_2017}{b}) provide promising results concerning the trial roulette method. Yet, directly obtaining distributional representations may be inappropriate given experts' unsatisfactory performance on specifying elements of this distribution. O'Hagan et al. (\protect\hyperlink{ref-ohagan_uncertain_2006}{2006}) refer to research by Hofstatter (\protect\hyperlink{ref-hofstatter_uber_1939}{1939}), Lathrop (\protect\hyperlink{ref-lathrop_perceived_1967}{1967}) and, Beach \& Scopp (\protect\hyperlink{ref-beach_intuitive_1968}{1968}) to show that experts are not good at interpreting and assigning numerical values to variances and relative variability. It might then be unreasonable to assume that experts are able to reliably specifying a probability distribution in one step.
Therefore, to assist experts in the process of creating a representation of their beliefs in a statistical distribution we propose to decompose the elicitation task in smaller steps to encourage and assist in structured reasoning. Decomposing a problem into more tractable and familiar components is suggested by for instance Fischhoff (\protect\hyperlink{ref-fischhoff_debiasing_1982}{1982}) to decrease the mismatch between the judge and the task. By decomposing the elicitation task we aim to reduce bias and incorporate more feedback to ensure that experts' opinions are properly calibrated and represented by the probability distributions that results from the elicitation. In the current paper, the statistical distribution of interest is the skewed normal (SN) distribution\footnote{Using the SN distribution we represent the tacit knowledge of experts by eliciting the location parameter of the distribution, in this case the mean. The uncertainty of the expert about his/her belief on the location parameter is represented by the scale and shape parameter (i.e., variance and skewness of the normal distribution). Eliciting the mean of a normal distribution offers the advantage of easily transformable scale for elicitation procedures. An adjustable scale means that even if one expert reasons in averages and the other expert in sums they can be transformed to be comparable, i.e., let \(\theta\) represent the parameter of interest and \(\theta \sim N(\mu, \sigma^2)\) and if we transform \(\theta\) via the following function \(\theta^* = a\theta+b\), then \(\theta^* \sim N[a\mu+b,(a\sigma)^2]\).} because uncertainty might typically not best be captured by a symmetric distribution. This (un)certainty is the key feature of Bayesian statistics, uncertainty reveals the extent of our knowledge and ignorance (Finetti, \protect\hyperlink{ref-de_finetti_theory_1974}{1974}).
We propose the Five-Step Method which consists out of the following steps:
\begin{enumerate}
\def\labelenumi{\arabic{enumi}.}
\tightlist
\item
Elicit the location parameter of the SN using the trial roulette method.
\item
Use software to provide instant feedback on the interpretation of the expert's beliefs by the researcher so the expert can accept this representation or adjust their input.
\item
Elicit the (un)certainty of the expert by determining the scale and shape parameters of the SN using expert's statements on the lower and upper bounds for a plausible range of the parameter values.
\item
Use software to provide instant feedback on the interpretation of the expert's (un)certainty about the location parameter by the researcher so expert can accept this representation or adjust their input.
\item
Use the elicited and calibrated probability distribution in a Bayesian analysis to update it with data or to compute a prior-data conflict.
\end{enumerate}
The remainder of the paper is ordered as follows. We first provide details on the Five-Step Method. Thereafter we present a user feasibility study in which we elicited beliefs regarding a trivial sports related question from respondents to investigate visual and procedural preferences of users for the digitized version of the trial roulette method. A second study was carried out by asking experts working at a staffing company about certain key performance indicators which we used to validate the internal validity of steps 1 and 2 of the elicitation procedure. A final study was done with regional directors working at a large financial institution. They provided actual forecasts concerning average turnover per professional in the first quarter of the year 2016 with the Five-Step Method. The participating companies already make predictions concerning the parameters we elicit, yet they do this in the form of point estimates. The experts are thus already used to thinking about these data and predicting these data which makes them highly suitable to include as experts in an elicitation exercise. Yet, it is an extension for them to actively specify and separate knowledge and uncertainty. Because the companies also provided us with data on the predicted parameters we were able to compare the forecasts of the experts with data and thereby get an indication of the internal validity of the elicitation procedure. The proposition to split the elicitation process results in a procedure differing from the existing elicitation procedures as, for example, proposed by Oakley (\protect\hyperlink{ref-oakley_eliciting_2010}{2010}), or that can be carried out through the use of existing software like MATCH. Therefore, we programmed our own software. All related materials for this study, including code and data, can be found on the Open Science Framework (OSF) webpage at \url{https://osf.io/wvujz}.
\hypertarget{five-step-method}{%
\section{Five-Step Method}\label{five-step-method}}
In this section we describe the technical details of the Five-Step Method which has been programmed in R (R Core Team, \protect\hyperlink{ref-r_core_team_r:_2017}{2017}\protect\hyperlink{ref-r_core_team_r:_2017}{b}) using the shiny package (Chang, Cheng, Allaire, Xie, \& McPherson, \protect\hyperlink{ref-R-shiny}{2019}).
\hypertarget{step-1}{%
\subsection{Step 1}\label{step-1}}
The first step of the Five-Step Method consists of a digitized version of the trial roulette, which can be seen in Figure \ref{fig:ch02fig1}. Instead of vertical ``bins'' a grid is used and the digital ``chips'' can be placed on the grid. Experts provide estimates for the expected minimum and maximum value of the parameter of interest, represented by the left and rightmost digital ``chips'' in the grid, based on which the range of the grid is determined. Thereafter they place additional ``chips'' in the grid. In specific, the input grid, denoted by \textbf{G}, is a matrix size 600 (columns) x 300 (rows) and cells are activated by the placement of a digital ``chips'' in the grid. The cells where a sticker is placed obtain a value of one, all other cells are set to non-available. A second matrix, denoted by \textbf{R} of the same dimensions is created in which all rows are equal and the columns are a sequence of numbers with equal intervals running from the reasonable lower to upper bound provided as input. We then create output matrix \textbf{O} which contains values from \textbf{R} activated by the placement of dots in \textbf{G} and after the deletion of all nonavailable values in \textbf{O}, the remaining values are stored in a vector.
\begin{figure}
{\centering \includegraphics[width=0.85\linewidth]{figures/chapter_2/Figure_1}
}
\caption{Shiny application for steps 1 and 2. On the left the input fields can be found for the reasonable lower and upper bound as minimum and maximum values. The input grid in which 'chips' can be placed is found on the lower right with the leftmost dot being the minimum value and the right most dot being the maximum value. Further 'chips' are placed by clicking the mouse drawing a maximum of 11 pixels left and right. On the top right feedback is provided, presenting the fitted distribution based on the input.}\label{fig:ch02fig1}
\end{figure}
\hypertarget{step-2}{%
\subsection{Step 2}\label{step-2}}
The vector of values that is elicited in step 1 are used to fit a SN distribution. The SN distribution is defined in this paper as a normal distribution with the additional shape parameter \(\gamma\). The shape parameter is based upon a general method for the transformation of symmetric distributions into skewed distributions as described in Fernández \& Steel (\protect\hyperlink{ref-fernandez_bayesian_1998}{1998}). The transformation of the symmetric distribution into a skewed distribution is done by allocating mass of the distribution to either side of the mode (\emph{M}) by controlling the error term (\(\epsilon\)) via the following function, taken from Fernandez and Steel Eq. 1:
\begin{equation}
p(\epsilon|\gamma) = \frac{2}{{\gamma + \frac{1}{\gamma}}} {f(\frac{\epsilon}{\gamma})I_{(M,\infty)}(\epsilon) + f(\gamma \epsilon)I_{(-\infty,M)}(\epsilon)}.
\label{eq:ch02eq1}
\end{equation}
The effect of the shape parameter on the allocation of mass can be seen in Figure \ref{fig:ch02fig2}. Note that the distributions would be exactly mirrored with respect to the mode if the \(\gamma\) values would be \(\frac{1}{\gamma}\).
To fit the SN distribution we make use of the snormFit function from the fGarch package (Wuertz et al., \protect\hyperlink{ref-R-fGarch}{2019}). This function uses an optimization algorithm to determine the optimal skewness parameter based on log-likelihood values. The mean and standard deviation are determined based on the vector of elicited values. The mean and standard deviation remain constant and thus there is only one parameter to optimize over, the shape parameter \(\gamma\).
The SN distribution that is fitted based upon the expert's input is provided as visual feedback to the expert, see Figure \ref{fig:ch02fig1}. The visual feedback indicates how we interpret the information that is provided by the expert. The expert can accept the representation of their beliefs or adjust input until the representation matches their beliefs. Once the expert approves the representation of their beliefs, the mean value is extracted from the distribution which is to be used in step 3.
\begin{figure}
{\centering \includegraphics[width=0.85\linewidth]{figures/chapter_2/Figure_2}
}
\caption{Example of the influence of shape parameter $\gamma$ on the allocation of mass for a normal distribution with a variance of 1.}\label{fig:ch02fig2}
\end{figure}
\hypertarget{step-3}{%
\subsection{Step 3}\label{step-3}}
Step 3 of the Five-Step Method is used to derive the distributional representation of the expert's prior beliefs concerning the parameter of interest and can be seen in Figure \ref{fig:ch02fig3}. We restricted the priors that represent the experts' beliefs to be SN distributions so \(\pi_d \sim SN(\mu_0, \sigma^2_0, \gamma_0)\), where subscript \emph{d} denotes expert \(d = 1,...,D\), \(\mu_0\) denotes the prior mean, \(\sigma^2_0\) denotes the prior variance, and \(\gamma_0\) denotes the prior skewness. The value for \(\mu_0\) is assumed to be known, either obtained through steps 1 and 2 or stated directly. In step 3 the expert is required to provide values for the reasonable lower and upper bounds they perceive as likely for their estimate of \(\mu_0\). The value for \(\mu_0\) is repeated 100 times, the values for the reasonable lower and upper bounds for the estimate are both repeated 10 times.
\begin{figure}
{\centering \includegraphics[width=0.9\linewidth]{figures/chapter_2/Figure_3}
}
\caption{Shiny application for steps 3 and 4. On the left the input fields require entering the estimate for $\mu_0$ and the reasonable lower and upper bound for the estimate. On the right the distribution that is fitted based on the input can be found.}\label{fig:ch02fig3}
\end{figure}
\hypertarget{step-4}{%
\subsection{Step 4}\label{step-4}}
Based on the input provided in step 3 we will obtain estimates for the scale parameter \(\sigma^2_0\) and the shape parameter \(\gamma_0\). The 120 values, \(\mu_0\) repeated 100 times and the values for the reasonable lower and upper bounds both repeated 10 times, are provided to the snormFit function by means of which a SN distribution is fitted. The estimates for \(\sigma^2_0\) and \(\gamma_0\) are obtained and \(\mu_0\) is constrained to the input value. Visual feedback is provided to the expert of the resulting SN distribution, which can be seen in Figure \ref{fig:ch02fig3}. The expert can accept the representation of their beliefs or adjust input until the representation matches their beliefs.
\hypertarget{step-5}{%
\subsection{Step 5}\label{step-5}}
Use the elicited distribution that represents the expert's beliefs.
\hypertarget{elicitation-studies}{%
\section{Elicitation Studies}\label{elicitation-studies}}
In this section we describe the three studies we conducted. During the user feasibility study R version 3.1.2 was used and R version 3.2.3 was used during the elicitations done with the staffing company and the large financial institution. We conducted the elicitations in a semi-structured face-toface setting so that the researcher could provide interpretations accompanying the visual feedback. An advantage of a face-to-face setting is that it allows clarification of procedural and elicitation related questions thereby improving the validity of the responses (O'Hagan et al., \protect\hyperlink{ref-ohagan_uncertain_2006}{2006}).
(Cooke \& Goossens, \protect\hyperlink{ref-cooke_procedures_1999}{1999}) describe that a panel of four experts can be sufficient for an elicitation, but they recommend a panel of about eight experts as a rule of thumb. In the user feasibility study nine respondents participated. In the staffing company only four experts were available in the entire company, therefore the sample was limited to a size of four. Regarding the study at the large financial institution four experts participated in the end.
\hypertarget{user-feasibility-study}{%
\subsection{User Feasibility Study}\label{user-feasibility-study}}
\hypertarget{design}{%
\subsubsection{Design}\label{design}}
With the user feasibility study we evaluated the usability of the first two steps of the Five-Step Method. Procedural and visual preferences were investigated. Four variations of the shiny application were tested. The respondents (\(D=9\)), obtained through convenience sampling from a population of university trained adults, were randomly allocated to two out of the four possible variations of the software.
In the first procedural option, we used the procedure of the trial roulette where 20 digital ``chips'', starting with the expected minimal and maximum value, each representing five percent of a distribution, were to be placed in a grid following the procedure described by Zondervan-Zwijnenburg et al. (\protect\hyperlink{ref-zondervan-zwijnenburg_application_2017}{2017}\protect\hyperlink{ref-zondervan-zwijnenburg_application_2017}{b}). After placing 20 ``chips'' the respondents could submit their input and they were provided with visual feedback on the distribution that was fitted based on these 20 ``chips''. They could accept the representation or adjust their input. The second procedural option required the placement of a minimum of seven ``chips'', starting with the expected minimal and maximum value. In this procedural variation the distribution that was fitted based on the input was constantly shown. The distribution changed with each placed ``chip'' and thus instant feedback was provided on the representation of the input. Respondents could, after placing a minimum of seven ``chips'', at each point accept the representation of their beliefs or add or adjust input. Next to these two options, we also varied the size of the digital grid in which the ``chips''" were placed: large and small.
The respondents evaluated the two variations they were appointed to with a questionnaire asking if the fitted
distribution was a good reflection of their beliefs and what visual and procedural preferences were. Additional questions were based on the taxonomy of Bloom, Engelhart, Furst, Hill, \& Krathwohl (\protect\hyperlink{ref-bloom_taxonomy_1956}{1956}) to identify weak points of the software and procedures. These questions investigated; the comprehension of the instructions, the ability to apply the tool, the understanding of the representation of the ``chips''", the relation between input and fitted distribution, and the relation between belief and fitted distribution. The full questionnaire can be found in
the data archive which is available on the OSF webpage at \url{https://osf.io/wvujz}.
\hypertarget{results}{%
\subsubsection{Results}\label{results}}
All respondents indicated that their beliefs where accurately represented. Five of the seven respondents allocated to both procedural variants preferred the second variation. Four of the six respondents allocated to both visual variants preferred the large grid, one abstained from answering. Three out of the nine respondents indicated for at least one of the variations that they did not understand the meaning of the ``chips''. In
the first procedural option the ``chips'' each represented 5\% of the data whilst in the second procedural option the meaning depended on the amount of chips that were placed. They allocated mass for the distribution that was fitted. The meaning of the chips was not completely understood by one person who used the first procedure and by two persons who used the second procedure. All three of them used a small grid variation. The three respondents all indicated that they knew what the distribution representing their opinion meant in the end and agreed that this accurately described their view. Based on the results we decided to continue working with the second procedural variation, requiring the minimal placement of seven ``chips without further restriction on the number of''chips", and a large grid.
\hypertarget{elicitation-staffing-company}{%
\subsection{Elicitation Staffing Company}\label{elicitation-staffing-company}}
\hypertarget{design-1}{%
\subsubsection{Design}\label{design-1}}
The goal of the second study was to test the internal validity of elicitations obtained with the first two steps of the Five-Step Method. We found a staffing company willing to participate with experts (\(D=4\)) providing predictions about five sales results concerning the first quarter of 2016: contract hours, hourly cost buying and selling, turnover and hourly sales margin. A staffing company is a link between companies that want to hire staff and staff looking to work at companies. They buy work from individuals and thereafter place them to work at other companies. The amount of hours they place an individual at another company in the quarter are the contract hours. The hourly cost buying is what it will cost them per hour to buy the work from the individuals and the hourly cost selling is the price which they charge the companies where they stall the individuals. The turnover is equal to the contract hours multiplied by the hourly cost selling and the hourly sales margin is equal to the hourly cost selling minus the hourly cost buying.
The experts were asked to predict the distribution of the data. In some sectors staffing companies staff a lot of individuals at low margins and thus generate a large turnover. In different sectors they staff few individuals at high margins thereby obtaining the same profit at lower turnover rates. These are all relevant considerations and the experts should know which is the case for their company. The company provided us with actual budgets they made which were indications of carefully constructed predictions. By comparing the predictions of the experts to the budget we could gain an indication of the internal validity of predictions made with the first two steps of the Five-Step Method. If the elicitation results match the budget this indicates that the procedure is able to represent the underlying construct of carefully constructed predictions.
\hypertarget{results-1}{%
\subsubsection{Results}\label{results-1}}
The results can be found in Figure \ref{fig:ch02fig4} in which we plotted the predictions of the four experts against the actual budgets for the first quarter. To conceal the true values, which is businesssensitive information, a linear transformation has been done on all variables. It can be seen, especially for the hourly sales margins and the turnover, that experts provided very similar predictions to the budgets, for more detailed information see Table \ref{tab:ch02tab1}. The resemblance of the predictions to the budget indicates internal validity for the use of the steps 1 and 2 of the Five-Step Method as the elicited predictions closely match carefully constructed expectations. Based on these results we decided not to further adjust the elicitation procedure.
\begin{figure}
{\centering \includegraphics[width=0.75\linewidth]{figures/chapter_2/Figure_4}
}
\caption{Results for elicitation with the staffing company. Experts' predictions plotted with actual budget for contract hours, hourly cost buying and selling, turnover and hourly sales margin concerning the first quarter of 2016.}\label{fig:ch02fig4}
\end{figure}
\newpage
\begin{longtable}[]{@{}llccc@{}}
\caption{\label{tab:ch02tab1} Results for elicitation with the staffing company. Experts' predictions with actual budget for contract hours, hourly cost buying and selling, turnover and hourly sales margin concerning the first quarter of 2016.}\tabularnewline
\toprule
\begin{minipage}[b]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[b]{0.13\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[b]{0.10\columnwidth}\centering
\(\mu\)\strut
\end{minipage} & \begin{minipage}[b]{0.13\columnwidth}\centering
\(\sigma\)\strut
\end{minipage} & \begin{minipage}[b]{0.18\columnwidth}\centering
\(\gamma\)\strut
\end{minipage}\tabularnewline
\midrule
\endfirsthead
\toprule
\begin{minipage}[b]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[b]{0.13\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[b]{0.10\columnwidth}\centering
\(\mu\)\strut
\end{minipage} & \begin{minipage}[b]{0.13\columnwidth}\centering
\(\sigma\)\strut
\end{minipage} & \begin{minipage}[b]{0.18\columnwidth}\centering
\(\gamma\)\strut
\end{minipage}\tabularnewline
\midrule
\endhead
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\textbf{Contract}
\textbf{Hours}\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Expert 1\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
3.88\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
0.83\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
4.01*10-6\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Expert 2\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
3.56\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
0.61\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
4.48*10-8\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Expert 3\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
3.85\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
0.70\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
0.51\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Expert 4\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
3.34\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
0.89\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
0.74\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Budget\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
3.37\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
0.91\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
5.52*10-4\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\textbf{Hourly Cost}
\textbf{Buying}\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Expert 1\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
3.21\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
0.99\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
1.10\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Expert 2\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
2.74\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
0.69\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
1.57\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Expert 3\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
2.91\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
0.80\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
1.78\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Expert 4\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
3.45\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
0.97\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
7.20\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Budget\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
3.09\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
1.05\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
566.00\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\textbf{Hourly Cost}
\textbf{Selling}\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Expert 1\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
3.99\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
1.14\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
1.73\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Expert 2\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
3.86\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
1.04\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
2.14\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Expert 3\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
3.72\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
0.81\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
1.41\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Expert 4\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
4.59\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
1.43\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
12.80\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Budget\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
3.87\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
0.99\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
1.29\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\textbf{Turnover}\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Expert 1\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
3.39\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
0.98\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
1.48\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Expert 2\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
3.21\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
0.81\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
1.53\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Expert 3\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
2.71\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
0.72\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
0.93\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Expert 4\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
3.16\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
1.17\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
0.98\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Budget\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
2.71\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
0.99\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
0.76\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\textbf{Hourly}
\textbf{Sales}
\textbf{Margin}\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Expert 1\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
2.18\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
0.94\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
1.68\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Expert 2\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
2.46\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
0.97\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
1.31\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Expert 3\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
2.25\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
0.76\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
1.69\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Expert 4\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
2.18\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
0.84\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
1.97\strut
\end{minipage}\tabularnewline
\begin{minipage}[t]{0.30\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\raggedright
Budget\strut
\end{minipage} & \begin{minipage}[t]{0.10\columnwidth}\centering
2.06\strut
\end{minipage} & \begin{minipage}[t]{0.13\columnwidth}\centering
0.96\strut
\end{minipage} & \begin{minipage}[t]{0.18\columnwidth}\centering
1.51\strut
\end{minipage}\tabularnewline
\bottomrule
\end{longtable}
\newpage
\hypertarget{elicitation-large-financial-institution}{%
\subsection{Elicitation Large Financial Institution}\label{elicitation-large-financial-institution}}
In the third elicitation study the experts (\(D=4\)) were regional directors working at a large financial institution. They are considered experts in knowledge concerning market opportunities, market dynamics and estimating the capabilities of the professionals to seize opportunities. Based on these skills we expected that they could predict the average turnover per professional in the entire country in the first quarter of 2016. In this study the experts did not predict the distribution of the data \(\textbf{y}\), but construct a prior for the mean denoted by \(\pi_d(\theta)\). As \(\pi_d(\theta) \sim SN(\mu_0, \sigma^2_0, \gamma_0)\) the elicitation results in the representation of each expert's beliefs expressed in the hyper parameters \(\mu_0\), \(\sigma^2_0\) and \(\gamma_0\). We compare the predictions of the experts against actual results, expressed as the posterior distribution of the average turnover per professional, denoted by \(\pi(\theta|y)\). \(\pi(\theta|y) \sim SN(\mu_1, \sigma^2_1, \gamma_1)\), \(\mu_1\) denotes the posterior mean, \(\sigma^2_1\) denotes the posterior variance and \(\gamma1\) the posterior skewness. The prior for \(\pi(\theta|y)\) is a \(N(0,100)\) prior which is uninformative given the scale of the data.
\hypertarget{design-2}{%
\subsubsection{Design}\label{design-2}}
The team that participated consisted of 11 experts, 10 regional directors and one director. All were eligible to be included in the study. To comply with conditions set by the Ethics Committee, we ensured that experts whom did not wish to participate could do so without it being known that they refused. Therefore we randomly selected seven out of the 11 experts and invited them to participate. Out of the seven selected experts that we approached, three indicated that they did not want to participate in the study and four indicated that they were willing to participate. All four experts that agreed to participate, did participate and completed the elicitation. The participating experts first performed a practice elicitation for their own sales team before moving on to their estimate for the whole country, enabling them to acquaint themselves with the elicitation applications. Offering this practice elicitation could improve the quality of the elicitations (Johnson et al., \protect\hyperlink{ref-johnson_methods_2010}{2010}). Only in the case that the director participated the practice run was be possible. The study receive ethical approval from our internal Ethics Committee of the Faculty of Social and Behavioural Sciences of Utrecht University. The letter of approval can be found in the data archive on the OSF website at \url{https://osf.io/wvujz}.
The Five-Step Method was used in this elicitation study and it consists of the following two parts: the first step is designed to support the expert in the use of reasoned and structured thoughts to obtain an estimate for the location parameter \(\mu_0\). In the second step the estimate for \(\mu_0\) is used and the expert is asked to provide a reasonable lower and upper bound for their estimate so the prior distribution for the mean turnover per professional can be constructed.
The ``chips'' placed in the first step were intended to represent individual professionals in the trial run and
clusters of similar professionals in the elicitation concerning the whole country. Visual feedback was provided on the elicited distribution, accompanied by a description of the value for \(\mu_0\) by the researcher. The expert could accept the representation of their beliefs or adjust input until the representation matched their beliefs. Results concerning country wide performance where discussed in terms of total turnover for all professionals within the team, therefore the estimate for \(\mu_0\) was transformed using the following function
\begin{equation}
\theta^* = a\theta + b,
\label{eq:ch02eq2}
\end{equation}
where \(\theta\) represents the parameter of interest and \(\theta \sim N(\mu, \sigma^2)\) so that \(\theta^* \sim N[a\mu + b, (a\sigma)^2]\).
The use of the mean as location parameter offered additional options to accommodate differences in reasoning of experts, e.g., a sales expert might feel comfortable to provide estimates for the total turnover of a store, represented by \(\theta^*\) in Eq. \eqref{eq:ch02eq2}, but not be comfortable providing estimates for the mean turnover per product sold in the store, represented by \(\theta\) in Eq. \eqref{eq:ch02eq2}. By knowing the total amount of products that are sold in the store, entering the amount as value for a and 0 for b in Eq. \eqref{eq:ch02eq2}, the prior beliefs regarding the total turnover can be transformed to prior beliefs regarding mean turnover per product and compared to predictions by other experts. The transformation procedure ensures no expert is forced to adhere to a certain scale. To illustrate this flexibility let us imagine that a store sells nine different types of products and in total sells 104 products. In steps 1 and 2 we wish to elicit and verify the location parameter for the mean turnover. Two experts feel comfortable supplying estimates for turnover per product whilst two other experts only feel comfortable supplying estimates for turnover per product type. They can both adhere to the scale they feel comfortable with as we can use a linear transformation to get them onto the same scale for steps 3 and 4. In Table \ref{tab:ch02tab2} we supply a numerical example to show how location parameters, elicited on a different scale, can be transformed using Eq. \eqref{eq:ch02eq2} to be on the same scale for steps three and four of the elicitation.
In step 3 of the Five-Step Method, we asked the experts to provide a reasonable lower and upper bound for the total turnover of all professionals: Based on the input a distribution was fitted and visual feedback was provided. The researcher supported the visual feedback with a description explaining that more density on places of the axis indicate more perceived likeliness for that value. The expert could accept the representation of their beliefs or adjust the input for the reasonable lower and upper bound until the representation matched their beliefs. The elicited distribution was transformed back to represent the average turnover per professional using Eq. \eqref{eq:ch02eq2}.
\begin{longtable}[]{@{}lcccc@{}}
\caption{\label{tab:ch02tab2} Illustration of linear transformations using Eq. \eqref{eq:ch02eq2}. Experts 1 and 2 choose to specify turnover per product, resulting in a location parameter on the product scale. Experts 3 and 4 choose to specify turnover per product type, resulting in a location parameter on the product type scale. In steps 3 and 4 we can use either the location parameter on the total turnover scale or the mean turnover scale for experts to provide a reasonable lower and upper bound. All experts' elicited location parameters can be transformed to both scales.}\tabularnewline
\toprule
\begin{minipage}[b]{0.16\columnwidth}\raggedright
\strut
\end{minipage} & \begin{minipage}[b]{0.16\columnwidth}\centering
Steps 1 and 2
product scale
mean result
\((n=104)\)\strut
\end{minipage} & \begin{minipage}[b]{0.17\columnwidth}\centering
Steps 1 and 2
product type