Learning Bayesian Statistics

Proudly sponsored by PyMC Labs, the Bayesian Consultancy. Book a call, or get in touch!

In this episode, Jonathan Templin, Professor of Psychological and Quantitative Foundations at the University of Iowa, shares insights into his journey in the world of psychometrics.

Jonathan’s research focuses on diagnostic classification models — psychometric models that seek to provide multiple reliable scores from educational and psychological assessments. He also studies Bayesian statistics, as applied in psychometrics, broadly. So, naturally, we discuss the significance of psychometrics in psychological sciences, and how Bayesian methods are helpful in this field.

We also talk about challenges in choosing appropriate prior distributions, best practices for model comparison, and how you can use the Multivariate Normal distribution to infer the correlations between the predictors of your linear regressions.

This is a deep-reaching conversation that concludes with the future of Bayesian statistics in psychological, educational, and social sciences — hope you’ll enjoy it!

Our theme music is « Good Bayesian », by Baba Brinkman (feat MC Lars and Mega Ran). Check out his awesome work at https://bababrinkman.com/ !

Thank you to my Patrons for making this episode possible!

Yusuke Saito, Avi Bryant, Ero Carrera, Giuliano Cruz, Tim Gasser, James Wade, Tradd Salvo, William Benton, James Ahloy, Robin Taylor,, Chad Scherrer, Zwelithini Tunyiswa, Bertrand Wilden, James Thompson, Stephen Oates, Gian Luca Di Tanna, Jack Wells, Matthew Maldonado, Ian Costley, Ally Salim, Larry Gill, Ian Moran, Paul Oreto, Colin Caprani, Colin Carroll, Nathaniel Burbank, Michael Osthege, Rémi Louf, Clive Edelsten, Henri Wallen, Hugo Botha, Vinh Nguyen, Marcin Elantkowski, Adam C. Smith, Will Kurt, Andrew Moskowitz, Hector Munoz, Marco Gorelli, Simon Kessell, Bradley Rode, Patrick Kelley, Rick Anderson, Casper de Bruin, Philippe Labonde, Michael Hankin, Cameron Smith, Tomáš Frýda, Ryan Wesslen, Andreas Netti, Riley King, Yoshiyuki Hamajima, Sven De Maeyer, Michael DeCrescenzo, Fergal M, Mason Yahr, Naoya Kanai, Steven Rowland, Aubrey Clayton, Jeannine Sue, Omri Har Shemesh, Scott Anthony Robson, Robert Yolken, Or Duek, Pavel Dusek, Paul Cox, Andreas Kröpelin, Raphaël R, Nicolas Rode, Gabriel Stechschulte, Arkady, Kurt TeKolste, Gergely Juhasz, Marcus Nölke, Maggi Mackintosh, Grant Pezzolesi, Avram Aelony, Joshua Meehl, Javier Sabio, Kristian Higgins, Alex Jones, Gregorio Aguilar, Matt Rosinski, Bart Trudeau, Luis Fonseca and Dante Gates.

Visit https://www.patreon.com/learnbayesstats to unlock exclusive Bayesian swag 😉

Links from the show:

Abstract

by Christoph Bamberg

You have probably unknowingly already been exposed to this episode’s topic – psychometric testing – when taking a test at school or university. Our guest, Professor Jonathan Templin, tries to increase the meaningfulness of these tests by improving the underlying psychometric models, the bayesian way of course!

Jonathan explains that it is not easy to judge the ability of a student based on exams since they have errors and are only a snapshot. Bayesian statistics helps by naturally propagating this uncertainty to the results.

In the field of psychometric testing, Marginal Maximum Likelihood is commonly used. This approach quickly becomes unfeasible though when trying to marginalise over multidimensional test scores. Luckily, Bayesian probabilistic sampling does not suffer from this.

A further reason to prefer Bayesian statistics is that it provides a lot of information in the posterior. Imagine taking a test that tells you what profession you should pursue at the end of high school. The field with the best fit is of course interesting, but the second best fit may be as well. The posterior distribution can provide this kind of information.

After becoming convinced that Bayes is the right choice for psychometrics, we also talk about practical challenges like choosing a prior for the covariance in a multivariate normal distribution, model selection procedures and more.

In the end we learn about a great Bayesian holiday destination, so make sure to listen till the end!

Transcript

This is an automatic transcript and may therefore contain errors. Please get in touch if you’re willing to correct them.

Transcript
Speaker:

In this episode, Jonathan Templin,

professor of Psychological and

2

00:00:08,034 --> 00:00:12,936

Quantitative Foundations at the University

of Iowa, shares insight into his journey

3

00:00:12,936 --> 00:00:14,517

in the world of psychometrics.

4

00:00:14,537 --> 00:00:20,401

Jonathan's research focuses on diagnostic

classification models, psychometric models

5

00:00:20,401 --> 00:00:24,683

that seek to provide multiple reliable

scores from educational and psychological

6

00:00:24,683 --> 00:00:25,663

assessment.

7

00:00:25,683 --> 00:00:29,962

He also studies patient statistics as

applied in psychometrics, broadly.

8

00:00:29,962 --> 00:00:34,185

So naturally, we discussed the

significance of psychometrics in

9

00:00:34,185 --> 00:00:38,989

psychological sciences and how Bayesian

methods are helpful in this field.

10

00:00:38,989 --> 00:00:42,972

We also talked about challenges in

choosing appropriate prior distributions,

11

00:00:42,972 --> 00:00:46,815

best practices for model comparison, and

how you can use the multivariate normal

12

00:00:46,815 --> 00:00:51,098

distribution to infer the correlations

between the predictors of your linear

13

00:00:51,098 --> 00:00:52,159

progressions.

14

00:00:52,379 --> 00:00:55,522

This is a deep, reaching conversation that

15

00:00:55,522 --> 00:00:59,625

concludes with the future of Bayesian

statistics in Psychological, Educational,

16

00:00:59,625 --> 00:01:01,206

and Social Sciences.

17

00:01:01,306 --> 00:01:02,887

Hope you'll enjoy it.

18

00:01:02,888 --> 00:01:08,752

This is Learning Bayesian Statistics,

episode 94, recorded September 11, 2023.

19

00:01:12,856 --> 00:01:16,859

Hello, my dear Bayesians!

20

00:01:16,959 --> 00:01:21,383

This time, I have the pleasure to welcome

three new members to our Bayesian crew,

21

00:01:21,383 --> 00:01:25,005

Bart Trudeau, Noes Fonseca, and Dante

Gates.

22

00:01:25,154 --> 00:01:27,134

Thank you so much for your support, folks.

23

00:01:27,134 --> 00:01:29,535

It's the main way this podcast gets

funded.

24

00:01:29,655 --> 00:01:35,518

And Bart and Dante, get ready to receive

your exclusive merch in the coming month.

25

00:01:35,738 --> 00:01:37,458

Send me a picture, of course.

26

00:01:37,679 --> 00:01:42,220

Now let's talk psychometrics and modeling

with Jonathan Templin.

27

00:01:43,441 --> 00:01:47,283

Jonathan Templin, welcome to learning

patient statistics.

28

00:01:47,283 --> 00:01:49,324

Thank you for having me.

29

00:01:49,324 --> 00:01:50,604

It's a pleasure to be here.

30

00:01:54,474 --> 00:01:56,314

Yeah, thanks a lot.

31

00:01:56,935 --> 00:02:01,977

Quite a few patrons have mentioned you in

the Slack of the show.

32

00:02:01,977 --> 00:02:08,879

So I'm very honored to honor their request

and have you on the show.

33

00:02:09,540 --> 00:02:15,042

And actually thank you folks for bringing

me all of those suggestions and allowing

34

00:02:15,042 --> 00:02:20,444

me to discover so many good patients out

there in the world doing awesome things in

35

00:02:20,444 --> 00:02:24,025

a lot of different fields using our.

36

00:02:24,242 --> 00:02:27,602

favorite tools to all of us based in

statistics.

37

00:02:28,143 --> 00:02:34,925

So Jonathan, before talking about all of

those good things, let's dive into your

38

00:02:34,925 --> 00:02:36,245

origin story.

39

00:02:36,245 --> 00:02:40,746

How did you come to the world of

psychometrics and psychological sciences

40

00:02:40,746 --> 00:02:43,407

and how sinuous of a path was it?

41

00:02:43,407 --> 00:02:47,288

That's a good question.

42

00:02:47,288 --> 00:02:48,609

So I was an odd student, I dropped out of

high school.

43

00:02:48,609 --> 00:02:49,429

So I started my...

44

00:02:52,142 --> 00:02:54,423

college degree and community college, that

would be the only place that would take

45

00:02:54,423 --> 00:02:55,023

me.

46

00:02:55,283 --> 00:02:58,225

I happened to be really lucky to do that

though, because I had some really great

47

00:02:58,225 --> 00:03:03,888

professors and I took a, once I discovered

that I probably could do school, I took a

48

00:03:03,888 --> 00:03:07,690

statistics course, you know, typical

undergraduate basic statistics.

49

00:03:07,790 --> 00:03:09,091

I found that I loved it.

50

00:03:09,331 --> 00:03:12,993

I decided that I wanted to do something

with statistics and then in the process, I

51

00:03:12,993 --> 00:03:19,237

took a research methods class in

psychology and I decided somehow I wanted

52

00:03:19,237 --> 00:03:21,197

to do statistics in psychology.

53

00:03:21,470 --> 00:03:25,991

So moved on from community college, went

to my undergraduate for two years at

54

00:03:25,991 --> 00:03:29,473

Sacramento state and Sacramento,

California also was really lucky because I

55

00:03:29,473 --> 00:03:32,014

had professor there that said, Hey,

there's this field called quantitative

56

00:03:32,014 --> 00:03:33,154

psychology.

57

00:03:33,334 --> 00:03:34,415

You should look into it.

58

00:03:34,415 --> 00:03:38,136

If you're interested in statistics and

psychology along the same time, he was

59

00:03:38,136 --> 00:03:40,357

teaching me something called factor

analysis.

60

00:03:40,357 --> 00:03:43,539

I now look at it as more principal

components analysis, but I wanted to know

61

00:03:43,539 --> 00:03:46,520

what was happening underneath the hood of

factor analysis.

62

00:03:46,520 --> 00:03:49,921

And so that's where he said, no, really,

you should go to the graduate school for

63

00:03:49,921 --> 00:03:50,621

that.

64

00:03:50,874 --> 00:03:53,036

And so that's what started me.

65

00:03:53,036 --> 00:03:57,920

I was fortunate enough to be able to go to

the University of Illinois for graduate

66

00:03:57,920 --> 00:03:58,801

studies.

67

00:03:58,901 --> 00:04:02,965

I did a master's, a PhD there, and in the

process, that's where I learned all about

68

00:04:02,965 --> 00:04:03,465

Bates.

69

00:04:03,465 --> 00:04:07,849

So it was a really lucky route, but it all

wouldn't have happened if I didn't go to

70

00:04:07,849 --> 00:04:13,174

community college, so I'm really proud to

say I'm a community college graduate, if

71

00:04:13,174 --> 00:04:13,414

you will.

72

00:04:13,414 --> 00:04:13,594

Yeah.

73

00:04:13,594 --> 00:04:14,475

Nice.

74

00:04:15,736 --> 00:04:15,936

Yeah.

75

00:04:15,936 --> 00:04:17,357

So it kind of happened.

76

00:04:18,378 --> 00:04:20,378

somewhat easily in a way, right?

77

00:04:20,378 --> 00:04:23,579

Good meeting at the right time and boom.

78

00:04:23,579 --> 00:04:25,600

That's right.

79

00:04:25,600 --> 00:04:30,282

And the call of the eigenvalue is what

really sent me to graduate school.

80

00:04:30,282 --> 00:04:32,342

So I wanted to figure out what that was

about.

81

00:04:34,023 --> 00:04:37,004

Yes, that is a good point.

82

00:04:38,444 --> 00:04:41,145

And so nowadays,

83

00:04:41,994 --> 00:04:42,694

What are you doing?

84

00:04:42,694 --> 00:04:46,815

How would you define the work you're doing

and what are the topics that you are

85

00:04:46,815 --> 00:04:48,575

particularly interested in?

86

00:04:49,496 --> 00:04:55,597

I would put my work into the field of item

response theory, largely.

87

00:04:55,737 --> 00:04:57,738

I do a lot of multidimensional item

response theory.

88

00:04:57,738 --> 00:05:02,679

There are derivative fields I think I'm

probably most known for, one of which is

89

00:05:02,780 --> 00:05:06,361

something called cognitive diagnosis or

diagnostic classification modeling.

90

00:05:06,361 --> 00:05:09,221

Basically, it's a classification based

method to try to...

91

00:05:09,934 --> 00:05:14,436

Classify students, or I work in the

College of Education, so most of this is

92

00:05:14,436 --> 00:05:20,139

applied to educational data from

assessments, and our goal is to, whenever

93

00:05:20,139 --> 00:05:23,461

you take a test, not just give you one

score, give you multiple valid scores, try

94

00:05:23,461 --> 00:05:26,082

to maximize the information we can give

you.

95

00:05:26,583 --> 00:05:30,805

My particular focus these days is in doing

so in classroom-based assessments, so how

96

00:05:30,805 --> 00:05:34,447

do we understand what a student knows at a

given point in the academic year and try

97

00:05:34,447 --> 00:05:37,629

to help make sure that they make the most

progress they can.

98

00:05:38,209 --> 00:05:38,542

Not.

99

00:05:38,542 --> 00:05:42,084

to remove the impact of the teacher

actually to provide the teacher with the

100

00:05:42,084 --> 00:05:45,807

best data to work with the child, to work

with the parents, to try to move forward.

101

00:05:45,807 --> 00:05:50,491

But all that boils down to interesting

measurements, psychometric issues, and

102

00:05:50,491 --> 00:05:55,555

interesting ways that we look at test data

that come out of classrooms.

103

00:05:55,555 --> 00:05:56,275

Okay.

104

00:05:56,275 --> 00:05:57,796

Yeah, that sounds fascinating.

105

00:05:58,697 --> 00:06:03,621

Basically trying to give a distribution of

results instead of just one point

106

00:06:03,621 --> 00:06:04,521

estimate.

107

00:06:04,814 --> 00:06:07,554

That's it also and tests have a lot of

error.

108

00:06:07,554 --> 00:06:11,635

So making sure that we don't over deliver

when we have a test score.

109

00:06:11,675 --> 00:06:15,556

Basically understanding what that is and

accurately quantifying how much

110

00:06:15,817 --> 00:06:20,998

measurement error is or lack of

reliability there is in the score itself.

111

00:06:20,998 --> 00:06:22,698

Yeah, that's fascinating.

112

00:06:24,019 --> 00:06:25,839

I mean, we can already dive into that.

113

00:06:25,839 --> 00:06:29,040

I have a lot of questions for you, but it

sounds very interesting.

114

00:06:29,040 --> 00:06:31,581

So yeah.

115

00:06:31,581 --> 00:06:34,241

So what does it look like concretely?

116

00:06:35,410 --> 00:06:42,732

these measurement errors and the test

scores attached to them, and basically how

117

00:06:42,732 --> 00:06:44,392

do you try to solve that?

118

00:06:44,392 --> 00:06:49,993

Maybe you can take an example from your

work where you are trying to do that.

119

00:06:50,354 --> 00:06:51,254

Absolutely.

120

00:06:51,254 --> 00:06:52,514

Let me start with the classical example.

121

00:06:52,514 --> 00:06:54,235

If this is too much information, I

apologize.

122

00:06:54,235 --> 00:07:00,997

But to set the stage, for a long time in

item response theory, we understand that a

123

00:07:00,997 --> 00:07:01,957

person's...

124

00:07:04,642 --> 00:07:07,324

Latentability estimate, if you want to

call it that, is applied in education.

125

00:07:07,324 --> 00:07:11,627

So this latent variable that represents

what a person knows, it's put onto the

126

00:07:11,627 --> 00:07:13,048

continuum where items are.

127

00:07:13,048 --> 00:07:15,729

So basically items and people are sort of

ordered.

128

00:07:16,831 --> 00:07:21,954

However, the properties of the model are

such that how much error there might be in

129

00:07:21,954 --> 00:07:25,637

a person's point estimate of their score

depends on where the score is located on

130

00:07:25,637 --> 00:07:26,458

the continuum.

131

00:07:26,458 --> 00:07:30,879

So this is what, you know, theory gave

rise to, you know, theory in the 1970s

132

00:07:30,879 --> 00:07:34,320

gave rise to our modern computerized

adaptive assessments and so forth, that

133

00:07:34,320 --> 00:07:40,161

sort of pick an item that would minimize

the error, if you will, different ways of

134

00:07:40,161 --> 00:07:42,062

describing what we pick an item for.

135

00:07:42,062 --> 00:07:44,122

But that's basically the idea.

136

00:07:44,923 --> 00:07:51,384

And so from a perspective of where I'm at

with what I do, a complicating factor in

137

00:07:51,384 --> 00:07:55,185

this, so that architecture that I just

mentioned that

138

00:07:55,870 --> 00:08:00,533

historic version of adaptive assessments

that really been built on large scale

139

00:08:00,533 --> 00:08:00,934

measures.

140

00:08:00,934 --> 00:08:05,337

So thousands of students and really what

happens in a classical census you would

141

00:08:05,337 --> 00:08:08,960

take a marginal maximum likelihood

estimate of certain parameter values from

142

00:08:08,960 --> 00:08:09,861

the model.

143

00:08:10,381 --> 00:08:14,905

You'd fix those values as if you knew them

with certainty and then you would go and

144

00:08:14,905 --> 00:08:18,628

estimate a person's parameter value along

with their standard error conditional

145

00:08:18,628 --> 00:08:20,129

standard error measurement.

146

00:08:21,170 --> 00:08:25,433

The situations I work in don't have large

sample size but we all in addition to

147

00:08:25,926 --> 00:08:31,568

a problem with sort of the asthmatotic

convergence, if you will, of those models,

148

00:08:32,309 --> 00:08:37,812

we also have a, not only we have not have

large sample sizes, we also have multiple,

149

00:08:38,152 --> 00:08:42,855

multiple scores effectively, multiple

latent freqs that we can't possibly do.

150

00:08:42,855 --> 00:08:50,419

So when you look at the same problem from

a Bayesian lens, sort of an interesting

151

00:08:50,419 --> 00:08:54,262

feature happens that we don't often see,

you know,

152

00:08:54,262 --> 00:08:59,304

frequentness or a classical framework in

that process of fixing the parameters of

153

00:08:59,304 --> 00:09:04,667

the model, the item parameters to a value,

you know, disregards any error in the

154

00:09:04,667 --> 00:09:05,348

estimate as well.

155

00:09:05,348 --> 00:09:09,530

Whereas if you're in a simultaneous

estimate, for instance, in a markup chain

156

00:09:09,530 --> 00:09:12,472

where you're sampling these values from a

posterior in addition to sampling

157

00:09:12,472 --> 00:09:16,734

students, it turns out those that error

around those parameters can propagate to

158

00:09:16,734 --> 00:09:20,376

the students and provide a wider interval

around them, which I think is a bit more

159

00:09:20,376 --> 00:09:23,017

accurate, particularly in smaller sample

size.

160

00:09:23,646 --> 00:09:24,126

situation.

161

00:09:24,126 --> 00:09:26,888

So I hope that's the answer to your

question.

162

00:09:26,888 --> 00:09:29,811

I may have taken a path that might have

been a little different there, but that's

163

00:09:29,811 --> 00:09:34,034

where I see the value at least in using

Bayesian statistics and what I do.

164

00:09:34,534 --> 00:09:36,255

Yeah, no, I love it.

165

00:09:36,616 --> 00:09:42,480

Don't shy away from technical explanation

on these podcasts.

166

00:09:43,362 --> 00:09:45,583

That's the good thing of the podcast.

167

00:09:45,683 --> 00:09:46,944

Don't have to shy away from it.

168

00:09:46,944 --> 00:09:47,865

It came at a good time.

169

00:09:47,865 --> 00:09:52,109

I've been working on this, some problems

like this all day, so I'm probably in the

170

00:09:52,109 --> 00:09:53,282

weeds a little bit.

171

00:09:53,282 --> 00:09:57,523

Forgive me if I go at the deep end of it.

172

00:09:57,523 --> 00:09:58,604

No, that's great.

173

00:09:58,784 --> 00:10:02,525

And we already mentioned item response

theory on the show.

174

00:10:02,525 --> 00:10:08,928

So hopefully people will refer back to

these episodes and that will give them a

175

00:10:08,928 --> 00:10:09,908

heads up.

176

00:10:11,229 --> 00:10:18,252

Well, actually you mentioned it, but do

you remember how you first got introduced

177

00:10:18,252 --> 00:10:21,933

to Bayesian methods and why did they stick

with you?

178

00:10:22,522 --> 00:10:24,002

Very, very much.

179

00:10:24,843 --> 00:10:28,645

I was introduced because in graduate

school, I had the opportunity to work for

180

00:10:28,645 --> 00:10:34,608

a lab run by Bill Stout at the University

of Illinois with other very notable people

181

00:10:34,608 --> 00:10:38,450

in my career, at least Jeff Douglas, Louis

Roussos, among others.

182

00:10:38,791 --> 00:10:41,532

And I was hired as a graduate research

assistant.

183

00:10:41,532 --> 00:10:46,715

And my job was to take a program that was

a metropolis Hastings algorithm and to

184

00:10:46,715 --> 00:10:48,636

make it run.

185

00:10:48,696 --> 00:10:49,717

And it was written in Fortran.

186

00:10:49,717 --> 00:10:50,738

So basically, I

187

00:10:50,738 --> 00:10:54,660

It was Metropolis Hastings, Bayesian, and

it was written in language that I didn't

188

00:10:54,660 --> 00:10:56,401

know with methods I didn't know.

189

00:10:56,401 --> 00:10:59,383

And so I was hired and said, yeah, figure

it out with good luck.

190

00:10:59,423 --> 00:11:02,665

Thankfully, I had colleagues that could

help actually probably figure it out more

191

00:11:02,665 --> 00:11:03,205

than I did.

192

00:11:03,205 --> 00:11:07,348

But I was very fortunate to be there

because it's like a trial by fire.

193

00:11:07,348 --> 00:11:09,729

I was basically going line by line through

that.

194

00:11:11,210 --> 00:11:15,693

This was a little bit in the later part

of, I think it was the year 2001, maybe a

195

00:11:15,693 --> 00:11:16,733

little early 2002.

196

00:11:17,334 --> 00:11:23,358

But something instrumental to me at the

time were a couple papers by a couple

197

00:11:23,358 --> 00:11:29,402

scholars in education at least, Rich Pates

and Brian Junker had a paper in 1999,

198

00:11:30,023 --> 00:11:34,266

actually two papers in 1999, I can even,

you know, it's like Journal of Educational

199

00:11:34,266 --> 00:11:35,007

Behavioral Statistics.

200

00:11:35,007 --> 00:11:36,868

It's like I have that memorized.

201

00:11:36,868 --> 00:11:40,331

But in their algorithm, they had written

down the algorithm itself and it was a

202

00:11:40,331 --> 00:11:43,953

matter of translating that to the

diagnostic models that we were working on.

203

00:11:44,422 --> 00:11:47,824

But that is why it stuck with me because

it was my job, but then it was also

204

00:11:47,824 --> 00:11:48,705

incredibly interesting.

205

00:11:48,705 --> 00:11:52,787

It was not like a lot of the research that

I was reading and not like a lot of the

206

00:11:52,787 --> 00:11:55,069

work I was doing in a lot of the classes I

was in.

207

00:11:55,069 --> 00:11:58,611

So I found it really mentally stimulating,

entirely challenging.

208

00:11:58,611 --> 00:12:00,993

It took the whole of my brain to figure

out.

209

00:12:00,993 --> 00:12:02,574

And even then I don't know that I figured

it out.

210

00:12:02,574 --> 00:12:08,278

So that helps answer that question.

211

00:12:08,278 --> 00:12:09,018

Yeah.

212

00:12:10,179 --> 00:12:14,026

So basically it sounds like you were

thrown into the Beijing pool.

213

00:12:14,026 --> 00:12:15,247

Like you didn't have any choice.

214

00:12:15,247 --> 00:12:17,008

I was.

215

00:12:17,088 --> 00:12:23,373

When I was Bayesian, it was nice because

at the time, you know, this is 2001, 2002,

216

00:12:24,934 --> 00:12:27,536

in education, no measurement in

psychology.

217

00:12:27,917 --> 00:12:30,959

You know, we knew of Bayes certainly, you

know, there's some great papers from the

218

00:12:30,959 --> 00:12:35,583

nineties that were around, but, you know,

we weren't, it wasn't prominent.

219

00:12:35,583 --> 00:12:38,265

It wasn't, you know, I was in graduate

school, but at the same time I wasn't

220

00:12:38,265 --> 00:12:42,329

learning it, I mean, I knew the textbook

Bayes, like the introductory Bayes, but

221

00:12:42,329 --> 00:12:43,789

not, definitely not.

222

00:12:43,946 --> 00:12:45,846

Like the estimation side.

223

00:12:46,086 --> 00:12:51,568

And so it was timing wise, you know,

people would look back now and say, okay,

224

00:12:51,568 --> 00:12:56,509

why didn't I go grab Stan or grab, at the

time I think we had, Jets didn't exist,

225

00:12:56,509 --> 00:12:57,829

there was bugs.

226

00:12:58,149 --> 00:13:02,010

And it was basically, you have to, you

know, like roll your own to do anything.

227

00:13:02,010 --> 00:13:04,151

So it was, it was good.

228

00:13:04,151 --> 00:13:05,111

No, for sure.

229

00:13:05,331 --> 00:13:11,253

Like, yeah, no, it's like telling, it's

like asking Christopher Columbus or

230

00:13:15,086 --> 00:13:15,606

That's right.

231

00:13:15,606 --> 00:13:16,467

It's a lot more direct.

232

00:13:16,467 --> 00:13:18,627

Just hop on the plane and...

233

00:13:18,627 --> 00:13:19,288

Wasn't an option.

234

00:13:19,288 --> 00:13:22,189

Exactly.

235

00:13:22,709 --> 00:13:23,469

Good point.

236

00:13:23,469 --> 00:13:26,571

But actually nowadays, what are you using?

237

00:13:26,571 --> 00:13:31,333

Are you still doing your own sampler like

that in Fortran or are you using some open

238

00:13:31,333 --> 00:13:32,933

source software?

239

00:13:33,374 --> 00:13:37,375

I can hopefully say I retired from Fortran

as much as possible.

240

00:13:38,836 --> 00:13:42,517

Most of what I do is install these days a

little bit of JAGS, but then occasionally

241

00:13:42,517 --> 00:13:43,257

I will...

242

00:13:43,342 --> 00:13:46,283

trying to write my own here or there.

243

00:13:46,904 --> 00:13:50,767

The latter part I'd love to do more of,

because you can get a little highly

244

00:13:50,767 --> 00:13:51,347

specialized.

245

00:13:51,347 --> 00:13:55,830

I just like that, I feel like the time to

really deeply do the development work in a

246

00:13:55,830 --> 00:14:00,213

way that doesn't just have an R package or

some package in Python that would just

247

00:14:00,213 --> 00:14:01,134

break all the time.

248

00:14:01,134 --> 00:14:05,297

So I'm sort of stuck right now with that,

but it is something that I'm grateful for

249

00:14:05,297 --> 00:14:08,919

having the contributions of others to be

able to rely upon to do estimation.

250

00:14:08,919 --> 00:14:09,840

Sorry.

251

00:14:09,840 --> 00:14:11,321

Yeah, no, exactly.

252

00:14:11,321 --> 00:14:12,121

I mean,

253

00:14:12,974 --> 00:14:16,235

So first, Stan, I've heard he's quite

good.

254

00:14:17,016 --> 00:14:18,577

Of course, it's amazing.

255

00:14:19,498 --> 00:14:23,581

A lot of Stan developers have been on this

show, and they do absolutely tremendous

256

00:14:23,581 --> 00:14:24,341

work.

257

00:14:24,462 --> 00:14:32,127

And yeah, as you were saying, why code

your own sampler when you can rely on

258

00:14:32,267 --> 00:14:38,632

samplers that are actually waterproof,

that are developed by a bunch of very

259

00:14:38,632 --> 00:14:41,293

smart people who do a lot of math.

260

00:14:41,690 --> 00:14:45,673

and who do all the heavy lifting for you,

well, just do that.

261

00:14:45,933 --> 00:14:49,816

And thanks to that, Bayesian computing and

statistics are much more accessible

262

00:14:49,816 --> 00:14:55,081

because you don't have to actually know

how to code your own MCMC sampler to do

263

00:14:55,081 --> 00:14:55,701

it.

264

00:14:55,902 --> 00:15:03,288

You can stand on the shoulders of giants

and just use that and superpower your own

265

00:15:03,288 --> 00:15:03,768

analysis.

266

00:15:03,768 --> 00:15:10,254

So it's definitely something we tell

people, don't code your own samplers now.

267

00:15:10,254 --> 00:15:13,196

You don't need to do that unless you

really, really have to do it.

268

00:15:13,196 --> 00:15:16,819

But usually, when you have to do that, you

know what you're doing.

269

00:15:17,460 --> 00:15:20,182

Otherwise, people have figured that out

for you.

270

00:15:20,182 --> 00:15:25,947

Just use the automatic samplers from Stan

or Pimsy or Numpyro or whatever you're

271

00:15:25,947 --> 00:15:26,647

using.

272

00:15:26,848 --> 00:15:35,295

It's usually extremely robust and checked

by a lot of different pairs of eyes and

273

00:15:35,295 --> 00:15:36,255

keyboards.

274

00:15:38,378 --> 00:15:43,039

having that team and like you said, full

of people who are experts in not only just

275

00:15:43,039 --> 00:15:45,940

mathematics, but also computer science

makes a big difference.

276

00:15:46,881 --> 00:15:47,381

Yeah.

277

00:15:47,381 --> 00:15:53,223

I mean, I would not be able to use patient

statistics nowadays if these samplers

278

00:15:53,264 --> 00:15:53,984

didn't exist, right?

279

00:15:53,984 --> 00:15:55,305

Because I'm not a mathematician.

280

00:15:55,305 --> 00:16:01,047

So if I had to write my own sample each

time, I would just be discouraged even

281

00:16:01,047 --> 00:16:02,187

before starting.

282

00:16:02,928 --> 00:16:03,688

Yeah.

283

00:16:04,288 --> 00:16:05,869

It's just a challenge in and of itself.

284

00:16:05,869 --> 00:16:07,569

I remember the old days where

285

00:16:07,606 --> 00:16:08,206

That would be it.

286

00:16:08,206 --> 00:16:09,246

That's my dissertation.

287

00:16:09,246 --> 00:16:10,667

That was what I had to do.

288

00:16:10,667 --> 00:16:13,308

So it was like six months work on just the

sampler.

289

00:16:13,349 --> 00:16:15,090

And even then it wasn't very good.

290

00:16:15,090 --> 00:16:17,731

And then they might actually do the

studying.

291

00:16:17,731 --> 00:16:23,054

Yeah, exactly.

292

00:16:24,795 --> 00:16:25,215

Yeah.

293

00:16:25,215 --> 00:16:32,440

I mean, to me really that probabilistic

programming is one of the super power of

294

00:16:32,440 --> 00:16:36,181

the Beijing community because that really

allows.

295

00:16:36,670 --> 00:16:44,592

almost anybody who can code in R or Python

or Julia to just use what's being done by

296

00:16:44,592 --> 00:16:47,833

very competent and smart people and for

free.

297

00:16:47,833 --> 00:16:47,973

Right.

298

00:16:47,973 --> 00:16:48,673

Yeah.

299

00:16:49,813 --> 00:16:50,373

Also true.

300

00:16:50,373 --> 00:16:50,574

Yeah.

301

00:16:50,574 --> 00:16:52,074

What a great community.

302

00:16:52,074 --> 00:16:55,495

I'm really, really impressed with the size

and the scope and how things have

303

00:16:55,495 --> 00:16:56,775

progressed in just 20 years.

304

00:16:56,775 --> 00:16:58,236

It's really something.

305

00:16:58,516 --> 00:16:59,216

Yeah.

306

00:16:59,316 --> 00:17:00,957

Exactly.

307

00:17:00,957 --> 00:17:02,857

And so actually...

308

00:17:04,898 --> 00:17:11,420

Do you know why, well, do you have an idea

why Bayesian statistics is useful in your

309

00:17:11,420 --> 00:17:12,101

field?

310

00:17:12,101 --> 00:17:17,323

What do they bring that you don't get with

the classical framework?

311

00:17:18,524 --> 00:17:25,006

Yeah, in particular, we have a really

nasty...

312

00:17:26,087 --> 00:17:31,529

If we were to do a classical framework,

typically the gold standard in...

313

00:17:31,830 --> 00:17:34,911

the field I work in is sort of a marginal

maximum likelihood.

314

00:17:34,911 --> 00:17:39,933

The marginal mean we get rid of the latent

variable to estimate models.

315

00:17:40,013 --> 00:17:44,255

So that process of marginalization is done

numerically.

316

00:17:44,255 --> 00:17:48,576

We numerically integrate across likelihood

function.

317

00:17:49,837 --> 00:17:53,798

Most cases, there are some special case

models that we really are too simplistic

318

00:17:53,798 --> 00:17:56,399

to use for what we do where we don't have

it.

319

00:17:57,560 --> 00:18:00,741

So if we want to do multidimensional

versions

320

00:18:01,886 --> 00:18:05,929

If you think about numeric integration,

for one dimension you have this sort of

321

00:18:06,330 --> 00:18:11,154

discretized set of a likelihood to take

sums across different, what we call

322

00:18:11,154 --> 00:18:14,296

quadrature points of some type of curve.

323

00:18:15,237 --> 00:18:20,863

For the multidimensional sense now, going

from one to two, you effectively squared

324

00:18:20,863 --> 00:18:22,064

the number of points you have.

325

00:18:22,064 --> 00:18:23,465

So that's just too latent variable.

326

00:18:23,465 --> 00:18:26,988

So if you want two bits of information

from an assessment from somebody, now

327

00:18:26,988 --> 00:18:28,949

you've just made your

328

00:18:29,542 --> 00:18:34,124

marginalization process exponentially more

difficult, more time-consuming.

329

00:18:34,124 --> 00:18:39,026

But really, the benefit of having two

scores is very little compared to having

330

00:18:39,026 --> 00:18:39,686

one.

331

00:18:39,846 --> 00:18:47,329

So if we wanted to do five or six or 300

scores, that marginalization process

332

00:18:47,329 --> 00:18:49,390

becomes really difficult.

333

00:18:49,790 --> 00:18:56,433

So from a brute force perspective, if we

take the a Bayesian sampler perspective,

334

00:18:57,078 --> 00:19:01,799

there is not the exponential increase of

computation in the linear increase in the

335

00:19:01,799 --> 00:19:02,600

latent variables.

336

00:19:02,600 --> 00:19:08,422

And so from a number of steps the process

has to take from calculation is much

337

00:19:08,422 --> 00:19:09,062

smaller.

338

00:19:09,062 --> 00:19:11,703

Now, of course, Markov chains have a lot

of calculations.

339

00:19:11,703 --> 00:19:16,906

So, you know, maybe overall the process is

longer, but it is, I found it to be

340

00:19:16,906 --> 00:19:22,068

necessity, basing statistics to estimate

in some form shows up in this

341

00:19:22,068 --> 00:19:25,669

multidimensional likelihood, basically

evaluation.

342

00:19:26,830 --> 00:19:31,091

created sort of hybrid versions of EM

algorithms where the E-step is replaced

343

00:19:31,091 --> 00:19:34,211

with the Bayesian type method.

344

00:19:34,632 --> 00:19:38,473

But for me, I like the full Bayesian

approach to everything.

345

00:19:38,473 --> 00:19:45,855

So I would say that just in summary

though, what Bayes brings from a brute

346

00:19:45,855 --> 00:19:49,516

force perspective is the ability to

estimate our models in a reasonable amount

347

00:19:49,516 --> 00:19:51,996

of time with a reasonable amount of

computations.

348

00:19:52,457 --> 00:19:55,274

There's the added benefit of what I

mentioned previously, which

349

00:19:55,274 --> 00:19:59,955

which is the small sample size, sort of

the, I think, a proper accounting or

350

00:19:59,955 --> 00:20:03,676

allowing of error to propagate in the

right way if you're going to report scores

351

00:20:03,676 --> 00:20:05,337

and so forth, I think that's an added

benefit.

352

00:20:05,337 --> 00:20:10,158

But from a primary perspective, I'm here

because I have a really tough integral to

353

00:20:10,158 --> 00:20:12,499

solve and Bayes helps me get around it.

354

00:20:13,859 --> 00:20:16,940

Yeah, that's a good point.

355

00:20:16,940 --> 00:20:21,081

And yeah, like as you were saying, I'm

guessing that having priors

356

00:20:21,122 --> 00:20:27,007

And generative modeling helps for low

sample sizes, which tends to be the case a

357

00:20:27,007 --> 00:20:28,688

lot in your field.

358

00:20:29,229 --> 00:20:29,769

Also true.

359

00:20:29,769 --> 00:20:30,270

Yeah.

360

00:20:30,270 --> 00:20:33,993

The prior distributions can help.

361

00:20:33,993 --> 00:20:40,118

A lot of the frustration with

multidimensional models and psychometrics,

362

00:20:40,459 --> 00:20:42,861

at least in practical sense.

363

00:20:42,861 --> 00:20:45,523

You get a set of data, you think it's

multidimensional.

364

00:20:45,523 --> 00:20:48,065

The next process is to estimate a model.

365

00:20:48,418 --> 00:20:51,580

in the classic sense that those models

sometimes would fail to converge.

366

00:20:51,580 --> 00:20:56,323

Uh, and very little reason why, um,

oftentimes it's failed to emerge.

367

00:20:56,323 --> 00:20:59,825

I had a class I taught four or five years

ago where I just asked people to estimate

368

00:20:59,825 --> 00:21:04,628

five dimensions and not a single person

couldn't could get, I had a set of data

369

00:21:04,628 --> 00:21:05,148

for each person.

370

00:21:05,148 --> 00:21:07,870

Not a single person could get it in

marriage with the default options that

371

00:21:07,870 --> 00:21:10,311

you'd see that like an IRT package.

372

00:21:10,591 --> 00:21:14,253

Um, so having the ability to sort of.

373

00:21:14,310 --> 00:21:17,791

Understand potentially where

non-convergence or why that's happening,

374

00:21:17,791 --> 00:21:20,111

which parameters are finding a difficult

spot.

375

00:21:20,111 --> 00:21:25,113

Then using priors to sort of aid an

estimation as one part, but then also sort

376

00:21:25,113 --> 00:21:27,533

of the idea of the Bayesian updating.

377

00:21:27,533 --> 00:21:31,374

If you're trying to understand what a

student knows throughout the year,

378

00:21:31,374 --> 00:21:33,475

Bayesian updating is perfect for such

things.

379

00:21:33,475 --> 00:21:36,876

You know, you can assess a student in

November and update their results that you

380

00:21:36,876 --> 00:21:40,077

have potentially from previous parts in

the year as well, too.

381

00:21:40,077 --> 00:21:41,037

So there's a lot of benefits.

382

00:21:41,037 --> 00:21:42,397

I guess I could keep going.

383

00:21:42,702 --> 00:21:48,345

I'm talking to a BASE podcast, so probably

I already know most of it.

384

00:21:48,345 --> 00:21:48,445

Yeah.

385

00:21:48,445 --> 00:21:52,068

I mean, a lot of people are also listening

to understand what BASE is all about and

386

00:21:52,068 --> 00:21:53,709

how that could help them in their own

field.

387

00:21:53,709 --> 00:22:00,353

So that's definitely useful if we have

some psychometricians in the audience who

388

00:22:00,353 --> 00:22:05,737

haven't tried yet some BASE, well, I'm

guessing that would be useful for them.

389

00:22:06,358 --> 00:22:08,359

And actually, could you share an example?

390

00:22:08,359 --> 00:22:12,430

If you have one of a research project

where BASE and stats played a

391

00:22:12,430 --> 00:22:16,992

a crucial role, ideally in uncovering

insights that might have been missed

392

00:22:16,992 --> 00:22:20,154

otherwise, especially using traditional

stats approaches?

393

00:22:20,154 --> 00:22:27,518

Yeah, I mean, just honestly, a lot of what

we do just estimating the model itself, it

394

00:22:27,518 --> 00:22:29,639

sounds like it should be trivial.

395

00:22:30,199 --> 00:22:36,462

But to do so with a full information

likelihood function is so difficult.

396

00:22:36,563 --> 00:22:41,745

I would say almost every single analysis

I've done using a multidimensional

397

00:22:42,706 --> 00:22:47,088

has been made possible because of the

Bayesian analyses themselves.

398

00:22:47,088 --> 00:22:49,890

Again, there are shortcut methods you

would call that.

399

00:22:49,890 --> 00:22:53,432

I think there are good methods, but again,

there are people, like I mentioned, that

400

00:22:53,432 --> 00:22:55,633

sort of a hybrid marginal maximum

likelihood.

401

00:22:55,633 --> 00:22:59,175

There's what we would call limited

information approaches that you might see

402

00:22:59,175 --> 00:23:04,138

in programs like M plus, or there's an R

package named Laban that do such things.

403

00:23:04,138 --> 00:23:10,361

But those only use functions of the data,

not the full data themselves.

404

00:23:11,150 --> 00:23:16,494

I mean, it's still good, but it's sort of

I have this sense that the full likelihood

405

00:23:16,494 --> 00:23:17,555

is what we should be using.

406

00:23:17,555 --> 00:23:23,520

So to me, just a simple example, take a, I

was working this morning with a four

407

00:23:23,520 --> 00:23:28,683

dimensional assessment, an assessment, you

know, 20 item test, kids in schools.

408

00:23:28,924 --> 00:23:37,231

And you know, I would have a difficult

time trying to estimate that with a full

409

00:23:37,231 --> 00:23:38,992

maximum likelihood method.

410

00:23:39,232 --> 00:23:40,822

And so Bayes made that possible.

411

00:23:40,822 --> 00:23:46,525

But beyond that, if we ever want to do

something with the test scores afterwards,

412

00:23:46,525 --> 00:23:46,825

right?

413

00:23:46,825 --> 00:23:51,327

So now we have a bunch of Markov chains of

people's scores themselves.

414

00:23:51,527 --> 00:23:57,191

This makes it easy to be able to then not

forget that these scores are not measured

415

00:23:57,191 --> 00:23:58,151

perfectly.

416

00:23:58,791 --> 00:24:02,493

And take a posterior distribution and use

that in a secondary analysis as well, too.

417

00:24:02,493 --> 00:24:08,637

So I was doing some work with one of the

Persian Gulf states where they were trying

418

00:24:08,637 --> 00:24:09,177

to

419

00:24:11,142 --> 00:24:15,044

like a vocational interest survey.

420

00:24:15,245 --> 00:24:19,548

And some of the classical methods for

this, sort of they disregarded any error

421

00:24:19,548 --> 00:24:20,048

whatsoever.

422

00:24:20,048 --> 00:24:23,951

And they basically said, oh, you're

interested in, I don't know, artistic work

423

00:24:23,951 --> 00:24:27,833

or you know, numeric work of some sort.

424

00:24:27,833 --> 00:24:29,334

And they would just tell you, oh, that's

it.

425

00:24:29,334 --> 00:24:29,895

That's your story.

426

00:24:29,895 --> 00:24:31,736

Like, I don't know if you've ever taken

one of those.

427

00:24:31,776 --> 00:24:32,937

What are you gonna do in a career?

428

00:24:32,937 --> 00:24:35,438

You're in a high school student and you're

trying to figure this out.

429

00:24:36,199 --> 00:24:39,782

But if you propagate, if you allow that

error to sort of propagate,

430

00:24:39,782 --> 00:24:44,226

through the way Bayesian methods make it

very easy to do, you'll see that while

431

00:24:44,226 --> 00:24:49,590

that may be the most likely choice of what

you're interested in or what your sort of

432

00:24:49,590 --> 00:24:54,374

dimensions that may be most salient to you

in your interests, there are many other

433

00:24:54,374 --> 00:24:57,697

choices that may even be close to that as

well.

434

00:24:57,697 --> 00:24:59,499

And that would be informative as well too.

435

00:24:59,499 --> 00:25:04,223

So we sort of forget, we sort of overstate

how certain we are in results.

436

00:25:04,223 --> 00:25:07,065

And I think a lot of the Bayesian methods

built around it.

437

00:25:07,065 --> 00:25:07,446

So

438

00:25:07,446 --> 00:25:11,148

That was one actually project where I did

write the own algorithm for it to try to

439

00:25:11,148 --> 00:25:14,410

estimate these things because it was just

a little more streamlined.

440

00:25:14,410 --> 00:25:21,015

But it seemed it seemed that would rather

than telling a high school student, hey,

441

00:25:21,015 --> 00:25:22,776

you're best at artistic things.

442

00:25:22,776 --> 00:25:26,458

What we could say is, hey, yeah, you may

be best at artistic, but really close to

443

00:25:26,458 --> 00:25:29,400

that is something that's numeric, you

know, like something along those lines.

444

00:25:29,400 --> 00:25:31,606

So while you're strong at art.

445

00:25:31,606 --> 00:25:32,866

You're really strong at math too.

446

00:25:32,866 --> 00:25:36,367

Maybe you should consider one of these two

rather than just go down a path that may

447

00:25:36,367 --> 00:25:38,348

or may not really reflect your interests.

448

00:25:38,348 --> 00:25:41,770

Hope that's a good example.

449

00:25:41,770 --> 00:25:42,430

Yeah.

450

00:25:42,430 --> 00:25:43,490

Yeah, definitely.

451

00:25:45,131 --> 00:25:45,671

Yeah, thanks.

452

00:25:45,671 --> 00:25:49,853

And I understand how that would be useful

for sure.

453

00:25:51,093 --> 00:25:57,776

And how does, I'm curious about the role

of priors in all that, because that's

454

00:25:57,776 --> 00:26:01,297

often something that puzzles beginners.

455

00:26:01,578 --> 00:26:08,222

And so you obviously have a lot of

experience in the Bayesian way of life in

456

00:26:08,222 --> 00:26:09,243

your field.

457

00:26:09,623 --> 00:26:18,089

So I'm curious, I'm guessing that you kind

of teach the way to do psychometric

458

00:26:18,089 --> 00:26:21,191

analysis in the Bayesian framework to a

lot of people.

459

00:26:21,952 --> 00:26:27,616

And I'm curious, especially on the prior

side, and if there are other interesting

460

00:26:27,616 --> 00:26:29,857

things that you would like to share on

that, feel free.

461

00:26:30,614 --> 00:26:32,415

My question is on the priors.

462

00:26:32,775 --> 00:26:38,540

How do you approach the challenge of

choosing appropriate prior distributions,

463

00:26:38,780 --> 00:26:41,402

especially when you're dealing with

complex models?

464

00:26:42,003 --> 00:26:43,004

Great question.

465

00:26:43,004 --> 00:26:48,148

And I'm sure each field does it a little

bit differently.

466

00:26:48,148 --> 00:26:54,093

I mean, as it probably should, because

each field has its own data and models and

467

00:26:54,093 --> 00:26:56,495

already established scientific knowledge.

468

00:26:56,495 --> 00:26:58,837

So that's my way of saying.

469

00:26:59,830 --> 00:27:01,110

This is my approach.

470

00:27:02,071 --> 00:27:04,813

I'm 100% confident that it's the approach

that everybody should take.

471

00:27:04,813 --> 00:27:06,854

But let me back it up a little bit.

472

00:27:08,075 --> 00:27:15,380

So generally speaking, I teach a lot of

students who are going into, um, many of

473

00:27:15,380 --> 00:27:19,063

our students end up in the industry for

educational measurement here in the United

474

00:27:19,063 --> 00:27:19,764

States.

475

00:27:19,764 --> 00:27:25,147

Um, I like, we usually denote our score

parameters with theta.

476

00:27:25,147 --> 00:27:27,809

I like to go around saying that, yeah, I'm

teaching you to have to sell

477

00:27:28,438 --> 00:27:32,020

That's sort of what they do, you know, in

a lot of these industry settings, they're

478

00:27:32,020 --> 00:27:33,381

selling test scores.

479

00:27:34,122 --> 00:27:39,045

So if you think that that's what you're

trying to do, I think that guides to me a

480

00:27:39,045 --> 00:27:45,009

set of prior choices that try to do the

least amount of speculation.

481

00:27:45,009 --> 00:27:45,910

So what I mean by that.

482

00:27:45,910 --> 00:27:49,493

So if you look at a measurement model,

like an item response model, you know,

483

00:27:49,493 --> 00:27:51,113

there's a set of parameters to it.

484

00:27:51,134 --> 00:27:55,057

One parameter in particular, in item

response theory, we call it the

485

00:27:55,057 --> 00:27:56,706

discrimination parameter or

486

00:27:56,706 --> 00:27:59,788

Factor analysis, we call it factor

loading, and linear regression, it would

487

00:27:59,788 --> 00:28:00,849

be a slope.

488

00:28:00,929 --> 00:28:06,054

This parameter tends to govern the extent

to which an item relates to the latent

489

00:28:06,054 --> 00:28:06,834

variable.

490

00:28:06,834 --> 00:28:10,997

So the higher that parameter is, the more

that item relates.

491

00:28:11,478 --> 00:28:15,702

Then when we go and do a Bayes theorem to

get a point estimate of a person's score

492

00:28:15,702 --> 00:28:19,465

or a posterior distribution of that

person's score, the contribution of that

493

00:28:19,465 --> 00:28:20,122

item.

494

00:28:20,122 --> 00:28:22,683

is largely reflected by the magnitude of

that parameter.

495

00:28:22,683 --> 00:28:26,245

The higher the parameter that is, the more

that item has weight on that distribution,

496

00:28:26,245 --> 00:28:28,266

the more we think we know about a person.

497

00:28:28,486 --> 00:28:33,649

So in doing that, when I look at setting

prior choices, what I try to do for that

498

00:28:33,729 --> 00:28:40,473

is to set a prior that would be toward

zero, mainly, actually at zero mostly, try

499

00:28:40,473 --> 00:28:44,795

to set it so that we want our data to tell

more of the job than our prior,

500

00:28:44,795 --> 00:28:49,037

particularly if we're trying to, if this

score has a big,

501

00:28:49,330 --> 00:28:53,853

uh, meaning to somebody you think of, um,

well, in the United States, the assessment

502

00:28:53,853 --> 00:28:57,395

culture is a little bit out of control,

but, you know, we have to take tests to go

503

00:28:57,395 --> 00:28:57,895

to college.

504

00:28:57,895 --> 00:29:00,377

We have to take tests to go to graduate

school and so forth.

505

00:29:00,377 --> 00:29:03,659

Uh, then of course, if you go and work in

certain industries, there's assessments to

506

00:29:03,659 --> 00:29:05,140

do licensure, right?

507

00:29:05,140 --> 00:29:08,703

So if you, you know, for instance, my

family is a, I come from that family of

508

00:29:08,703 --> 00:29:13,806

nurses, uh, it's a very noble profession,

but to, to be licensed in a nurse in

509

00:29:13,806 --> 00:29:15,947

California, you have to pass an exam.

510

00:29:19,254 --> 00:29:23,216

provide that score for the exam that we're

not, that score reflects as much of the

511

00:29:23,216 --> 00:29:25,237

data as possible unless a prior choice.

512

00:29:25,237 --> 00:29:32,120

And so there are ways that, you know,

people can sort of use priors, they're

513

00:29:32,221 --> 00:29:36,143

sort of not necessarily empirical science

benefit, you can sort of put too much

514

00:29:36,143 --> 00:29:37,584

subjective weight onto it.

515

00:29:37,584 --> 00:29:42,126

So when I talk about priors, when I talk

about the, I try to talk about the

516

00:29:42,126 --> 00:29:46,028

ramifications of the choice of prior on

certain parameters, that discrimination

517

00:29:46,028 --> 00:29:48,869

parameter or slope, I tend to want

518

00:29:48,970 --> 00:29:52,553

to have the data to force it to be further

away from zero because then I'm being more

519

00:29:52,553 --> 00:29:54,034

conservative, I feel like.

520

00:29:54,034 --> 00:29:58,738

The rest of the parameters, I tend to not

use heavy priors on what I do.

521

00:29:58,738 --> 00:30:03,361

I tend to use some very uninformative

priors unless I have to.

522

00:30:03,802 --> 00:30:08,686

And then the most complicated prior for

what we do, and the one that's caused

523

00:30:08,686 --> 00:30:12,549

historically the biggest challenge,

although it's, I think, relatively in good

524

00:30:12,549 --> 00:30:16,072

place these days thanks to research and

science, is the prior that goes on a

525

00:30:16,072 --> 00:30:18,053

covariance or correlation matrix.

526

00:30:18,650 --> 00:30:22,732

That had been incredibly difficult to try

to estimate back in the day.

527

00:30:22,972 --> 00:30:28,256

But now things are much, much easier in

modern computing, in modern ways of

528

00:30:28,256 --> 00:30:29,976

looking, modern priors actually.

529

00:30:31,978 --> 00:30:33,038

Yeah, interesting.

530

00:30:33,038 --> 00:30:36,440

Would you like to walk us a bit through

that?

531

00:30:36,440 --> 00:30:41,083

What are you using these days on priors on

correlation or covariance matrices?

532

00:30:41,083 --> 00:30:45,310

Because, yeah, I do teach those also

because...

533

00:30:45,310 --> 00:30:46,170

I love it.

534

00:30:46,250 --> 00:30:52,596

Basically, if you're using, for instance,

a linear regression and want to estimate

535

00:30:52,596 --> 00:30:57,520

not only the correlation of the

parameters, the predictors on the outcome,

536

00:30:57,520 --> 00:31:03,926

but also the correlation between the

predictors themselves and then using that

537

00:31:03,926 --> 00:31:08,169

additional information to make even better

prediction on the outcome, you would, for

538

00:31:08,169 --> 00:31:13,322

instance, use a multivariate normal on the

parameters on your slopes.

539

00:31:13,322 --> 00:31:18,224

of your linear regression, for instance,

what primaries do you use on that

540

00:31:18,364 --> 00:31:19,305

multivariate?

541

00:31:19,305 --> 00:31:21,266

What does the multivariate normal mean?

542

00:31:21,266 --> 00:31:24,308

And a multivariate normal needs a

covariance matrix.

543

00:31:24,308 --> 00:31:26,709

So what primaries do you use on the

covariance matrix?

544

00:31:26,709 --> 00:31:28,930

So that's basically the context for

people.

545

00:31:29,711 --> 00:31:34,274

Now, John, basically try and take it from

there.

546

00:31:34,274 --> 00:31:37,215

What are you using in your field these

days?

547

00:31:37,395 --> 00:31:41,417

Yeah, so going with your example, I have

no idea.

548

00:31:41,526 --> 00:31:45,207

You know, like, if you have a set of

regression coefficients that you say are

549

00:31:45,207 --> 00:31:49,929

multivariate normal, yes, there is a place

for a covariance in the prior.

550

00:31:51,930 --> 00:31:54,231

I never try to speculate what that is.

551

00:31:54,231 --> 00:31:58,153

I don't think I have, like, the human

judgment that it takes to figure out what

552

00:31:58,153 --> 00:32:00,774

the, like, the belief, your prior belief

is for that.

553

00:32:00,774 --> 00:32:06,256

I think you're talking about what would be

analogous to sort of the, like, the

554

00:32:06,256 --> 00:32:08,277

asthmatotic covariance matrix.

555

00:32:09,078 --> 00:32:11,639

The posterior distribution of these

parameters where you look at the

556

00:32:11,639 --> 00:32:15,682

covariance between them is like the

asymptotic covariance matrix in ML, and we

557

00:32:15,682 --> 00:32:19,985

just rarely ever speculate off of the

diagonal, it seems like, on that.

558

00:32:19,985 --> 00:32:23,248

I mean, there are certainly uses for

linear combinations and whatnot, but

559

00:32:23,248 --> 00:32:24,729

that's tough.

560

00:32:24,729 --> 00:32:28,832

I'm more thinking about, like, when I have

a handful of latent variables and try to

561

00:32:28,832 --> 00:32:34,036

estimate, now the problem is I need a

covariance matrix between them, and

562

00:32:34,036 --> 00:32:36,117

they're likely to be highly correlated,

right?

563

00:32:36,117 --> 00:32:36,777

So...

564

00:32:37,606 --> 00:32:42,510

In our field, we tend to see correlations

of psychological variables that are 0.7,

565

00:32:42,550 --> 00:32:43,531

0.8, 0.9.

566

00:32:44,212 --> 00:32:49,816

These are all academic skills in my field

that are coming from the same brain.

567

00:32:50,277 --> 00:32:54,440

The child has a lot of reasons why those

are going to be highly correlated.

568

00:32:55,962 --> 00:33:01,246

And so these days, I love the LKJ prior

for it.

569

00:33:01,246 --> 00:33:04,609

It makes it easy to put a prior on a

covariance matrix and then if you want to

570

00:33:04,609 --> 00:33:05,629

rescale it.

571

00:33:05,802 --> 00:33:09,403

That's one of the other weird features of

the psychometric world is that because

572

00:33:09,403 --> 00:33:14,345

these variables don't exist, to estimate

covariance matrix, we'd have to make

573

00:33:14,345 --> 00:33:18,267

certain constraints on the, on some of the

item parameters, the measurement model for

574

00:33:18,267 --> 00:33:19,128

instance.

575

00:33:19,128 --> 00:33:24,710

If we want a variance of the factor, we

have to set one of the parameters of the

576

00:33:25,350 --> 00:33:27,872

discrimination parameters to a value to be

able to estimate it.

577

00:33:27,872 --> 00:33:29,512

Otherwise, it's not identified.

578

00:33:32,758 --> 00:33:35,798

work that we talk about for calibration

when we're trying to build scores or build

579

00:33:35,798 --> 00:33:40,840

assessments and their data for it, we fix

that value of the variance of a factor to

580

00:33:40,840 --> 00:33:40,960

one.

581

00:33:40,960 --> 00:33:45,981

We standardize the factor zero, meaning

variance one, very simple idea.

582

00:33:46,301 --> 00:33:50,783

The models are equivalent in a classic

sense, in that the likelihoods are

583

00:33:50,783 --> 00:33:52,623

equivalent, whether we do one way or the

other.

584

00:33:52,623 --> 00:33:56,184

When we put products on the posteriors

aren't entirely equivalent, but that's a

585

00:33:56,184 --> 00:33:59,345

matter of a typical Bayesian issue with

transformations.

586

00:33:59,345 --> 00:34:00,045

But

587

00:34:00,894 --> 00:34:06,376

In the sense where we want a correlation

matrix, prior to the LKJ, prior, there

588

00:34:06,376 --> 00:34:11,798

were all these sort of, one of my mentors,

Rod McDonald, called devices, little hacks

589

00:34:11,798 --> 00:34:16,900

or tricks that we would do to sort of keep

covariance matrix, sample it, right?

590

00:34:16,900 --> 00:34:22,543

I mean, you think about statistically to

sample it, I like a lot of rejection

591

00:34:22,543 --> 00:34:23,523

sampling methods.

592

00:34:23,523 --> 00:34:27,925

So if you were to basically propose a

covariance or correlation matrix, it has

593

00:34:27,925 --> 00:34:28,945

to be positive.

594

00:34:28,970 --> 00:34:32,894

semi-definite, that's a hard term.

595

00:34:32,894 --> 00:34:38,321

It has to be, you have to make sure that

the correlation is bounded and so forth.

596

00:34:38,321 --> 00:34:43,167

But LKJ takes care of almost all of that

for me in a way that allows me to just

597

00:34:43,167 --> 00:34:47,132

model the straight correlation matrix,

which has really made life a lot easier

598

00:34:47,132 --> 00:34:48,653

when it comes to estimation.

599

00:34:50,838 --> 00:34:54,719

Yeah, I mean, I'm not surprised that does.

600

00:34:54,719 --> 00:35:00,902

I mean, that is also the kind of priors I

tend to use personally and that I teach

601

00:35:00,902 --> 00:35:01,542

also.

602

00:35:02,263 --> 00:35:07,085

In this example, for instance, of the

linear regression, that's what I probably

603

00:35:07,085 --> 00:35:15,008

end up using LKJPrior on the predictors on

the slopes of the linear regression.

604

00:35:15,989 --> 00:35:18,130

And for people who don't know,

605

00:35:18,130 --> 00:35:19,610

Never used LKJ prior.

606

00:35:19,610 --> 00:35:26,112

LKJ is decomposition of the covariance

matrix.

607

00:35:26,932 --> 00:35:30,413

That way, we can basically sample it.

608

00:35:30,793 --> 00:35:34,875

Otherwise, it's extremely hard to sample

from a covariance matrix.

609

00:35:34,995 --> 00:35:43,357

But the LKJ decomposition of the matrix is

a way to basically an algebraic trick.

610

00:35:44,578 --> 00:35:49,400

that makes use of the Cholesky

decomposition of a covariance matrix that

611

00:35:49,420 --> 00:35:54,382

allows us to sample the Cholesky

decomposition instead of the covariance

612

00:35:54,382 --> 00:35:56,023

matrix fully, and that helps the sampling.

613

00:35:56,023 --> 00:35:56,383

Thank you.

614

00:35:56,383 --> 00:35:57,243

Thank you for putting that out there.

615

00:35:57,243 --> 00:36:02,826

I'm glad you put that on.

616

00:36:03,566 --> 00:36:05,267

Yeah, so yeah.

617

00:36:05,267 --> 00:36:09,129

And basically, the way you would

parametrize that, for instance, in Poem C,

618

00:36:09,129 --> 00:36:09,889

you would

619

00:36:10,610 --> 00:36:18,553

use pm.lkj, and basically you would have

to parameterize that with at least three

620

00:36:18,553 --> 00:36:20,414

parameters, the number of dimensions.

621

00:36:20,414 --> 00:36:24,695

So for instance, if you have three

predictors, that would be n equals 3.

622

00:36:25,576 --> 00:36:31,518

The standard deviation that you are

expecting on the predictors on the slopes

623

00:36:31,518 --> 00:36:34,820

of the linear regression, so that's

something you're used to, right?

624

00:36:34,820 --> 00:36:39,662

If you're using a normal prior on the

slope, then the sigma of the slope is just

625

00:36:39,662 --> 00:36:45,506

standard deviation that you're expecting

on that effect for your data and model.

626

00:36:45,587 --> 00:36:54,454

And then you have to specify a prior on

the correlation of these slopes.

627

00:36:54,454 --> 00:36:58,737

And that's where you get into the

covariance part.

628

00:36:58,737 --> 00:37:02,581

And so basically, you can specify a prior.

629

00:37:02,581 --> 00:37:07,104

So that would be called eta in PIME-Z on

the LKJ prior.

630

00:37:07,485 --> 00:37:08,606

And the

631

00:37:08,606 --> 00:37:17,853

bigger eta, the more suspicious of high

correlations your prior would be.

632

00:37:17,853 --> 00:37:23,197

So if eta equals 1, you're basically

expecting a uniform distribution of

633

00:37:23,197 --> 00:37:23,957

correlations.

634

00:37:23,957 --> 00:37:26,679

That could be minus 1, that could be 1,

that could be 0.

635

00:37:26,679 --> 00:37:29,121

All of those have the same weight.

636

00:37:29,121 --> 00:37:33,164

And then if you go to eta equals 8, for

instance, you would put much more prior

637

00:37:33,164 --> 00:37:35,225

weight on correlations eta.

638

00:37:35,834 --> 00:37:41,638

Close to zero, much of them will be close

to zero in 0.5 minus 0.5, but it would be

639

00:37:41,638 --> 00:37:45,500

very suspicious of very big correlations,

which I guess would make a lot of sense,

640

00:37:45,500 --> 00:37:47,221

for instance, social science.

641

00:37:47,902 --> 00:37:50,524

I don't know in your field, but yeah.

642

00:37:50,524 --> 00:37:54,947

I typically use the uniform, the one

setting, at least to start with, but yeah,

643

00:37:54,947 --> 00:37:57,288

I think that's a great description.

644

00:37:57,288 --> 00:37:59,710

Very good description.

645

00:37:59,710 --> 00:38:04,373

Yeah, I really love these kinds of models

because they make linear regression even

646

00:38:04,373 --> 00:38:05,313

more powerful.

647

00:38:05,514 --> 00:38:11,278

To me, linear regression is so powerful

and very underrated.

648

00:38:11,319 --> 00:38:19,846

You can go so far with plain linear

regression and often it's hard to really

649

00:38:19,846 --> 00:38:20,467

do better.

650

00:38:20,467 --> 00:38:24,390

You have to work a lot to do better than a

really good linear regression.

651

00:38:25,191 --> 00:38:26,692

I completely agree with you.

652

00:38:26,692 --> 00:38:28,793

Yeah, I'm 100% right there.

653

00:38:30,455 --> 00:38:33,337

And actually then you get into sort of

the...

654

00:38:33,742 --> 00:38:37,144

quadratic or the nonlinear forms in linear

regression that map onto it that make it

655

00:38:37,144 --> 00:38:37,524

even more powerful.

656

00:38:37,524 --> 00:38:42,928

So yeah, it's absolutely wonderful.

657

00:38:42,928 --> 00:38:46,030

Yeah, yeah.

658

00:38:46,030 --> 00:38:51,313

And I mean, as Spider-Man's uncle said,

great power comes with great

659

00:38:51,313 --> 00:38:52,394

responsibility.

660

00:38:52,394 --> 00:38:59,219

So you have to be very careful about the

priors when you have all those features,

661

00:38:59,219 --> 00:39:03,121

so inversing functions because they

662

00:39:03,974 --> 00:39:10,319

the parameter space, but same thing, well,

if you're using a multivariate normal, I

663

00:39:10,319 --> 00:39:11,300

mean, that's more complex.

664

00:39:11,300 --> 00:39:15,464

So of course you have to think a bit more

about your model structure, about your

665

00:39:15,464 --> 00:39:16,144

prior.

666

00:39:16,144 --> 00:39:21,769

And also the more structure you add, if

the size of the data is kept equal, well,

667

00:39:21,769 --> 00:39:27,054

that means you have more risk for

overfitting and you have less informative

668

00:39:27,054 --> 00:39:29,416

power per data point.

669

00:39:29,416 --> 00:39:30,516

Let's say so.

670

00:39:30,597 --> 00:39:31,877

That means the prior.

671

00:39:32,102 --> 00:39:37,344

increase in importance, so you have to

think about them more.

672

00:39:37,425 --> 00:39:45,149

But you get a much more powerful model

after once and the goal is to get much

673

00:39:45,149 --> 00:39:49,491

more powerful predictions after once.

674

00:39:49,491 --> 00:39:50,472

I do agree.

675

00:39:51,172 --> 00:39:53,533

These weapons are hard to wield.

676

00:39:55,054 --> 00:39:57,816

They require time and effort.

677

00:39:58,816 --> 00:40:00,854

And on my end, I don't know for you.

678

00:40:00,854 --> 00:40:07,917

Jonathan, but on my end, they also require

a lot of caffeine from time to time.

679

00:40:07,917 --> 00:40:08,177

Maybe.

680

00:40:08,177 --> 00:40:08,657

Yeah.

681

00:40:08,657 --> 00:40:12,758

I mean, so that's the key.

682

00:40:12,979 --> 00:40:15,660

You see how I did the segue.

683

00:40:15,660 --> 00:40:16,620

I should have a podcast.

684

00:40:16,620 --> 00:40:17,841

Yeah.

685

00:40:17,841 --> 00:40:24,464

So as a first time I do that in the

podcast, but I had that.

686

00:40:24,464 --> 00:40:24,744

Yeah.

687

00:40:24,744 --> 00:40:26,665

So I'm a big coffee drinker.

688

00:40:26,665 --> 00:40:27,385

I love coffee.

689

00:40:27,385 --> 00:40:28,905

I'm a big coffee nerd.

690

00:40:29,246 --> 00:40:36,608

But from time to time, I try to decrease

my caffeine usage, you know, also because

691

00:40:36,608 --> 00:40:38,668

you have some habituation effects.

692

00:40:38,668 --> 00:40:46,071

So if I want to keep the caffeine shot

effect, well, I have to sometimes do a

693

00:40:46,071 --> 00:40:48,231

decrease of my usage.

694

00:40:48,231 --> 00:40:54,033

And funnily enough, when I was thinking

about that, a small company called Magic

695

00:40:54,033 --> 00:40:55,953

Mind, they came to me...

696

00:40:56,190 --> 00:40:59,652

They sent me an email and they listened to

the show and they were like, hey, you've

697

00:40:59,652 --> 00:41:01,013

got a cool show.

698

00:41:01,514 --> 00:41:08,458

I would be happy to send you some bottles

for you to try and to talk about it on the

699

00:41:08,458 --> 00:41:08,839

show.

700

00:41:08,839 --> 00:41:11,260

And I thought that was fun.

701

00:41:11,260 --> 00:41:15,243

So I got some Magic Mind myself.

702

00:41:15,764 --> 00:41:21,908

I drank it, but I'm not going to buy

Jonathan because I got Magic Mind to send

703

00:41:21,908 --> 00:41:24,429

some samples to Jonathan.

704

00:41:24,566 --> 00:41:31,849

And if you are watching the YouTube video,

Jonathan is going to try the Magic Mind

705

00:41:31,989 --> 00:41:33,970

right now, live.

706

00:41:34,090 --> 00:41:37,251

So yeah, take it away, Jon.

707

00:41:37,632 --> 00:41:41,334

Yeah, this is interesting because you

reached out to me for the podcast and I

708

00:41:41,334 --> 00:41:45,035

had not met you, but you know, it's a

conversation, it's a podcast, you have to

709

00:41:45,035 --> 00:41:46,156

do great work.

710

00:41:46,156 --> 00:41:47,116

Yes, I'll say yes to that.

711

00:41:47,116 --> 00:41:48,897

Then you said, how would you like to try

the Magic Mind?

712

00:41:48,897 --> 00:41:49,837

And I thought...

713

00:41:50,146 --> 00:41:53,368

being a psych major as an undergraduate,

this is an interesting social psychology

714

00:41:53,368 --> 00:41:56,010

experiment where a random person from the

internet says, hey, I'll send you

715

00:41:56,010 --> 00:41:56,770

something.

716

00:41:56,931 --> 00:42:01,674

So I thought there's a little bit of

safety in that by drinking it in front of

717

00:42:01,694 --> 00:42:03,415

you while we're talking on the podcast.

718

00:42:03,415 --> 00:42:10,381

But of course, I know you can cut this out

if I hit the floor, but here it comes.

719

00:42:10,381 --> 00:42:12,362

So you're drinking it like, sure.

720

00:42:12,362 --> 00:42:16,145

Yeah, I decided to drink it like a shot,

if you will.

721

00:42:17,410 --> 00:42:19,671

It was actually tasted much better than I

expected.

722

00:42:19,671 --> 00:42:21,752

It came in a bottle with green.

723

00:42:21,752 --> 00:42:25,074

It tasted tangy, so very good.

724

00:42:25,074 --> 00:42:32,759

And now the question will be, if I get

better at my answers to your questions by

725

00:42:32,759 --> 00:42:35,920

the end of the podcast, therefore we have

now a nice experiment.

726

00:42:35,920 --> 00:42:39,843

But no, I noticed it has a bit of

caffeine, certainly less than a cup of

727

00:42:39,843 --> 00:42:40,163

coffee.

728

00:42:40,163 --> 00:42:44,885

But at the same time, it doesn't seem

offensive whatsoever.

729

00:42:45,942 --> 00:42:46,742

Yeah, that's pretty good.

730

00:42:46,742 --> 00:42:50,364

Yeah, I mean, I'm still drinking caffeine,

if that's all right.

731

00:42:50,364 --> 00:42:52,245

But yeah, from time to time, I like to

drink it.

732

00:42:52,245 --> 00:42:53,926

My habituation, my answer to that is just

drink more.

733

00:42:53,926 --> 00:42:55,007

That's fine.

734

00:42:55,007 --> 00:42:57,268

Yeah, exactly.

735

00:42:57,868 --> 00:42:59,729

Oh yeah, and decaf and stuff like that.

736

00:42:59,729 --> 00:43:03,351

But yeah, I love the idea of the product

is cool.

737

00:43:03,691 --> 00:43:04,292

I liked it.

738

00:43:04,292 --> 00:43:07,253

So I was like, yeah, I'm going to give it

a shot.

739

00:43:07,253 --> 00:43:13,457

And so the way I drank it was also

basically making myself a latte

740

00:43:14,954 --> 00:43:20,637

coffee, I would use the Magic Pint and

then I would put my milk in the milk foam.

741

00:43:21,378 --> 00:43:22,819

And that is really good.

742

00:43:22,999 --> 00:43:23,600

I have to say.

743

00:43:23,600 --> 00:43:24,540

See how that works.

744

00:43:26,542 --> 00:43:26,962

Yeah.

745

00:43:26,962 --> 00:43:31,926

So it's based on, I mean, the thing you

taste most is the matcha, I think.

746

00:43:32,526 --> 00:43:38,391

And usually I'm not a big fan of matcha

and that's why I give it the green color.

747

00:43:38,391 --> 00:43:42,673

I think usually I'm not, but I had to say,

I really appreciated that.

748

00:43:43,238 --> 00:43:44,999

You and me both, I was feeling the same

way.

749

00:43:44,999 --> 00:43:49,282

When I saw it come in the mail, I was

like, ooh, that added to my skepticism,

750

00:43:49,282 --> 00:43:49,442

right?

751

00:43:49,442 --> 00:43:50,803

I'm trying to be a good scientist.

752

00:43:50,803 --> 00:43:52,204

I'm trying to be like, yeah.

753

00:43:52,204 --> 00:43:57,989

But yeah, it was actually surprisingly,

tasted more like a juice, like a citrus

754

00:43:57,989 --> 00:43:59,390

juice than it was matcha.

755

00:43:59,390 --> 00:44:04,053

So it was much nicer than I expected.

756

00:44:04,053 --> 00:44:07,816

Yeah, I love that because me too, I'm

obviously extremely skeptical about all

757

00:44:07,816 --> 00:44:08,557

those stuff.

758

00:44:08,557 --> 00:44:09,197

So.

759

00:44:10,290 --> 00:44:13,631

I like doing that.

760

00:44:13,631 --> 00:44:19,853

It's way better, way more fun to do it

with you or any other nerd from the

761

00:44:19,853 --> 00:44:23,574

community than doing it with normal people

from the street because I'm way too

762

00:44:23,574 --> 00:44:24,854

skeptical for them.

763

00:44:24,854 --> 00:44:29,396

They wouldn't even understand my

skepticism.

764

00:44:29,396 --> 00:44:29,756

I agree.

765

00:44:29,756 --> 00:44:32,357

I felt like in a scientific community,

I've seen some of the people you've had on

766

00:44:32,357 --> 00:44:35,537

the podcast, we're all a little bit

skeptical about what we do.

767

00:44:36,618 --> 00:44:40,079

I could bring that skepticism here and I'd

feel like at home, hopefully.

768

00:44:40,079 --> 00:44:42,139

I'm glad that you allowed me to do that.

769

00:44:42,139 --> 00:44:42,820

Yeah.

770

00:44:42,820 --> 00:44:44,360

And that's the way of life.

771

00:44:48,241 --> 00:44:55,103

Thanks for trusting me because I agree

that seeing from a third party observer,

772

00:44:55,103 --> 00:44:57,644

you'd be like, that sounds like a scam.

773

00:44:58,504 --> 00:45:01,885

That guy is just inviting me to sell him

something to me.

774

00:45:02,970 --> 00:45:06,091

In a week, he's going to send me an email

to tell me he's got some financial

775

00:45:06,091 --> 00:45:08,551

troubles and I have to wire him $10,000.

776

00:45:10,452 --> 00:45:14,053

Waiting for that or is it, what level of

paranoia do I have this morning?

777

00:45:14,053 --> 00:45:18,314

I was like, well, who are my enemies and

who really wants to do something bad to

778

00:45:18,314 --> 00:45:18,514

me?

779

00:45:18,514 --> 00:45:18,834

Right?

780

00:45:18,834 --> 00:45:22,075

So, I don't believe I'm at that level.

781

00:45:22,075 --> 00:45:23,996

So I don't think I have anything to worry

about.

782

00:45:23,996 --> 00:45:25,797

It seems like a reputable company.

783

00:45:25,797 --> 00:45:27,317

So it was, it was amazing.

784

00:45:27,317 --> 00:45:28,597

Yeah.

785

00:45:29,002 --> 00:45:30,102

No, that was good.

786

00:45:30,102 --> 00:45:35,046

Thanks a lot MagicMine for sending me

those samples, that was really fun.

787

00:45:35,046 --> 00:45:39,690

Feel free to give it a try, other people

if you want, if that sounded like

788

00:45:39,690 --> 00:45:41,771

something you'd be interested in.

789

00:45:41,851 --> 00:45:47,075

And if you have any other product to send

me, send them to me, I mean, that sounds

790

00:45:47,075 --> 00:45:47,395

fun.

791

00:45:47,395 --> 00:45:50,838

I mean, I'm not gonna say yes to

everything, you know, I have standards on

792

00:45:50,838 --> 00:45:53,500

the show, and especially scientific

standards.

793

00:45:53,900 --> 00:45:56,001

But you can always send me something.

794

00:45:56,022 --> 00:45:57,283

And I will always analyze it.

795

00:45:57,283 --> 00:45:59,545

You know, somehow you can work out an

agreement with the World Cup, right?

796

00:45:59,545 --> 00:46:01,126

Some World Cup tickets for the next time.

797

00:46:01,126 --> 00:46:01,246

True.

798

00:46:01,246 --> 00:46:01,626

That would be nice.

799

00:46:01,626 --> 00:46:03,908

True.

800

00:46:03,908 --> 00:46:07,411

Yeah, exactly.

801

00:46:07,651 --> 00:46:07,972

Awesome.

802

00:46:07,972 --> 00:46:14,437

Well, what we did is actually kind of

related, I think, I would say to the

803

00:46:14,437 --> 00:46:16,479

other, another aspect of your work.

804

00:46:16,479 --> 00:46:18,740

And that is model comparison.

805

00:46:20,102 --> 00:46:25,045

So, and it's again, a topic that's asked a

lot by students.

806

00:46:25,378 --> 00:46:29,860

Especially when they come from the

classical machine learning framework where

807

00:46:29,901 --> 00:46:33,663

model comparison is just everywhere.

808

00:46:33,984 --> 00:46:37,446

So often they ask how they can do that in

the Bayesian framework.

809

00:46:37,926 --> 00:46:42,970

Again, as usual, I am always skeptical

about just doing model comparison and just

810

00:46:42,970 --> 00:46:46,032

picking your model based on some one

statistic.

811

00:46:46,032 --> 00:46:50,755

I always say there is no magic one

matching bullet, you know, in the Bayesian

812

00:46:50,755 --> 00:46:54,262

framework where it's just, okay, model

comparisons say that, so for sure.

813

00:46:54,262 --> 00:46:55,542

That's the best model.

814

00:46:56,083 --> 00:46:57,824

I wouldn't say that's how it works.

815

00:46:57,824 --> 00:47:02,988

And you would need a collection of

different indicators, including, for

816

00:47:02,988 --> 00:47:08,451

instance, the LOO, the LOO factor, that

tells you, yeah, that model is better.

817

00:47:08,451 --> 00:47:11,453

But not only that, what about the

posterior predictions?

818

00:47:11,593 --> 00:47:12,694

What about the model structure?

819

00:47:12,694 --> 00:47:14,035

What about the priors?

820

00:47:14,095 --> 00:47:17,617

What about just the generative story about

the model?

821

00:47:17,818 --> 00:47:23,501

But talking about model comparison, what

can you tell us, John, about the

822

00:47:23,794 --> 00:47:28,217

some best practices for carrying out

effective model comparisons?

823

00:47:29,238 --> 00:47:30,359

Kajen is best practice.

824

00:47:30,359 --> 00:47:32,421

I'll just give you what my practice is.

825

00:47:32,421 --> 00:47:35,083

I will make no claim that it's best.

826

00:47:35,083 --> 00:47:36,444

It's difficult.

827

00:47:36,444 --> 00:47:41,548

I think you hit on all the aspects of it

in introducing the topic.

828

00:47:42,148 --> 00:47:45,952

If you have a set of models that you're

considering, the first thing I'd like to

829

00:47:45,952 --> 00:47:50,575

think about is not the comparison between

them as much as how each model would fit a

830

00:47:50,575 --> 00:47:52,316

data set of data

831

00:47:53,750 --> 00:47:58,833

post-serial predictive model checking is,

you know, from an amazing sense is where

832

00:47:58,833 --> 00:48:02,735

really a lot of the work for me is focused

around.

833

00:48:04,577 --> 00:48:11,242

Interestingly, what you choose to check

against is a bit of a challenge,

834

00:48:11,242 --> 00:48:14,784

particularly, you know, in certain fields

in psychometrics, at least the ones I'm

835

00:48:14,784 --> 00:48:15,844

familiar with.

836

00:48:16,445 --> 00:48:21,068

I do see a lot of, first of all, model

fit,

837

00:48:23,262 --> 00:48:26,403

well-researched area in psychometrics in

general.

838

00:48:27,124 --> 00:48:31,586

Really, there's millions of papers in the

1980s, maybe not millions, but it seems

839

00:48:31,586 --> 00:48:32,527

like that many.

840

00:48:32,527 --> 00:48:35,329

And then another, it's always been

something that people have studied.

841

00:48:35,329 --> 00:48:39,591

I think recently there's been a resurgence

of new ideas in it as well.

842

00:48:39,591 --> 00:48:43,493

So it's well-covered territory from the

psychometric literature.

843

00:48:43,493 --> 00:48:47,916

It's less well-covered, at least in my

view, in Bayesian psychometrics.

844

00:48:48,576 --> 00:48:50,837

So what I've tried to do,

845

00:48:51,466 --> 00:48:58,650

with my work to try to see if a model fits

absolutely is to look at, there's this,

846

00:48:59,311 --> 00:49:02,153

one of the complicating factors is that a

lot of my data is discrete.

847

00:49:02,153 --> 00:49:05,455

So it's correct and correct scored items.

848

00:49:05,696 --> 00:49:12,901

And in that sense, in the last 15, 20

years, there's been some good work in the

849

00:49:13,241 --> 00:49:18,005

non-Bayesian world about how to use what

we call limited information methods to

850

00:49:18,005 --> 00:49:18,825

assess model fit.

851

00:49:18,825 --> 00:49:19,765

So instead of,

852

00:49:19,786 --> 00:49:22,587

looking at model fit to the entire

contingency table.

853

00:49:22,587 --> 00:49:28,409

So if you have a set of binary data, let's

say 10 variables that you've observed,

854

00:49:28,670 --> 00:49:32,672

technically you have 1,024 different

probabilities that have permutations of

855

00:49:32,672 --> 00:49:34,592

ways they could be zeros and ones.

856

00:49:34,953 --> 00:49:40,495

And model fit should be built toward that

1,024 vector of probabilities.

857

00:49:40,495 --> 00:49:41,596

Good luck with that, right?

858

00:49:41,596 --> 00:49:43,677

You're not gonna collect enough data to do

that.

859

00:49:43,677 --> 00:49:45,777

And so...

860

00:49:45,822 --> 00:49:52,485

What a group of scientists Alberto Medeo

Alavarez, Lissai and others have created

861

00:49:52,485 --> 00:49:56,208

are sort of model fit to lower level

contingency tables.

862

00:49:56,208 --> 00:50:01,491

So each marginal moment of the day, each

mean effectively, and then like a two-way

863

00:50:01,491 --> 00:50:04,772

table between all pairs of observed

variables.

864

00:50:05,193 --> 00:50:08,455

In work that I've done with a couple of

students recently, we've tried to

865

00:50:08,455 --> 00:50:12,257

replicate that idea, but more on a

Bayesian sentence.

866

00:50:12,257 --> 00:50:13,717

So could we come up with

867

00:50:14,058 --> 00:50:16,779

and M, like a statistic, this is called an

M2 statistic.

868

00:50:16,779 --> 00:50:20,900

Could we come up with a version of a

posterior predictive check for what a

869

00:50:20,900 --> 00:50:23,461

model says the two-way table should look

like?

870

00:50:23,642 --> 00:50:29,104

And then similar to that, could we create

a model such that we know saturates that?

871

00:50:29,104 --> 00:50:34,306

So for instance, if we have 10 observed

variables, we could create a model that

872

00:50:34,306 --> 00:50:39,308

has all 10 shoes to two-way tables

estimated perfect, what we would expect to

873

00:50:39,308 --> 00:50:39,708

be perfect.

874

00:50:39,708 --> 00:50:43,349

Now, of course, there's posterior

distributions, but you would expect with

875

00:50:43,502 --> 00:50:47,725

you know, plenty of data and, you know,

very diffused priors that you would get

876

00:50:47,725 --> 00:50:50,467

point estimates, EAP estimates, and that

should be right about where you can

877

00:50:50,467 --> 00:50:52,308

observe the frequencies of data.

878

00:50:52,569 --> 00:50:54,110

Quick check.

879

00:50:54,110 --> 00:50:58,954

So, um, the idea then is now we have two

models, one of which we know should fit

880

00:50:58,954 --> 00:51:00,616

the data absolutely.

881

00:51:00,616 --> 00:51:05,320

And one of which we know, uh, we're, we're

wondering if it fits now that the

882

00:51:05,320 --> 00:51:06,621

comparison comes together.

883

00:51:06,621 --> 00:51:08,982

So we have these two predictive

distributions.

884

00:51:09,303 --> 00:51:10,484

Um, how do we compare them?

885

00:51:10,484 --> 00:51:12,662

Uh, and that's where, you know,

886

00:51:12,662 --> 00:51:13,842

different approaches we've taken.

887

00:51:13,842 --> 00:51:16,624

One of those is just simply looking at the

distributional overlaps.

888

00:51:16,624 --> 00:51:20,446

We tried to calculate a, we use the

Kilnogorov Smirnov distribution, sort of

889

00:51:20,446 --> 00:51:25,209

the sea where moments are percent wise of

the distributions with overlap, because if

890

00:51:25,209 --> 00:51:30,191

your model's data overlaps with what you

think that the data should look like, you

891

00:51:30,191 --> 00:51:31,732

think the model fits well.

892

00:51:31,732 --> 00:51:34,814

And if it doesn't, it should be far apart

and won't fit well.

893

00:51:34,814 --> 00:51:36,635

That's how we've been trying to build.

894

00:51:36,895 --> 00:51:39,637

It's weird because it's a model

comparison, but one of the comparing

895

00:51:39,637 --> 00:51:41,174

models we know to be

896

00:51:41,174 --> 00:51:45,176

what we call saturated, it should fit the

data the best and no other model, all the

897

00:51:45,176 --> 00:51:47,197

other models should be subsumed into it.

898

00:51:47,337 --> 00:51:51,640

So that's the approach I've taken recently

with posterior predictive checks, but then

899

00:51:51,640 --> 00:51:52,701

a model comparison.

900

00:51:52,701 --> 00:51:57,103

We could have used, as you mentioned, the

LOO factor or the LOO statistic.

901

00:51:57,103 --> 00:52:00,726

And maybe that's something that we should

look into also.

902

00:52:00,726 --> 00:52:05,408

We haven't yet, but one of my recent

graduates, new assistant professor at

903

00:52:05,408 --> 00:52:07,469

University of Arkansas here in the United

States.

904

00:52:07,574 --> 00:52:13,359

Ji Hang Zhang had done a lot of work on

this in his dissertation and other studies

905

00:52:13,359 --> 00:52:13,639

here.

906

00:52:13,639 --> 00:52:16,161

So that's sort of the approach I take.

907

00:52:16,181 --> 00:52:19,564

The other thing I want to mention though

is when you're comparing amongst models,

908

00:52:19,564 --> 00:52:22,106

you have to establish that model for that

absolute fit first.

909

00:52:22,106 --> 00:52:26,410

So the way I envision this is you sort of

compare your model to this sort of

910

00:52:26,410 --> 00:52:27,731

saturated model.

911

00:52:28,392 --> 00:52:32,656

You do that for multiple versions of your

models and then effectively choose amongst

912

00:52:32,656 --> 00:52:35,277

the set of models you're comparing that

sort of fit.

913

00:52:35,370 --> 00:52:39,452

But what that absolute fit is, is like you

mentioned, it's nearly impossible to tell

914

00:52:39,452 --> 00:52:39,833

exactly.

915

00:52:39,833 --> 00:52:44,936

There's a number of ideas that go into

what makes a good for a good fitting

916

00:52:44,936 --> 00:52:45,716

model.

917

00:52:46,257 --> 00:52:47,638

Yeah.

918

00:52:48,118 --> 00:52:54,222

And definitely I encourage people to go

take a look at the Lou paper.

919

00:52:54,723 --> 00:52:57,025

I will put a link in the show note to that

paper.

920

00:52:57,025 --> 00:53:03,589

And also if you're using Arvies, whether

in Julia or Python, we do have.

921

00:53:04,302 --> 00:53:06,424

implementation of the Loo algorithm.

922

00:53:06,424 --> 00:53:10,987

So comparing your models with obviously

extremely simple, it's just a call to

923

00:53:11,088 --> 00:53:14,671

compare and then you can even do a plot of

that.

924

00:53:15,252 --> 00:53:19,115

And yeah, as you were saying, the Loo

algorithm doesn't have any meaning by

925

00:53:19,115 --> 00:53:19,496

itself.

926

00:53:19,496 --> 00:53:20,056

Right?

927

00:53:20,056 --> 00:53:22,238

The Loo score of a model doesn't mean

anything.

928

00:53:22,238 --> 00:53:25,201

It's in comparison to another, to other

models.

929

00:53:25,201 --> 00:53:29,665

So yeah, basically having a baseline model

that you think is already good enough.

930

00:53:30,422 --> 00:53:34,183

And then all the other models have to

compare to that one, which basically could

931

00:53:34,183 --> 00:53:42,927

be like the placebo, if you want, or the

already existing solution that there is

932

00:53:42,927 --> 00:53:43,387

for that.

933

00:53:43,387 --> 00:53:49,790

And then any model that's more complicated

than that should be in competition with

934

00:53:49,790 --> 00:53:55,732

that one and should have a reason to be

used, because otherwise, why are you using

935

00:53:55,732 --> 00:53:58,693

a more complicated model if you could just

use

936

00:53:59,358 --> 00:54:02,420

a simple linear regression, because that's

what I use most of the time for my

937

00:54:02,420 --> 00:54:03,180

baseline model.

938

00:54:03,180 --> 00:54:03,701

Right?

939

00:54:03,701 --> 00:54:08,404

Baseline model, just use a simple linear

regression, and then do all the fancy

940

00:54:08,404 --> 00:54:14,248

modeling you want and compare that to the

linear regression, both in predictions and

941

00:54:14,248 --> 00:54:15,649

with the Loo algorithm.

942

00:54:15,649 --> 00:54:21,914

And well, if there is a good reason to

make your life more difficult, then use

943

00:54:21,914 --> 00:54:22,234

it.

944

00:54:22,234 --> 00:54:25,256

But otherwise, why would you?

945

00:54:29,454 --> 00:54:35,698

And yeah, actually talking about these

complexities, something I see is also that

946

00:54:35,798 --> 00:54:40,802

many, many people, many practitioners

might be hesitant to adopt the patient

947

00:54:40,802 --> 00:54:45,705

methods due to the fact that they perceive

them as complex.

948

00:54:45,866 --> 00:54:53,651

So I'm wondering yourself, what resources

or strategies would you recommend to those

949

00:54:53,651 --> 00:54:57,110

who want to learn and apply patient

techniques in their research?

950

00:54:57,110 --> 00:54:59,571

And especially in your field of

psychometrics.

951

00:55:00,291 --> 00:55:00,632

Yeah.

952

00:55:00,632 --> 00:55:05,575

I think, um, starting with an

understanding of sort of just the output,

953

00:55:05,575 --> 00:55:09,277

you know, the basics of if you're, if you

have data and if your responsibility is

954

00:55:09,277 --> 00:55:15,380

providing analysis for it, uh, finding

either a package or somebody else's

955

00:55:15,380 --> 00:55:18,342

program that makes the coding quick.

956

00:55:18,342 --> 00:55:22,204

So like you've mentioned linear

regression, if you use VRMS and R, you

957

00:55:22,204 --> 00:55:24,405

know, which will translate that into Stan.

958

00:55:24,554 --> 00:55:29,576

You can quickly go about getting a

Bayesian result fast.

959

00:55:29,576 --> 00:55:33,617

And I found that to me, the conceptual

consideration of what a posterior

960

00:55:33,617 --> 00:55:37,679

distribution is actually is less complex

than we think about when we think about

961

00:55:37,679 --> 00:55:41,040

all the things that we're drilled into in

the classical methods, like, you know,

962

00:55:41,040 --> 00:55:43,722

what, where does the standard error come

from and all this other, you know,

963

00:55:43,722 --> 00:55:48,143

asymptotic features in Bayes it's, it's

visible, like you can see a posterior

964

00:55:48,143 --> 00:55:51,305

distribution, you can plot it, you can,

you know, touch it, almost like touch it

965

00:55:51,305 --> 00:55:51,985

and feel it, right?

966

00:55:51,985 --> 00:55:53,425

It's right there in front of you.

967

00:55:53,658 --> 00:55:58,659

So for me, I think the thing I try to get

people to first is just to understand what

968

00:55:58,659 --> 00:55:59,919

the outputs are.

969

00:56:00,379 --> 00:56:03,260

Sort of what are the key parts of it.

970

00:56:03,260 --> 00:56:07,521

And then, you know, hopefully that gives

that mental representation of where that,

971

00:56:07,521 --> 00:56:08,662

where they're moving toward.

972

00:56:08,662 --> 00:56:11,182

And then at that point, start to add in

all the complexities.

973

00:56:11,182 --> 00:56:16,104

Um, but it is, I think it's, it's

incredibly challenging to try to, to teach

974

00:56:16,104 --> 00:56:21,265

Bayesian methods and I actually think the

further along a person goes, not learning

975

00:56:21,265 --> 00:56:22,890

the Bayesian version of things.

976

00:56:22,890 --> 00:56:27,531

Makes it even harder because now you have

all this well-established, um, can we say

977

00:56:27,531 --> 00:56:32,213

routines or statistics that you're used to

seeing that are not Bayesian, uh, that may

978

00:56:32,213 --> 00:56:36,235

or may not have a direct, um, analog in

the Bayes world.

979

00:56:36,235 --> 00:56:37,956

Um, but that may not be a bad thing.

980

00:56:37,956 --> 00:56:41,917

So, um, thinking about it, actually, I'm

going to take a step back here.

981

00:56:41,917 --> 00:56:47,220

Can conceptually, I think it's, this is

the challenge, um, we face in a program

982

00:56:47,220 --> 00:56:48,200

like I do right here.

983

00:56:48,200 --> 00:56:49,100

I'm working right now.

984

00:56:49,100 --> 00:56:52,201

I work with, um, nine other tenure track.

985

00:56:52,238 --> 00:56:55,680

or Tender to Tender Tech faculty, which is

a very large program.

986

00:56:55,680 --> 00:57:00,483

And we have a long-running curriculum, but

sort of the question I like to ask is,

987

00:57:00,483 --> 00:57:01,223

what do we do with Bayes?

988

00:57:01,223 --> 00:57:03,204

Do we have a parallel track in Bayes?

989

00:57:03,204 --> 00:57:05,145

Do we do Bayes in every class?

990

00:57:05,145 --> 00:57:07,807

Because that's a heavy lift for a lot of

people as well.

991

00:57:07,807 --> 00:57:12,630

Right now, it's, I teach the Bayes

classes, and occasionally some of my

992

00:57:12,630 --> 00:57:16,592

colleagues will put Bayesian statistics in

their classes, but it's tough.

993

00:57:16,652 --> 00:57:17,813

I think if I were

994

00:57:18,654 --> 00:57:21,776

you know, anointed myself king of how we

do all the curriculum.

995

00:57:21,776 --> 00:57:23,137

I don't know the answer I'd come to.

996

00:57:23,137 --> 00:57:24,838

I go back and forth each way.

997

00:57:24,838 --> 00:57:29,121

So, um, I would love to see what a

curriculum looks like where they only

998

00:57:29,121 --> 00:57:30,902

started with base and only kept it in

base.

999

00:57:30,902 --> 00:57:32,583

Cause I think that would be a lot of fun.

Speaker:

00:57:32,723 --> 00:57:35,665

Um, and the quit, the thought question I

asked myself that I don't have an answer

Speaker:

00:57:35,665 --> 00:57:40,488

for is would that be a better mechanism to

get students up to speed on the models

Speaker:

00:57:40,488 --> 00:57:45,251

they're using, then it would be in other

contexts and other classical contexts, I

Speaker:

00:57:45,251 --> 00:57:45,832

don't, I don't know.

Speaker:

00:57:45,832 --> 00:57:47,873

Yeah.

Speaker:

00:57:47,873 --> 00:57:48,398

Yay.

Speaker:

00:57:48,398 --> 00:57:49,258

Good point.

Speaker:

00:57:49,859 --> 00:57:51,199

Yeah, two things.

Speaker:

00:57:51,199 --> 00:57:54,742

First, King of Curriculum, amazing title.

Speaker:

00:57:54,822 --> 00:57:59,145

I think it should actually be renamed to

that title in all campuses around the

Speaker:

00:57:59,145 --> 00:57:59,945

world.

Speaker:

00:58:00,466 --> 00:58:03,728

The world's worst kingdom is the

curriculum.

Speaker:

00:58:03,728 --> 00:58:06,170

Yeah.

Speaker:

00:58:06,170 --> 00:58:07,731

I mean, that's really good.

Speaker:

00:58:07,731 --> 00:58:10,593

Like you're going to party, you know, and

so what are we doing on King of

Speaker:

00:58:10,593 --> 00:58:11,613

Curriculum?

Speaker:

00:58:12,494 --> 00:58:15,136

So long as the crown is on the head,

that's all that matters, right?

Speaker:

00:58:15,136 --> 00:58:17,477

That would drop some jaws for sure.

Speaker:

00:58:23,191 --> 00:58:29,173

And second, I definitely would like the

theory of the multiverse to be true,

Speaker:

00:58:29,193 --> 00:58:33,735

because that means in one of these

universes, there is at least one where

Speaker:

00:58:33,735 --> 00:58:36,135

Bayesian methods came first.

Speaker:

00:58:36,315 --> 00:58:42,197

And I am definitely curious to see what

that world looks like and see how...

Speaker:

00:58:42,657 --> 00:58:43,550

Yeah, what...

Speaker:

00:58:43,550 --> 00:58:47,912

What's that world where people were

actually exposed to patient methods first

Speaker:

00:58:47,933 --> 00:58:50,955

and maybe to frequency statistics later?

Speaker:

00:58:50,955 --> 00:58:56,398

Were they actually exposed to frequency

statistics later?

Speaker:

00:58:56,398 --> 00:58:57,619

That's the question.

Speaker:

00:58:57,739 --> 00:59:01,341

No, but yeah, jokes aside, I would be

definitely curious about that.

Speaker:

00:59:02,302 --> 00:59:07,266

Yeah, well, I don't know that I'll have

that experiment in my lifetime, but maybe

Speaker:

00:59:07,266 --> 00:59:09,727

like in a parallel universe somewhere.

Speaker:

00:59:15,010 --> 00:59:22,713

Before we close up the show, I'm wondering

if you have a personal anecdote or example

Speaker:

00:59:22,713 --> 00:59:27,315

of a challenging problem you encountered

in your research or teaching related to

Speaker:

00:59:27,315 --> 00:59:30,817

vision stats and how you were able to

navigate through it?

Speaker:

00:59:30,817 --> 00:59:30,917

Yeah.

Speaker:

00:59:30,917 --> 00:59:40,301

I mean, maybe it's too much in the weeds,

but that first experience I was in

Speaker:

00:59:40,301 --> 00:59:41,941

graduate school trying to learn.

Speaker:

00:59:45,151 --> 00:59:45,631

code.

Speaker:

00:59:45,631 --> 00:59:53,176

It was coding a correlation matrix of

tetrachore correlations.

Speaker:

00:59:53,176 --> 00:59:56,657

And that was incredibly difficult.

Speaker:

00:59:57,138 --> 01:00:02,021

One day, one of my colleagues, Bob Henson,

figured it out with the likelihood

Speaker:

01:00:02,021 --> 01:00:02,841

function and so forth.

Speaker:

01:00:02,841 --> 01:00:04,882

But that was the holdup that we had.

Speaker:

01:00:05,723 --> 01:00:09,910

And it's incredible because I say this

because again, we're not, I mentioned it.

Speaker:

01:00:09,910 --> 01:00:11,630

do a lot of my own package coding or

whatnot.

Speaker:

01:00:11,630 --> 01:00:16,473

But I think you see a similar phenomenon

if you misspecify something in your model

Speaker:

01:00:16,473 --> 01:00:20,996

in general and you get results and the

results are either all over the place or

Speaker:

01:00:20,996 --> 01:00:21,776

entire number line.

Speaker:

01:00:21,776 --> 01:00:24,858

For me, it was the correlations, posterior

distribution looked like a uniform

Speaker:

01:00:24,858 --> 01:00:26,339

distribution from negative one to one.

Speaker:

01:00:26,339 --> 01:00:28,980

That was, that's a bad thing to see,

right?

Speaker:

01:00:28,980 --> 01:00:35,884

So just the, the anecdote I have with this

is, it's less, I guess it's less like

Speaker:

01:00:35,884 --> 01:00:38,318

awesome, like when you're like, oh, Bayes

did this and then.

Speaker:

01:00:38,318 --> 01:00:42,339

couldn't have done it otherwise, but it's

more the perseverance that goes to

Speaker:

01:00:42,339 --> 01:00:47,981

sticking with the Bayesian side, which is,

um, Bayes also provides you the ability to

Speaker:

01:00:47,981 --> 01:00:53,083

check a little bit of your work to see if

it's completely gone sideways.

Speaker:

01:00:53,083 --> 01:00:53,404

Right.

Speaker:

01:00:53,404 --> 01:00:55,404

So, uh, you see a result like that.

Speaker:

01:00:55,404 --> 01:00:57,665

You have that healthy dose of skepticism.

Speaker:

01:00:57,865 --> 01:01:02,727

You start to investigate more in my case,

it took years, a couple of years of my

Speaker:

01:01:02,727 --> 01:01:08,169

life, uh, working in concert with other

people, uh, as grad students, but, um,

Speaker:

01:01:08,242 --> 01:01:10,544

was fixed, it was almost obvious that it

was.

Speaker:

01:01:10,544 --> 01:01:15,488

I mean, it was, you went from this uniform

distribution across negative one to one to

Speaker:

01:01:15,488 --> 01:01:18,010

something that looked very much like a

posterior distribution that we're used to

Speaker:

01:01:18,010 --> 01:01:21,192

seeing, send around a certain value of the

correlation.

Speaker:

01:01:21,373 --> 01:01:25,957

And again, it was, for us, it was figuring

out what the likelihood was, but for most

Speaker:

01:01:25,957 --> 01:01:27,738

packages, at least that's not a big deal.

Speaker:

01:01:27,738 --> 01:01:31,161

I think it's already specified in your

choice of model and prior.

Speaker:

01:01:31,201 --> 01:01:36,185

But at the same time, just remembering

that

Speaker:

01:01:36,270 --> 01:01:40,031

Uh, it's sort of the, the frustration part

of it, not making it work is actually

Speaker:

01:01:40,031 --> 01:01:40,791

really informative.

Speaker:

01:01:40,791 --> 01:01:44,472

Uh, you get that and you, you can build

and you can sort of check your work if you

Speaker:

01:01:44,472 --> 01:01:45,912

go forward analytically.

Speaker:

01:01:45,912 --> 01:01:50,153

I mean, not analytically brute force, the

sampling part, but that's sort of a check

Speaker:

01:01:50,153 --> 01:01:51,174

on your work.

Speaker:

01:01:51,794 --> 01:01:57,235

Trying to say, so not a great example, not

a super inspiring example, but, um, more

Speaker:

01:01:57,235 --> 01:01:59,536

perseverance pays off in days and in life.

Speaker:

01:01:59,536 --> 01:02:01,617

So it's sort of the analog that I get from

it.

Speaker:

01:02:01,617 --> 01:02:03,037

Yeah.

Speaker:

01:02:03,037 --> 01:02:04,377

Yeah, no, for sure.

Speaker:

01:02:04,377 --> 01:02:05,297

I mean, um,

Speaker:

01:02:06,066 --> 01:02:11,950

is perseverance is so important because

you're definitely going to encounter

Speaker:

01:02:12,091 --> 01:02:12,411

issues.

Speaker:

01:02:12,411 --> 01:02:18,336

I mean, none of your models is going to

work as you thought it would.

Speaker:

01:02:18,336 --> 01:02:23,400

So if you don't have that drive and that

passion for the thing that you're

Speaker:

01:02:23,400 --> 01:02:30,466

standing, it's going to be extremely hard

to just get it through the finish line

Speaker:

01:02:30,466 --> 01:02:32,267

because it's not going to be easy.

Speaker:

01:02:32,267 --> 01:02:35,186

So, you know, it's like choosing a new

sport.

Speaker:

01:02:35,186 --> 01:02:40,867

If you don't like what the sport is all

about, you're not going to stick with it

Speaker:

01:02:40,867 --> 01:02:42,788

because it's going to be hard.

Speaker:

01:02:42,788 --> 01:02:51,370

So that perseverance, I would say, come

from your curiosity and your passion for

Speaker:

01:02:51,510 --> 01:02:54,351

your field and the methods you're using.

Speaker:

01:02:54,851 --> 01:02:57,592

And the other thing I was going to add,

this is tangential, but let me just add

Speaker:

01:02:57,592 --> 01:03:01,553

it, you have the chance to go visit Bay's

grave in London, take it.

Speaker:

01:03:01,553 --> 01:03:03,570

I had to do that last summer.

Speaker:

01:03:03,570 --> 01:03:06,891

I just, I was in London, I had my children

with me and we all picked some spot we

Speaker:

01:03:06,891 --> 01:03:07,851

wanted to go to.

Speaker:

01:03:07,851 --> 01:03:12,373

And I was like, I'm going to go find and

take a picture in front of Bayes grave.

Speaker:

01:03:12,373 --> 01:03:14,234

And I sort of brought up an interesting

question.

Speaker:

01:03:14,234 --> 01:03:18,136

Like I don't know the etiquette of taking

photographs in front of a deceased grave

Speaker:

01:03:18,136 --> 01:03:18,756

site.

Speaker:

01:03:18,756 --> 01:03:20,736

This is at least providing it.

Speaker:

01:03:21,417 --> 01:03:25,298

But then ironically, as you're sitting

there, as I was sitting there on the tube,

Speaker:

01:03:25,499 --> 01:03:29,700

leaving, I sat next to a woman and she had

Bayes theorem on her shirt.

Speaker:

01:03:29,700 --> 01:03:31,681

It was the Bayes School of Economics.

Speaker:

01:03:31,681 --> 01:03:32,874

So something like this.

Speaker:

01:03:32,874 --> 01:03:36,757

in London, I was like, it was like, okay,

I have reached the Mecca.

Speaker:

01:03:36,757 --> 01:03:41,722

Like the perseverance led to like, like a

trip, you know, my own version of the trip

Speaker:

01:03:41,722 --> 01:03:42,983

to, to London.

Speaker:

01:03:42,983 --> 01:03:45,465

Uh, but definitely, uh, definitely worth

the time to go.

Speaker:

01:03:45,465 --> 01:03:49,669

If you want to be surrounded, uh, once you

reach that, that level of perseverance,

Speaker:

01:03:49,669 --> 01:03:52,271

uh, you're part of the club and then you

can do things like that.

Speaker:

01:03:52,711 --> 01:03:56,475

Fine vacations around, you know, holidays

around base, base graves.

Speaker:

01:03:56,475 --> 01:03:59,117

Yeah.

Speaker:

01:03:59,117 --> 01:03:59,877

I mean.

Speaker:

01:03:59,970 --> 01:04:02,811

I am definitely gonna do that.

Speaker:

01:04:02,811 --> 01:04:06,732

Thank you very much for giving me another

idea of a nerd holiday.

Speaker:

01:04:06,732 --> 01:04:10,894

My girlfriend is gonna hate me, but she

always wanted to visit London, so you

Speaker:

01:04:10,894 --> 01:04:12,755

know, that's gonna be my bait.

Speaker:

01:04:13,796 --> 01:04:17,417

It's not bad to get to, it's off of Old

Street, you know, actually well marked.

Speaker:

01:04:17,417 --> 01:04:21,059

I mean the grave site's a little

weathered, but it's in a good spot, a good

Speaker:

01:04:21,059 --> 01:04:25,341

part of town, so you know, not really

heavily touristy, amazingly.

Speaker:

01:04:25,341 --> 01:04:26,401

Oh yeah, I'm guessing.

Speaker:

01:04:26,401 --> 01:04:27,381

But you know.

Speaker:

01:04:28,314 --> 01:04:30,355

I am guessing that's the good thing.

Speaker:

01:04:31,015 --> 01:04:34,697

Yeah, no, I already know how I'm gonna ask

her.

Speaker:

01:04:34,697 --> 01:04:36,238

Honey, when I go to London?

Speaker:

01:04:36,278 --> 01:04:36,898

Perfect.

Speaker:

01:04:36,898 --> 01:04:37,599

Let's go to Bay's.

Speaker:

01:04:37,599 --> 01:04:38,579

Let's go check out Bay's Grave.

Speaker:

01:04:38,579 --> 01:04:42,362

Yeah, I mean, that's perfect.

Speaker:

01:04:42,362 --> 01:04:43,882

That's amazing.

Speaker:

01:04:43,882 --> 01:04:48,045

So say, I mean, you should send me that

picture and that should be your picture

Speaker:

01:04:48,045 --> 01:04:49,746

for these episodes.

Speaker:

01:04:49,746 --> 01:04:55,409

I always take a picture from guests to

illustrate the episode icon, but you

Speaker:

01:04:55,409 --> 01:04:57,130

definitely need that.

Speaker:

01:04:57,130 --> 01:04:58,190

picture for your icon.

Speaker:

01:04:58,190 --> 01:04:58,590

I can do that.

Speaker:

01:04:58,590 --> 01:05:00,211

I'll be happy to.

Speaker:

01:05:00,211 --> 01:05:01,212

Yeah.

Speaker:

01:05:01,492 --> 01:05:02,652

Awesome.

Speaker:

01:05:03,113 --> 01:05:03,573

Definitely.

Speaker:

01:05:03,573 --> 01:05:08,856

So before asking you the last two

questions, I'm just curious how you see

Speaker:

01:05:09,036 --> 01:05:15,620

the future of patient stats in the context

of psychological sciences and

Speaker:

01:05:15,620 --> 01:05:16,740

psychometrics.

Speaker:

01:05:16,981 --> 01:05:22,684

And what are some exciting avenues for

research and application that you envision

Speaker:

01:05:22,684 --> 01:05:25,705

in the coming years or that you would

really like to see?

Speaker:

01:05:26,494 --> 01:05:28,754

Oh, that's a great question.

Speaker:

01:05:28,754 --> 01:05:29,134

Terrible.

Speaker:

01:05:29,134 --> 01:05:37,357

So I, you know, interestingly, in

psychology, you know, quantitative

Speaker:

01:05:37,357 --> 01:05:41,338

psychology sort of been on a downhill

swing for, I don't know, 5060 years,

Speaker:

01:05:41,338 --> 01:05:44,278

there's fewer and fewer programs, at least

in the United States, where people are

Speaker:

01:05:44,278 --> 01:05:45,139

training.

Speaker:

01:05:45,219 --> 01:05:49,020

But despite that, I feel like the use of

Bayesian statistics is up in a lot of a

Speaker:

01:05:49,020 --> 01:05:50,260

lot of different other areas.

Speaker:

01:05:50,260 --> 01:05:55,382

And I think that I think that affords a

bit.

Speaker:

01:05:55,382 --> 01:05:56,922

better model-based science.

Speaker:

01:05:56,922 --> 01:06:00,384

So you have to specify a model, you have

to model in mind, and then you go and do

Speaker:

01:06:00,384 --> 01:06:00,584

that.

Speaker:

01:06:00,584 --> 01:06:03,705

I think that benefit makes the science

much better.

Speaker:

01:06:03,705 --> 01:06:07,607

You're not just using sort of what's

always been done.

Speaker:

01:06:07,607 --> 01:06:10,848

You can sort of push the envelope

methodologically a bit more.

Speaker:

01:06:10,848 --> 01:06:14,109

And I think that that, and Bayesian

statistics in one way, another benefit of

Speaker:

01:06:14,109 --> 01:06:18,291

them is now you can code an algorithm that

likely will work without having to know,

Speaker:

01:06:18,291 --> 01:06:21,952

like you said, all of the underpinnings,

the technical side of things, you can use

Speaker:

01:06:21,952 --> 01:06:24,453

an existing package to do so.

Speaker:

01:06:25,670 --> 01:06:29,751

I like to say that that's going to

continue to make science a better

Speaker:

01:06:29,751 --> 01:06:30,812

practice.

Speaker:

01:06:31,332 --> 01:06:39,635

I think the fear that I have is sort of

the sea of the large language model-based

Speaker:

01:06:39,676 --> 01:06:43,137

version of what we're doing in machine

learning, artificial intelligence.

Speaker:

01:06:43,137 --> 01:06:49,360

But I will be interested to see how we

incorporate a lot of the Bayesian ideas,

Speaker:

01:06:49,360 --> 01:06:51,801

Bayesian methods into that as well.

Speaker:

01:06:51,801 --> 01:06:53,581

I think that there's potential.

Speaker:

01:06:53,846 --> 01:06:57,527

Clearly, people are doing this, I mean,

that's what runs a lot of what is

Speaker:

01:06:57,527 --> 01:06:58,608

happening anyway.

Speaker:

01:06:58,608 --> 01:07:00,948

So I look forward to seeing that as well.

Speaker:

01:07:01,349 --> 01:07:07,351

So I get a sense that what we're talking

about is really what may be the foundation

Speaker:

01:07:07,351 --> 01:07:08,872

for what the future will be.

Speaker:

01:07:08,872 --> 01:07:12,033

I mean, maybe we will, maybe instead of

that parallel universe, if we could come

Speaker:

01:07:12,033 --> 01:07:16,615

back or go into the future just in our own

universe in 50 years, maybe what we will

Speaker:

01:07:16,615 --> 01:07:19,356

see is curriculum entirely on Bayesian

methods.

Speaker:

01:07:19,356 --> 01:07:21,966

And from, you know, I just looked at your.

Speaker:

01:07:21,966 --> 01:07:26,027

topic list you had recently talking about

variational inference and so forth.

Speaker:

01:07:26,387 --> 01:07:32,910

The use of that in very large models

themselves, I think that is very important

Speaker:

01:07:32,910 --> 01:07:33,250

stuff.

Speaker:

01:07:33,250 --> 01:07:37,792

So it may just be the thing that crowds

out everything else, but that's

Speaker:

01:07:37,792 --> 01:07:42,114

speculative and I don't make a living

making prediction, unfortunately.

Speaker:

01:07:42,114 --> 01:07:43,874

So that's the best I can do.

Speaker:

01:07:43,874 --> 01:07:45,155

Yeah.

Speaker:

01:07:45,155 --> 01:07:46,015

Yeah, yeah.

Speaker:

01:07:46,015 --> 01:07:48,756

I mean, that's also more of a wishlist

question.

Speaker:

01:07:48,756 --> 01:07:50,297

So that's all good.

Speaker:

01:07:50,757 --> 01:07:51,217

Yeah.

Speaker:

01:07:51,217 --> 01:07:51,826

Awesome.

Speaker:

01:07:51,826 --> 01:07:53,847

Well, John, amazing.

Speaker:

01:07:54,708 --> 01:07:55,888

I learned a lot.

Speaker:

01:07:55,908 --> 01:07:57,309

We covered a lot of topics.

Speaker:

01:07:57,309 --> 01:07:58,670

I'm really happy.

Speaker:

01:07:59,531 --> 01:08:04,254

But of course, before letting you go, I'm

going to ask you the last two questions I

Speaker:

01:08:04,254 --> 01:08:06,295

ask every guest at the end of the show.

Speaker:

01:08:06,836 --> 01:08:10,778

Number one, you had unlimited time and

resources.

Speaker:

01:08:10,778 --> 01:08:14,001

Which problem would you try to solve?

Speaker:

01:08:14,001 --> 01:08:18,343

Well, I would be trying to figure out how

we know what a student knows every day of

Speaker:

01:08:18,343 --> 01:08:21,685

the year so that we can best teach them

where to go next.

Speaker:

01:08:22,062 --> 01:08:23,982

That would be it.

Speaker:

01:08:23,982 --> 01:08:29,285

Right now, there's not only the problem of

the technical issues of estimation,

Speaker:

01:08:29,285 --> 01:08:33,186

there's also the problem of how do we best

assess them, how much time do they spend

Speaker:

01:08:33,186 --> 01:08:34,387

doing it and so forth.

Speaker:

01:08:34,387 --> 01:08:39,429

That to me is what I would spend most of

my time on.

Speaker:

01:08:39,429 --> 01:08:41,510

That sounds like a good project.

Speaker:

01:08:41,510 --> 01:08:42,390

I love it.

Speaker:

01:08:43,510 --> 01:08:49,633

And second question, if you could have

dinner with any great scientific mind that

Speaker:

01:08:49,633 --> 01:08:51,173

life are fictional.

Speaker:

01:08:51,234 --> 01:08:52,914

who did be.

Speaker:

01:08:52,914 --> 01:08:53,294

All right.

Speaker:

01:08:53,294 --> 01:08:55,595

I got a really obscure choice, right?

Speaker:

01:08:55,595 --> 01:08:59,016

It's not like I'm picking Einstein or

anything.

Speaker:

01:08:59,016 --> 01:09:01,656

I really, I have like two actually, I've

sort of debated.

Speaker:

01:09:01,656 --> 01:09:06,238

One is economist Paul Krugman, who writes

for the New York Times, works at City

Speaker:

01:09:06,238 --> 01:09:07,418

University of New York now.

Speaker:

01:09:07,418 --> 01:09:09,299

You know, Nobel laureate.

Speaker:

01:09:09,299 --> 01:09:13,720

Loved his work, loved his understanding

of, for the interplay between model and

Speaker:

01:09:13,720 --> 01:09:18,121

data and understanding is fantastic.

Speaker:

01:09:18,341 --> 01:09:20,282

So I would just.

Speaker:

01:09:20,282 --> 01:09:23,204

sit there and just have to listen to

everything you had to say, I think.

Speaker:

01:09:23,224 --> 01:09:26,767

The other is there's a, again, obscure

thing.

Speaker:

01:09:26,767 --> 01:09:31,151

One of my things I'm fascinated by is

weather and weather forecasting.

Speaker:

01:09:31,151 --> 01:09:35,033

Uh, if you know, I'm in education or

psychological measurement.

Speaker:

01:09:35,234 --> 01:09:38,457

Uh, and there's a guy who started the

company called the weather underground.

Speaker:

01:09:38,457 --> 01:09:39,738

His name is Jeff Masters.

Speaker:

01:09:39,738 --> 01:09:43,941

Uh, you can read his work on a blog at

Yale these days, climate connections,

Speaker:

01:09:43,941 --> 01:09:45,262

something along those lines.

Speaker:

01:09:45,262 --> 01:09:49,385

Anyway, since sold the company, but he's

fascinating about modeling, you know,

Speaker:

01:09:49,546 --> 01:09:52,148

Right now we're in the peak of hurricane

season in the United States.

Speaker:

01:09:52,148 --> 01:09:56,532

We see these storms coming off of Africa

or spinning up everywhere and sort of the

Speaker:

01:09:56,532 --> 01:10:01,416

interplay between, unfortunately, the

climate change and then other atmospheric

Speaker:

01:10:01,416 --> 01:10:01,996

dynamics.

Speaker:

01:10:01,996 --> 01:10:07,060

This just makes for an incredibly complex

system that's just fascinating and how

Speaker:

01:10:07,201 --> 01:10:08,742

science approaches prediction there.

Speaker:

01:10:08,742 --> 01:10:10,404

So I find that to be great.

Speaker:

01:10:10,404 --> 01:10:11,464

But those are the two.

Speaker:

01:10:11,464 --> 01:10:14,107

I had to think a lot about that because

there's so many choices, but those two

Speaker:

01:10:14,107 --> 01:10:17,769

people are the ones I read the most,

certainly when it's not just in my field.

Speaker:

01:10:18,942 --> 01:10:19,702

Nice.

Speaker:

01:10:19,702 --> 01:10:21,983

Yeah, sounds fascinating.

Speaker:

01:10:22,063 --> 01:10:24,505

And weather forecasting is definitely

incredible.

Speaker:

01:10:25,445 --> 01:10:30,188

Also, because the great thing is you have

feedback every day.

Speaker:

01:10:30,828 --> 01:10:33,010

So that's really cool.

Speaker:

01:10:33,010 --> 01:10:34,070

You can improve your predictions.

Speaker:

01:10:34,070 --> 01:10:35,751

Like the missing data problem.

Speaker:

01:10:35,992 --> 01:10:37,973

You can't sample every part of the

atmosphere.

Speaker:

01:10:37,973 --> 01:10:41,895

So how do you incorporate that into your

analysis as well?

Speaker:

01:10:42,615 --> 01:10:43,856

No, that's incredible.

Speaker:

01:10:43,856 --> 01:10:45,697

Multiple average models and stuff.

Speaker:

01:10:45,697 --> 01:10:46,646

Anyway, yeah.

Speaker:

01:10:46,646 --> 01:10:51,529

Yeah, that's also a testimony to the power

of modeling and parsimony, you know, where

Speaker:

01:10:51,529 --> 01:10:56,533

it's like, because I worked a lot on

electoral forecasting models and, you

Speaker:

01:10:56,533 --> 01:11:01,937

know, classic way people dismiss models in

these areas.

Speaker:

01:11:01,937 --> 01:11:06,340

Well, you cannot really predict what

people are going to do at an individual

Speaker:

01:11:06,340 --> 01:11:08,061

level, which is true.

Speaker:

01:11:08,061 --> 01:11:11,624

I mean, you cannot, people have free will,

you know, so you cannot predict at an

Speaker:

01:11:11,624 --> 01:11:14,705

individual level what they are going to

do, but you can.

Speaker:

01:11:14,766 --> 01:11:19,249

quite reliably predict what masses are

going to do.

Speaker:

01:11:19,329 --> 01:11:27,836

Yeah, basically, where the aggregation of

individual points, you can actually kind

Speaker:

01:11:27,836 --> 01:11:30,077

of reliably do it.

Speaker:

01:11:30,939 --> 01:11:35,002

And so the power of modeling here where

you get something that, yeah, you know,

Speaker:

01:11:35,002 --> 01:11:36,143

it's not good.

Speaker:

01:11:36,143 --> 01:11:44,829

It's, you know, the model is wrong, but it

works because it simplifies

Speaker:

01:11:45,378 --> 01:11:51,541

things, but doesn't simplify them to a

point where it doesn't make sense anymore.

Speaker:

01:11:51,801 --> 01:11:55,783

Kind of like the standard model in

physics, where we know it doesn't work, it

Speaker:

01:11:55,783 --> 01:12:02,027

breaks at some point, but it does a pretty

good job of predicting a lot of phenomena

Speaker:

01:12:02,027 --> 01:12:02,527

and we observe.

Speaker:

01:12:02,527 --> 01:12:04,988

So, do you prefer that?

Speaker:

01:12:04,988 --> 01:12:09,431

Is it free will or is it random error?

Speaker:

01:12:09,431 --> 01:12:11,852

Well, you have to come back for another

episode on that because otherwise, yes.

Speaker:

01:12:11,852 --> 01:12:13,893

That's a good one.

Speaker:

01:12:16,547 --> 01:12:16,787

Good point.

Speaker:

01:12:16,787 --> 01:12:16,888

Nice.

Speaker:

01:12:16,888 --> 01:12:22,172

Well, Jonathan, thank you so much for your

time.

Speaker:

01:12:22,172 --> 01:12:26,835

As usual, I will put resources and a link

to your website in the show notes for

Speaker:

01:12:26,835 --> 01:12:28,336

those who want to dig deeper.

Speaker:

01:12:28,436 --> 01:12:31,819

Thank you again, Jonathan, for taking the

time and being on this show.

Speaker:

01:12:32,440 --> 01:12:32,940

Happy to be here.

Speaker:

01:12:32,940 --> 01:12:34,521

Thanks for the opportunity.

Speaker:

01:12:34,521 --> 01:12:41,947

It was a pleasure to speak with you and I

hope it makes sense for a lot of people.

Speaker:

01:12:41,947 --> 01:12:43,488

Appreciate your time.

Previous post
Next post