Week 5 Ethics discussion

Watch the video about Gino Francesca and Dan Ariely presented in above in Canvas, then answer the questions below.

What exactly did Francesca and Ariely do that created the scandal?

Why was it so shocking that these two researchers in particular did this?

What was the ultimate outcome?

Be sure to post your opinions and respond to at least two other students by February 15. You do not have to provide peer-reviewed sources for your post, but if you do reference outside material, then you should cite it in APA 7 format.

I have several other videos on this case if anyone would like me to rip them into a file and post them. Or you can easily look it up for yourself. Judo has several videos on this particular topic.

here is a transcript of the Video:

0:00

Academia is broken universities are

0:03

broken the way that academic research is

0:05

published is broken that’s the message

0:07

that’s come through loud and clear over

0:09

the last few weeks thanks to three

0:11

articles concerning the research of

0:13

Francesca Geno if you don’t know what

0:15

I’m talking about Let Me Explain

0:16

Francesca Geno is a professor of

0:18

Behavioral Science at Harvard University

0:20

she is extremely well known in the field

0:23

I’ve talked about her research to

0:25

clients before I’ve recommended books on

0:27

this channel to you guys that use her

0:28

work as a key reference I’ve used her

0:30

research before as references in my own

0:32

essays and work that I did at University

0:34

when it comes to academic Fame Francesca

0:37

Gino is up there as you would expect

0:38

from someone who is a professor at

0:41

Harvard however the reason why she’s so

0:43

well known is because her research tends

0:44

to bring out a lot of very surprising

0:47

findings now some people just think this

0:49

research is cool and don’t think much

0:50

more about it but a lot of people in the

0:52

industry have been quite skeptical of

0:54

Francesco Gino and her work because her

0:56

results just seem a little bit too good

0:58

her hypotheses are really wacky but yeah

1:00

they always seem to be proved correct

1:02

the effect sizes from her studies seem

1:04

to be really large and her statistical

1:06

significance just seem a little bit too

1:08

significant so while some of us have

1:10

been skeptical of her work for a while

1:11

nobody has taken the time to actually

1:13

investigate her research and go into her

1:15

data to see if they can find anything

1:16

fishy

1:18

until now these three guys Yuri Joe and

1:20

laif are also professors of Behavioral

1:22

Science and other related subjects from

1:24

different universities across the world

1:26

and they took it upon themselves to

1:28

investigate Francesca Gino and her data

1:30

to see if there was anything fishy going

1:32

on and spoiler alert they found a lot of

1:34

fishy stuff in the data and that’s what

1:36

the three articles that they released

1:37

are talking about each article relates

1:39

to a different study by Francesco Gino

1:41

and in this video I’m going to be taking

1:43

you through each one the results of

1:45

their investigation are shocking damning

1:47

for Francesca Gino but I think they

1:49

speak even louder volumes about the

1:50

state of Academia in general and that’s

1:52

what I’m going to be concluding on at

1:53

the end of this video so without further

1:55

Ado let’s jump into the first study so

1:57

this first article is called cluster

1:59

fake and it’s referring to a paper

2:00

written by Gino in 2012 along with her

2:02

collaborators Shu Nina Mazar Dan arieli

2:05

and Max baseman given the fact that I

2:07

know the first names of all of those

2:09

researchers with the exception of Shu

2:10

should tell you that all of these

2:12

researchers are very well-known people

2:13

in the field of Behavioral Science so in

2:15

this study they were trying to get

2:17

participants to be more honest and the

2:19

hypothesis was that if you put an

2:21

honesty pledge at the top of a form

2:22

that’ll make people more honest when

2:24

they then fill out the rest of the form

2:26

so all of the studies in this paper by

2:28

these authors were looking at this idea

2:30

that if you put an honesty pledge at the

2:32

top of a form people will be more honest

2:34

than if you put the honesty pledge at

2:35

the bottom of a form now the first study

2:37

in this paper was led by Francesca Geno

2:39

our protagonist so in this study

2:41

students were brought into a lab to

2:43

complete 20 math puzzles in five minutes

2:45

the students were told that they would

2:46

be paid one dollar for each math puzzle

2:49

they solved correctly and the way that

2:50

this worked is that when students walked

2:52

into the room there were two pieces of

2:53

paper they had their work paper and

2:56

their report paper so on the work paper

2:57

they write down their workings for the

2:59

math questions and of course their

3:00

answers and then on the report paper

3:02

they would then have to report how many

3:04

answers they got correctly and therefore

3:05

how much they should get paid the

3:07

students were then told that before

3:08

handing in their report paper to the

3:10

researchers and getting paid that they

3:12

should shred their original work paper

3:13

the idea behind this is that by

3:15

shredding their work paper there’s then

3:17

a stronger incentive for them to cheat

3:19

on the report paper and lie about how

3:21

many answers they got correct since the

3:23

researchers in theory should never know

3:24

how many answers they got right on the

3:26

work paper but what the students didn’t

3:28

know was that the shredder at the back

3:30

of the room was not a normal Shredder

3:31

what the people in the experiment don’t

3:33

know is that the shredder has been fixed

3:36

so the shredder only showed the sides of

3:38

the page but the main body of the page

3:40

remains intact now in order to test the

3:43

hypothesis of the researchers on the

3:45

reporting paper the participants were

3:47

split into two groups half of them had

3:49

an honesty pledge at the top of the

3:50

paper and half the planet honesty

3:52

pledged at the bottom of the paper with

3:54

the idea being of course that those who

3:56

sign the honesty pledge at the top would

3:57

then cheat less going forward so what

4:00

was the result well the result showed a

4:02

massive effect from this simple

4:04

intervention according to what was

4:05

published in the study originally for

4:07

the students who silently honestly

4:08

pledge at the top of the form only 37

4:11

percent of them lied but when students

4:12

signed at the bottom of the form 79 of

4:15

students lied this is a massive effect

4:18

size that the researchers are reporting

4:19

and as a result of that this study

4:21

gained a lot of public attention and I

4:24

have talked about it with many people in

4:25

the past before because it is so

4:27

surprising but that’s why these

4:29

Vigilantes were suspicious the results

4:31

just seem a bit too good can it really

4:33

be the case that simply moving an

4:35

honestly pledge from the bottom to the

4:36

top of a form can have such a dramatic

4:39

effect on the amount of cheating that

4:40

happens it seems pretty unlikely so our

4:43

Vigilantes managed to Source the

4:45

original data set that was published by

4:47

the authors of the study and when they

4:49

looked into the data it just seemed a

4:51

little bit fishy if you look at this

4:53

table and specifically look at the left

4:54

hand column the P hash column this is

4:57

referring to participant ID this is the

5:00

unique ID given to each participant in a

5:02

study and as is highlighted in yellow

5:04

there are some weird anomalies in the

5:06

way that this data has been sorted

5:07

because when you look at this data it

5:09

seems obvious that this has been sorted

5:10

by first the condition so all of

5:12

condition 1 are together then all of

5:14

condition two are together and then in

5:16

ascending order of the participant ID

5:18

which means that the numbers should

5:20

consistently get bigger as you go down

5:22

the line and there should be no

5:23

duplicates remember each participant has

5:25

a unique ID so when you look at this

5:27

data it’s a bit weird we’ve got 249s

5:29

here that’s a duplicate that should

5:31

never happen and then at the end of the

5:33

condition one set of participants you

5:35

have participant 51 coming after 95 then

5:38

12 then 101 like that sequence doesn’t

5:40

make any sense and similarly when you

5:42

get to condition two we start with 7

5:44

then 91 then 52 then all the way back

5:46

down to 5 again these entries in the

5:48

data set look suspicious they look like

5:50

they’re out of sequence which suggests

5:52

that somebody maybe has tampered with

5:54

them so our Vigilantes are suspicious of

5:56

these rows so then you have to ask the

5:58

question why would the researchers want

6:00

to tamper with the data well it’s

6:02

because they would want to show a bigger

6:04

effect than those actually seen in the

6:06

real data the more dramatic the effect

6:09

of the intervention is the more

6:10

surprising the result of the study is

6:12

and therefore the more likely it is to

6:14

get published in a top journal the more

6:16

likely it is that this will make a lot

6:17

of press headlines that they will get

6:18

lots of interviews and work off the back

6:20

of it and so there’s a strong incentive

6:22

for the researchers to fudge the data a

6:24

little bit make the effect seem larger

6:26

than it really is and so that’s what our

6:28

Vigilantes were looking for they wanted

6:30

to see if these suspicious rows in the

6:32

data set showed a bigger effect than the

6:35

normal data that wasn’t suspicious and

6:37

sure enough that’s exactly what they

6:39

found if you look at this graph the red

6:41

circles with the cross show the

6:42

suspicious data and the blue dots show

6:44

the unsuspicious data and as you can see

6:46

the circles with the red crosses are the

6:48

most extreme ones meaning that these few

6:50

data points are inflating the effect

6:52

size now the article goes on to show how

6:55

our Vigilantes did some very clever work

6:57

to unpack the Excel file that this data

6:59

was stored in and they were able to show

7:01

quite clearly that these suspicious rows

7:02

were manually resorted in the data set I

7:05

won’t go into it on this video because

7:07

it’s quite technical but I’ll have a

7:08

link to all of these articles in the

7:10

description if you want to read them in

7:11

full but as you’ll soon see this theme

7:13

of suspicious data and then there’s data

7:16

showing extremely strong effect sizes

7:18

will be a recurring pattern so let’s

7:19

move on to study two now this second

7:22

article is called my class year is

7:24

Harvard and you’ll see why in a second

7:26

they’re looking at a study from 2015

7:27

written by Francesca Gino as well as

7:30

kuchaki and golinski again two fairly

7:32

well-known researchers in the field now

7:34

the hypothesis for this study in my

7:36

opinion pretty stupid the hypothesis is

7:39

that if you argue against something that

7:41

you really believe in that makes you

7:42

feel dirty which then increases your

7:45

desire for cleansing products which is

7:48

kind of silly in my opinion but

7:50

nevertheless this is what they were

7:51

researching so this study was done at

7:53

Harvard University with almost 500

7:56

students and what they asked the

7:57

participants to do was the following so

7:59

students of Harvard University were

8:01

brought into the lab and then asked how

8:02

they felt about this thing called the

8:03

queue guide I don’t really know what the

8:05

cue guide is but apparently it’s a Hot

8:06

Topic at Harvard and it’s very

8:08

controversial some people are for it

8:09

some people are against it so when they

8:11

were brought to the lab they were asked

8:12

how do you feel about the queue guide

8:13

and they either said they were for or

8:15

against it and then the participants

8:16

were split into two groups half the

8:18

participants were asked to write an

8:20

essay supporting the view that they just

8:22

gave so if they said I’m for the queue

8:24

guide they had to then write an essay

8:25

explaining why they were for the queue

8:27

guide but then half the participants

8:28

were asked to write an essay arguing

8:30

opposite to the point that they just

8:31

gave so if they said I’m for the queue

8:33

guide they would then have to write an

8:35

essay explaining why they should be

8:36

against the queue guide again the idea

8:39

being that those who are writing an

8:40

essay against what they actually believe

8:42

in would make them feel dirty because

8:44

after they’d written this essay they

8:45

were then shown five different cleansing

8:47

products and the participants in the

8:49

study had to rate how desirable they

8:51

felt these cleansing products were on a

8:53

scale of one to seven with one being

8:55

completely undesirable and seven being

8:58

completely desirable and again the

9:00

authors found a strong effect you can

9:02

see here that the p-value is less than

9:05

0.01. and for those of you who haven’t

9:07

had any academic training and statistics

9:09

basically when you’re doing a study like

9:11

this you’re looking for a p-value that’s

9:12

less than 0.05 that’s the industry

9:14

standard if it’s less than 0.05 you say

9:17

yes I’m confident that the effect that

9:18

I’m seeing is caused by the manipulation

9:20

that I just did so less than 0.1 is an

9:24

extremely strong effect you’re basically

9:26

100 confident that what you’re seeing in

9:29

the data is caused by the manipulation

9:31

that you did so once again our

9:33

Vigilantes are suspicious of this very

9:35

strong effect size so the managed to

9:37

Source the data online and do a little

9:39

bit of investigating and what they find

9:41

are some weird anomalies in the kind of

9:43

demographic data that the participants

9:44

have to give when they enter the study

9:46

and this is very common in psychological

9:47

studies that participants have to give a

9:49

little bit of demographic data about

9:50

themselves which gives the researchers a

9:52

little bit more flexibility about how

9:53

they cut up the data later on so in this

9:55

particular study the participants were

9:57

asked a number of demographic questions

9:58

including their age their gender and

10:00

then number six was what year in school

10:02

they were now the way this question is

10:04

structured isn’t very good in my opinion

10:05

in terms of research design but

10:07

nevertheless there are a number of

10:08

acceptable answers that you can give to

10:10

you in school because Harvard is an

10:12

American School you might say I’m a

10:14

senior right which is a common thing or

10:15

a sophomore you might write the year

10:17

that you’re supposed to graduate 2015

10:19

2016 Etc or you might indicate a one a

10:22

two a three or four or a five to

10:23

indicate how many years of school that

10:25

you’ve been in there these are all

10:26

different answers but they’re all

10:27

acceptable and make sense in the context

10:29

of being asked what year in school are

10:31

you and so when our Vigilantes go into

10:33

the data that’s exactly what they saw in

10:34

this column a range of different answers

10:36

that were all acceptable all except for

10:38

one there were 20 entries in this data

10:40

set where the answer to the question

10:41

what year in school are you was Harvard

10:45

that doesn’t make any sense what year in

10:47

school are you Harvard

10:49

what right that doesn’t make any sense

10:51

and the other thing that was suspicious

10:52

about these Harvard entries is that they

10:54

were all grouped together within 35 rows

10:56

again this was a data set of nearly 500

10:58

different participants and yet all of

11:00

these weird Harvard answers were within

11:02

35 rows so once again our Vigilantes

11:05

treat these Harvard answers as

11:07

suspicious data entries they mark them

11:09

in red circles with crosses and as you

11:11

can see the ones that are suspicious are

11:14

again the most extreme answers

11:16

supporting the hypothesis of the

11:18

researchers with the exception of this

11:20

one but come on it’s most suspicious

11:22

when you look at the ones on argued

11:24

other side so these are the people who

11:26

wrote an essay arguing against what they

11:28

didn’t believe in and therefore were

11:30

supposed to feel more dirty and find

11:31

cleansing products more appealing all of

11:33

these suspicious entries on that side of

11:35

the manipulation went for seven that

11:37

they found all of the cleaning products

11:39

completely desirable and so what are

11:41

Vigilantes go on to say is that these

11:43

were just the 20 entries in the data set

11:45

that looked suspicious because of this

11:46

Harvard answer to the demographic

11:48

question but who’s say that the other

11:50

data in the data set was not also

11:51

tampered with but just they were more

11:53

careful when they filled in this column

11:54

and didn’t put Harvard since it seems

11:56

pretty clear that at least these 20

11:58

entries were manipulated and tampered

12:00

with some way it probably means that

12:01

there are other entries within this data

12:03

set that were also tampered with are you

12:04

shocked yet I hope you are but it’s

12:06

about to get worse because there’s a

12:07

third article to do with Francesca Gino

12:09

so this third article was released

12:11

literally yesterday the day before I’m

12:12

filming this video and it’s called the

12:14

cheaters are out of order this is

12:16

written by Francesca Gino and a guy

12:18

called wiltermuth I don’t know

12:19

wiltermuth but again I find it

12:21

incredibly ironic that all of this

12:23

cheating and fake data is being

12:25

conducted by researchers who are

12:27

studying the science of honesty it is

12:29

incredibly ironic so in this third study

12:32

Gino and her co-author are investigating

12:34

the idea that people who cheat people

12:37

that lie who are dishonest are actually

12:40

more creative and they call the paper

12:42

Evil Genius how dishonesty can lead to

12:45

Greater creativity

12:48

really so let’s quickly go through how

12:51

the study worked participants were

12:52

brought into a lab where they were sat

12:53

at a machine with a virtual coin

12:56

flipping mechanism what the participants

12:58

are asked to do is to predict whether

13:00

the coin will flip heads or tails and

13:02

then they would push a button to

13:04

actually flip the coin and if they had

13:06

predicted correctly about whether it

13:07

would go heads or tails then they would

13:09

get a dollar so again there’s a strong

13:10

incentive to cheat so the participants

13:12

were right down on a piece of paper how

13:14

many predictions they got correct and

13:15

then they would hand that to the

13:16

researcher in order to get paid but then

13:18

of course the researchers would then go

13:19

back and look at the machine that they

13:21

were flipping the coin on to see how

13:23

many they actually got correct and then

13:24

they were able to tell how many times

13:26

that participant had cheated so after

13:28

they had completed the coin flipping

13:30

task they were then given a creativity

13:32

task and the creativity task was how

13:34

many different uses can you think of for

13:36

a piece of newspaper so in Psychology

13:38

this is a pretty common technique for

13:40

testing creativity you give somebody an

13:42

inanimate object and then you say how

13:43

many uses can you think of for this

13:46

inanimate object and again with this

13:48

study we see a very strong effect size

13:50

remember the magic number that academics

13:52

look for is p less than 0.05 and here we

13:56

have P less than 0.01 so basically what

13:59

that means is that there’s an extremely

14:00

high likelihood that the effect that the

14:02

academics are seeing is caused by the

14:04

manipulation that they did so again our

14:06

Vigilantes are suspicious but this one

14:09

is interesting because our Vigilantes

14:10

were able to actually get the data set

14:12

from Geno several years ago so they got

14:15

this data set directly from Geno so

14:17

again when our Vigilantes look into the

14:19

data they find some weird things going

14:21

on as you can see it seems to be sorted

14:23

by two things firstly by the number of

14:26

times the participant cheated so all the

14:28

people who didn’t cheat at all are zeros

14:30

and then the number of responses is the

14:32

number of different uses for a newspaper

14:34

that that participant could come up with

14:35

and those are clearly ranked in

14:37

ascending order but as you can see from

14:39

this next screenshot some of the

14:40

cheaters are out of order so these are

14:42

the people who cheated once who

14:44

basically over reported one time and the

14:46

number of uses that they could come up

14:47

with for the newspaper a route of

14:49

sequence here we have 3 4 13 then 9 and

14:53

then back down to five again then back

14:55

up to nine then five then nine and eight

14:56

the nine is just a total mess right so

14:59

these ones that are highlighted in

15:00

yellow are the… [Content truncated to 3000 words]

WRITE MY PAPER


Comments

Leave a Reply